diff --git a/.gitattributes b/.gitattributes
index 517cc278bfb10732bfccc9ac530f98efa1485385..93db076dedefb6e101c31fd03726b1e6ff43dbc5 100644
--- a/.gitattributes
+++ b/.gitattributes
@@ -151,3 +151,14 @@ runs/l2r90-ssl/checkpoint-360/tokenizer.json filter=lfs diff=lfs merge=lfs -text
runs/l2r90-ssl/checkpoint-3600/tokenizer.json filter=lfs diff=lfs merge=lfs -text
runs/l2r90-ssl/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lfs -text
runs/l2r90-ssl/checkpoint-720/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-wd0033-ssl/checkpoint-1080/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-wd0033-ssl/checkpoint-1440/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-wd0033-ssl/checkpoint-1800/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-wd0033-ssl/checkpoint-2160/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-wd0033-ssl/checkpoint-2520/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-wd0033-ssl/checkpoint-2880/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-wd0033-ssl/checkpoint-3240/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-wd0033-ssl/checkpoint-360/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-wd0033-ssl/checkpoint-3600/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-wd0033-ssl/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-wd0033-ssl/checkpoint-720/tokenizer.json filter=lfs diff=lfs merge=lfs -text
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1080/chat_template.jinja b/runs/l2r90-wd0033-ssl/checkpoint-1080/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1080/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1080/config.json b/runs/l2r90-wd0033-ssl/checkpoint-1080/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1080/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1080/generation_config.json b/runs/l2r90-wd0033-ssl/checkpoint-1080/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1080/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1080/model.safetensors b/runs/l2r90-wd0033-ssl/checkpoint-1080/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..ff5afabee3718065b7a4abdea3a0458498a9635b
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1080/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:05d276bba89241cdfe41ad12c21051d55c24c3dcde865d8c0653483a5be82d87
+size 583356232
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1080/optimizer.pt b/runs/l2r90-wd0033-ssl/checkpoint-1080/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..25d41f389f53e1fc65fcd8be829a421ad6927fee
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1080/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:349ea57798ecb57f2f42c2ab9677e165728876bb9949ba6bc65582770869e044
+size 1166825338
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1080/rng_state_0.pth b/runs/l2r90-wd0033-ssl/checkpoint-1080/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..ae48b9e23a260e22a8ccc20a31ba16b6e9b46e7b
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1080/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5e63a07811ad2d31f8c3bf0f3c024082f94aadfb69efe4f765095eae4cc09ead
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1080/rng_state_1.pth b/runs/l2r90-wd0033-ssl/checkpoint-1080/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..8482b4d45cac5b143de68ed6cc7f966677728d1f
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1080/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c6a55c1c2be49b9fbc3457ee2d0907b699099ec31ed5076ceb68dd33d2cc38fe
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1080/scheduler.pt b/runs/l2r90-wd0033-ssl/checkpoint-1080/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..6460bf7d5e8aadff08654c5c446813d18e721d62
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1080/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bd8f1c3c3cf1427d3265ecb044cf76510643402bb89ce35df8f605375b066d14
+size 1064
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1080/tokenizer.json b/runs/l2r90-wd0033-ssl/checkpoint-1080/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1080/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1080/tokenizer_config.json b/runs/l2r90-wd0033-ssl/checkpoint-1080/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1080/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1080/trainer_state.json b/runs/l2r90-wd0033-ssl/checkpoint-1080/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..ffaeaa4ef9cc0163b5b5e5248605a4b07bd5e16d
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1080/trainer_state.json
@@ -0,0 +1,7706 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 15.0,
+ "eval_steps": 500,
+ "global_step": 1080,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013961605584642234,
+ "grad_norm": 0.26733073592185974,
+ "learning_rate": 0.0,
+ "loss": 7.990724086761475,
+ "step": 1
+ },
+ {
+ "epoch": 0.027923211169284468,
+ "grad_norm": 0.26712340116500854,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 7.990110874176025,
+ "step": 2
+ },
+ {
+ "epoch": 0.041884816753926704,
+ "grad_norm": 0.2646932601928711,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 7.956935405731201,
+ "step": 3
+ },
+ {
+ "epoch": 0.055846422338568937,
+ "grad_norm": 0.24684545397758484,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 7.890539646148682,
+ "step": 4
+ },
+ {
+ "epoch": 0.06980802792321117,
+ "grad_norm": 0.21016643941402435,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 7.8146138191223145,
+ "step": 5
+ },
+ {
+ "epoch": 0.08376963350785341,
+ "grad_norm": 0.17992742359638214,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 7.724307060241699,
+ "step": 6
+ },
+ {
+ "epoch": 0.09773123909249563,
+ "grad_norm": 0.14356893301010132,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 7.645662307739258,
+ "step": 7
+ },
+ {
+ "epoch": 0.11169284467713787,
+ "grad_norm": 0.10830964893102646,
+ "learning_rate": 4.2e-05,
+ "loss": 7.582019329071045,
+ "step": 8
+ },
+ {
+ "epoch": 0.1256544502617801,
+ "grad_norm": 0.08265161514282227,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 7.5230793952941895,
+ "step": 9
+ },
+ {
+ "epoch": 0.13961605584642234,
+ "grad_norm": 0.06423710286617279,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 7.473812103271484,
+ "step": 10
+ },
+ {
+ "epoch": 0.15357766143106458,
+ "grad_norm": 0.05132246017456055,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 7.440786838531494,
+ "step": 11
+ },
+ {
+ "epoch": 0.16753926701570682,
+ "grad_norm": 0.040572404861450195,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 7.41290283203125,
+ "step": 12
+ },
+ {
+ "epoch": 0.18150087260034903,
+ "grad_norm": 0.03138705715537071,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 7.39749002456665,
+ "step": 13
+ },
+ {
+ "epoch": 0.19546247818499127,
+ "grad_norm": 0.02553008496761322,
+ "learning_rate": 7.8e-05,
+ "loss": 7.387092590332031,
+ "step": 14
+ },
+ {
+ "epoch": 0.2094240837696335,
+ "grad_norm": 0.022178977727890015,
+ "learning_rate": 8.4e-05,
+ "loss": 7.378993511199951,
+ "step": 15
+ },
+ {
+ "epoch": 0.22338568935427575,
+ "grad_norm": 0.019915293902158737,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 7.37220573425293,
+ "step": 16
+ },
+ {
+ "epoch": 0.23734729493891799,
+ "grad_norm": 0.017390629276633263,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 7.366759300231934,
+ "step": 17
+ },
+ {
+ "epoch": 0.2513089005235602,
+ "grad_norm": 0.015472771599888802,
+ "learning_rate": 0.000102,
+ "loss": 7.3652663230896,
+ "step": 18
+ },
+ {
+ "epoch": 0.26527050610820246,
+ "grad_norm": 0.013599214144051075,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 7.359714984893799,
+ "step": 19
+ },
+ {
+ "epoch": 0.2792321116928447,
+ "grad_norm": 0.012442861683666706,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 7.353302001953125,
+ "step": 20
+ },
+ {
+ "epoch": 0.2931937172774869,
+ "grad_norm": 0.010471965186297894,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 7.357417583465576,
+ "step": 21
+ },
+ {
+ "epoch": 0.30715532286212915,
+ "grad_norm": 0.009647224098443985,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 7.351763725280762,
+ "step": 22
+ },
+ {
+ "epoch": 0.32111692844677137,
+ "grad_norm": 0.008228210732340813,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 7.345434188842773,
+ "step": 23
+ },
+ {
+ "epoch": 0.33507853403141363,
+ "grad_norm": 0.007454514969140291,
+ "learning_rate": 0.000138,
+ "loss": 7.342520236968994,
+ "step": 24
+ },
+ {
+ "epoch": 0.34904013961605584,
+ "grad_norm": 0.0069856527261435986,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 7.338870048522949,
+ "step": 25
+ },
+ {
+ "epoch": 0.36300174520069806,
+ "grad_norm": 0.006091665010899305,
+ "learning_rate": 0.00015,
+ "loss": 7.328121662139893,
+ "step": 26
+ },
+ {
+ "epoch": 0.3769633507853403,
+ "grad_norm": 0.005525320768356323,
+ "learning_rate": 0.000156,
+ "loss": 7.335029602050781,
+ "step": 27
+ },
+ {
+ "epoch": 0.39092495636998253,
+ "grad_norm": 0.0061133443377912045,
+ "learning_rate": 0.000162,
+ "loss": 7.319417953491211,
+ "step": 28
+ },
+ {
+ "epoch": 0.4048865619546248,
+ "grad_norm": 0.006040909793227911,
+ "learning_rate": 0.000168,
+ "loss": 7.326376438140869,
+ "step": 29
+ },
+ {
+ "epoch": 0.418848167539267,
+ "grad_norm": 0.006003216840326786,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 7.319552421569824,
+ "step": 30
+ },
+ {
+ "epoch": 0.4328097731239092,
+ "grad_norm": 0.007096003741025925,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 7.323486328125,
+ "step": 31
+ },
+ {
+ "epoch": 0.4467713787085515,
+ "grad_norm": 0.007725459989160299,
+ "learning_rate": 0.000186,
+ "loss": 7.3190717697143555,
+ "step": 32
+ },
+ {
+ "epoch": 0.4607329842931937,
+ "grad_norm": 0.007338670082390308,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 7.317154884338379,
+ "step": 33
+ },
+ {
+ "epoch": 0.47469458987783597,
+ "grad_norm": 0.007709544617682695,
+ "learning_rate": 0.000198,
+ "loss": 7.311826705932617,
+ "step": 34
+ },
+ {
+ "epoch": 0.4886561954624782,
+ "grad_norm": 0.009249784983694553,
+ "learning_rate": 0.000204,
+ "loss": 7.322498798370361,
+ "step": 35
+ },
+ {
+ "epoch": 0.5026178010471204,
+ "grad_norm": 0.006950767710804939,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 7.314571380615234,
+ "step": 36
+ },
+ {
+ "epoch": 0.5165794066317626,
+ "grad_norm": 0.007583985570818186,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 7.309540748596191,
+ "step": 37
+ },
+ {
+ "epoch": 0.5305410122164049,
+ "grad_norm": 0.014731820672750473,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 7.307004928588867,
+ "step": 38
+ },
+ {
+ "epoch": 0.5445026178010471,
+ "grad_norm": 0.016276473179459572,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 7.299642562866211,
+ "step": 39
+ },
+ {
+ "epoch": 0.5584642233856894,
+ "grad_norm": 0.020018689334392548,
+ "learning_rate": 0.000234,
+ "loss": 7.298466682434082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5724258289703316,
+ "grad_norm": 0.014941536821424961,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 7.289968967437744,
+ "step": 41
+ },
+ {
+ "epoch": 0.5863874345549738,
+ "grad_norm": 0.018299812451004982,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 7.276564598083496,
+ "step": 42
+ },
+ {
+ "epoch": 0.6003490401396161,
+ "grad_norm": 0.01615075021982193,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.2575554847717285,
+ "step": 43
+ },
+ {
+ "epoch": 0.6143106457242583,
+ "grad_norm": 0.013063831254839897,
+ "learning_rate": 0.000258,
+ "loss": 7.253885269165039,
+ "step": 44
+ },
+ {
+ "epoch": 0.6282722513089005,
+ "grad_norm": 0.015211676247417927,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.245779037475586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6422338568935427,
+ "grad_norm": 0.02093500830233097,
+ "learning_rate": 0.00027,
+ "loss": 7.225130081176758,
+ "step": 46
+ },
+ {
+ "epoch": 0.6561954624781849,
+ "grad_norm": 0.01808053068816662,
+ "learning_rate": 0.000276,
+ "loss": 7.209654331207275,
+ "step": 47
+ },
+ {
+ "epoch": 0.6701570680628273,
+ "grad_norm": 0.02209644205868244,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.197124481201172,
+ "step": 48
+ },
+ {
+ "epoch": 0.6841186736474695,
+ "grad_norm": 0.023524878546595573,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.1841936111450195,
+ "step": 49
+ },
+ {
+ "epoch": 0.6980802792321117,
+ "grad_norm": 0.031148087233304977,
+ "learning_rate": 0.000294,
+ "loss": 7.188040733337402,
+ "step": 50
+ },
+ {
+ "epoch": 0.7120418848167539,
+ "grad_norm": 0.02931929938495159,
+ "learning_rate": 0.0003,
+ "loss": 7.17371940612793,
+ "step": 51
+ },
+ {
+ "epoch": 0.7260034904013961,
+ "grad_norm": 0.01611483097076416,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.1714324951171875,
+ "step": 52
+ },
+ {
+ "epoch": 0.7399650959860384,
+ "grad_norm": 0.02619907818734646,
+ "learning_rate": 0.000312,
+ "loss": 7.158261299133301,
+ "step": 53
+ },
+ {
+ "epoch": 0.7539267015706806,
+ "grad_norm": 0.049462560564279556,
+ "learning_rate": 0.000318,
+ "loss": 7.167076110839844,
+ "step": 54
+ },
+ {
+ "epoch": 0.7678883071553229,
+ "grad_norm": 0.09835070371627808,
+ "learning_rate": 0.000324,
+ "loss": 7.154292106628418,
+ "step": 55
+ },
+ {
+ "epoch": 0.7818499127399651,
+ "grad_norm": 0.11406510323286057,
+ "learning_rate": 0.00033,
+ "loss": 7.162016868591309,
+ "step": 56
+ },
+ {
+ "epoch": 0.7958115183246073,
+ "grad_norm": 0.027026303112506866,
+ "learning_rate": 0.000336,
+ "loss": 7.141416549682617,
+ "step": 57
+ },
+ {
+ "epoch": 0.8097731239092496,
+ "grad_norm": 0.06690579652786255,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.13959264755249,
+ "step": 58
+ },
+ {
+ "epoch": 0.8237347294938918,
+ "grad_norm": 0.020919431000947952,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.126513481140137,
+ "step": 59
+ },
+ {
+ "epoch": 0.837696335078534,
+ "grad_norm": 0.05618130415678024,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.124345779418945,
+ "step": 60
+ },
+ {
+ "epoch": 0.8516579406631762,
+ "grad_norm": 0.03402302414178848,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.100615501403809,
+ "step": 61
+ },
+ {
+ "epoch": 0.8656195462478184,
+ "grad_norm": 0.03646906837821007,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.116507530212402,
+ "step": 62
+ },
+ {
+ "epoch": 0.8795811518324608,
+ "grad_norm": 0.03076397068798542,
+ "learning_rate": 0.000372,
+ "loss": 7.097659111022949,
+ "step": 63
+ },
+ {
+ "epoch": 0.893542757417103,
+ "grad_norm": 0.03329646959900856,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.097290992736816,
+ "step": 64
+ },
+ {
+ "epoch": 0.9075043630017452,
+ "grad_norm": 0.01998630352318287,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.092268943786621,
+ "step": 65
+ },
+ {
+ "epoch": 0.9214659685863874,
+ "grad_norm": 0.03278940171003342,
+ "learning_rate": 0.00039,
+ "loss": 7.088292121887207,
+ "step": 66
+ },
+ {
+ "epoch": 0.9354275741710296,
+ "grad_norm": 0.024620421230793,
+ "learning_rate": 0.000396,
+ "loss": 7.078650951385498,
+ "step": 67
+ },
+ {
+ "epoch": 0.9493891797556719,
+ "grad_norm": 0.03750383481383324,
+ "learning_rate": 0.000402,
+ "loss": 7.052778244018555,
+ "step": 68
+ },
+ {
+ "epoch": 0.9633507853403142,
+ "grad_norm": 0.07558075338602066,
+ "learning_rate": 0.000408,
+ "loss": 7.045991897583008,
+ "step": 69
+ },
+ {
+ "epoch": 0.9773123909249564,
+ "grad_norm": 0.16021482646465302,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.023152828216553,
+ "step": 70
+ },
+ {
+ "epoch": 0.9912739965095986,
+ "grad_norm": 0.24981893599033356,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.067540645599365,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.10255520790815353,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 4.425640106201172,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8861758708953857,
+ "eval_runtime": 58.2699,
+ "eval_samples_per_second": 41.908,
+ "eval_steps_per_second": 0.669,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139616055846423,
+ "grad_norm": 0.13374797999858856,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.073373317718506,
+ "step": 73
+ },
+ {
+ "epoch": 1.0279232111692844,
+ "grad_norm": 0.05079387128353119,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.0243306159973145,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418848167539267,
+ "grad_norm": 0.10056068748235703,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.014476299285889,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558464223385688,
+ "grad_norm": 0.06995758414268494,
+ "learning_rate": 0.00045,
+ "loss": 7.011931419372559,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698080279232112,
+ "grad_norm": 0.0483892522752285,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 6.9822797775268555,
+ "step": 77
+ },
+ {
+ "epoch": 1.0837696335078535,
+ "grad_norm": 0.08408170193433762,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 6.964348793029785,
+ "step": 78
+ },
+ {
+ "epoch": 1.0977312390924956,
+ "grad_norm": 0.08027710020542145,
+ "learning_rate": 0.000468,
+ "loss": 6.96852445602417,
+ "step": 79
+ },
+ {
+ "epoch": 1.111692844677138,
+ "grad_norm": 0.05647159367799759,
+ "learning_rate": 0.000474,
+ "loss": 6.944484710693359,
+ "step": 80
+ },
+ {
+ "epoch": 1.12565445026178,
+ "grad_norm": 0.07464330643415451,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 6.9246745109558105,
+ "step": 81
+ },
+ {
+ "epoch": 1.1396160558464223,
+ "grad_norm": 0.04451674595475197,
+ "learning_rate": 0.000486,
+ "loss": 6.895403861999512,
+ "step": 82
+ },
+ {
+ "epoch": 1.1535776614310647,
+ "grad_norm": 0.05820256471633911,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 6.887338638305664,
+ "step": 83
+ },
+ {
+ "epoch": 1.1675392670157068,
+ "grad_norm": 0.02925296127796173,
+ "learning_rate": 0.000498,
+ "loss": 6.857797622680664,
+ "step": 84
+ },
+ {
+ "epoch": 1.181500872600349,
+ "grad_norm": 0.05096987634897232,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 6.854822158813477,
+ "step": 85
+ },
+ {
+ "epoch": 1.1954624781849912,
+ "grad_norm": 0.03483197093009949,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 6.840914726257324,
+ "step": 86
+ },
+ {
+ "epoch": 1.2094240837696335,
+ "grad_norm": 0.03323264792561531,
+ "learning_rate": 0.000516,
+ "loss": 6.834497451782227,
+ "step": 87
+ },
+ {
+ "epoch": 1.2233856893542758,
+ "grad_norm": 0.03290868178009987,
+ "learning_rate": 0.000522,
+ "loss": 6.820746898651123,
+ "step": 88
+ },
+ {
+ "epoch": 1.237347294938918,
+ "grad_norm": 0.025398844853043556,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 6.795746803283691,
+ "step": 89
+ },
+ {
+ "epoch": 1.2513089005235603,
+ "grad_norm": 0.030700596049427986,
+ "learning_rate": 0.000534,
+ "loss": 6.7867889404296875,
+ "step": 90
+ },
+ {
+ "epoch": 1.2652705061082026,
+ "grad_norm": 0.019589420408010483,
+ "learning_rate": 0.00054,
+ "loss": 6.77443790435791,
+ "step": 91
+ },
+ {
+ "epoch": 1.2792321116928447,
+ "grad_norm": 0.03313920646905899,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 6.762271881103516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2931937172774868,
+ "grad_norm": 0.025389468297362328,
+ "learning_rate": 0.000552,
+ "loss": 6.750258445739746,
+ "step": 93
+ },
+ {
+ "epoch": 1.307155322862129,
+ "grad_norm": 0.040296655148267746,
+ "learning_rate": 0.000558,
+ "loss": 6.729808807373047,
+ "step": 94
+ },
+ {
+ "epoch": 1.3211169284467714,
+ "grad_norm": 0.04735933244228363,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 6.737982749938965,
+ "step": 95
+ },
+ {
+ "epoch": 1.3350785340314135,
+ "grad_norm": 0.05529680848121643,
+ "learning_rate": 0.00057,
+ "loss": 6.735714912414551,
+ "step": 96
+ },
+ {
+ "epoch": 1.3490401396160558,
+ "grad_norm": 0.07148554176092148,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 6.708859920501709,
+ "step": 97
+ },
+ {
+ "epoch": 1.3630017452006982,
+ "grad_norm": 0.059708479791879654,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 6.708860397338867,
+ "step": 98
+ },
+ {
+ "epoch": 1.3769633507853403,
+ "grad_norm": 0.027111195027828217,
+ "learning_rate": 0.000588,
+ "loss": 6.686685085296631,
+ "step": 99
+ },
+ {
+ "epoch": 1.3909249563699826,
+ "grad_norm": 0.0500967763364315,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 6.676325798034668,
+ "step": 100
+ },
+ {
+ "epoch": 1.404886561954625,
+ "grad_norm": 0.044565796852111816,
+ "learning_rate": 0.0006,
+ "loss": 6.674530029296875,
+ "step": 101
+ },
+ {
+ "epoch": 1.418848167539267,
+ "grad_norm": 0.03608180209994316,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 6.653741359710693,
+ "step": 102
+ },
+ {
+ "epoch": 1.4328097731239091,
+ "grad_norm": 0.030394982546567917,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 6.643152236938477,
+ "step": 103
+ },
+ {
+ "epoch": 1.4467713787085514,
+ "grad_norm": 0.040185656398534775,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 6.628686904907227,
+ "step": 104
+ },
+ {
+ "epoch": 1.4607329842931938,
+ "grad_norm": 0.021049322560429573,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 6.610864639282227,
+ "step": 105
+ },
+ {
+ "epoch": 1.4746945898778359,
+ "grad_norm": 0.04164277762174606,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 6.602880954742432,
+ "step": 106
+ },
+ {
+ "epoch": 1.4886561954624782,
+ "grad_norm": 0.0389864556491375,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 6.599456787109375,
+ "step": 107
+ },
+ {
+ "epoch": 1.5026178010471205,
+ "grad_norm": 0.05805973336100578,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 6.588700294494629,
+ "step": 108
+ },
+ {
+ "epoch": 1.5165794066317626,
+ "grad_norm": 0.08825036883354187,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 6.587496757507324,
+ "step": 109
+ },
+ {
+ "epoch": 1.530541012216405,
+ "grad_norm": 0.12521523237228394,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 6.626501083374023,
+ "step": 110
+ },
+ {
+ "epoch": 1.5445026178010473,
+ "grad_norm": 0.07429645210504532,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 6.590241432189941,
+ "step": 111
+ },
+ {
+ "epoch": 1.5584642233856894,
+ "grad_norm": 0.06463009119033813,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 6.578714370727539,
+ "step": 112
+ },
+ {
+ "epoch": 1.5724258289703315,
+ "grad_norm": 0.07508203387260437,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 6.561270713806152,
+ "step": 113
+ },
+ {
+ "epoch": 1.5863874345549738,
+ "grad_norm": 0.04125197231769562,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 6.558821678161621,
+ "step": 114
+ },
+ {
+ "epoch": 1.600349040139616,
+ "grad_norm": 0.05457386001944542,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.541611671447754,
+ "step": 115
+ },
+ {
+ "epoch": 1.6143106457242582,
+ "grad_norm": 0.03684385493397713,
+ "learning_rate": 0.000599973080557108,
+ "loss": 6.547738075256348,
+ "step": 116
+ },
+ {
+ "epoch": 1.6282722513089005,
+ "grad_norm": 0.05888746306300163,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.529491424560547,
+ "step": 117
+ },
+ {
+ "epoch": 1.6422338568935428,
+ "grad_norm": 0.056668370962142944,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.519444942474365,
+ "step": 118
+ },
+ {
+ "epoch": 1.656195462478185,
+ "grad_norm": 0.09123896062374115,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.522984981536865,
+ "step": 119
+ },
+ {
+ "epoch": 1.6701570680628273,
+ "grad_norm": 0.12380023300647736,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.515375137329102,
+ "step": 120
+ },
+ {
+ "epoch": 1.6841186736474696,
+ "grad_norm": 0.06953322887420654,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.500491142272949,
+ "step": 121
+ },
+ {
+ "epoch": 1.6980802792321117,
+ "grad_norm": 0.08613869547843933,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.5097198486328125,
+ "step": 122
+ },
+ {
+ "epoch": 1.7120418848167538,
+ "grad_norm": 0.08296383172273636,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.487521171569824,
+ "step": 123
+ },
+ {
+ "epoch": 1.7260034904013961,
+ "grad_norm": 0.05744783207774162,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.485134124755859,
+ "step": 124
+ },
+ {
+ "epoch": 1.7399650959860384,
+ "grad_norm": 0.04108566418290138,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.466182708740234,
+ "step": 125
+ },
+ {
+ "epoch": 1.7539267015706805,
+ "grad_norm": 0.05374931916594505,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.479011535644531,
+ "step": 126
+ },
+ {
+ "epoch": 1.7678883071553229,
+ "grad_norm": 0.03867914527654648,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.465902328491211,
+ "step": 127
+ },
+ {
+ "epoch": 1.7818499127399652,
+ "grad_norm": 0.05477374792098999,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.453082084655762,
+ "step": 128
+ },
+ {
+ "epoch": 1.7958115183246073,
+ "grad_norm": 0.06508298963308334,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.446089744567871,
+ "step": 129
+ },
+ {
+ "epoch": 1.8097731239092496,
+ "grad_norm": 0.06388624757528305,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.451748847961426,
+ "step": 130
+ },
+ {
+ "epoch": 1.823734729493892,
+ "grad_norm": 0.04223188757896423,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.4492387771606445,
+ "step": 131
+ },
+ {
+ "epoch": 1.837696335078534,
+ "grad_norm": 0.04528658092021942,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.409390926361084,
+ "step": 132
+ },
+ {
+ "epoch": 1.8516579406631761,
+ "grad_norm": 0.04962967336177826,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.415774345397949,
+ "step": 133
+ },
+ {
+ "epoch": 1.8656195462478184,
+ "grad_norm": 0.031944390386343,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.411528587341309,
+ "step": 134
+ },
+ {
+ "epoch": 1.8795811518324608,
+ "grad_norm": 0.028411149978637695,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.392329216003418,
+ "step": 135
+ },
+ {
+ "epoch": 1.8935427574171029,
+ "grad_norm": 0.03516148403286934,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.392238616943359,
+ "step": 136
+ },
+ {
+ "epoch": 1.9075043630017452,
+ "grad_norm": 0.027800120413303375,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.39179801940918,
+ "step": 137
+ },
+ {
+ "epoch": 1.9214659685863875,
+ "grad_norm": 0.02362987957894802,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.378960609436035,
+ "step": 138
+ },
+ {
+ "epoch": 1.9354275741710296,
+ "grad_norm": 0.033999308943748474,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.36802339553833,
+ "step": 139
+ },
+ {
+ "epoch": 1.949389179755672,
+ "grad_norm": 0.03290879726409912,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.366308689117432,
+ "step": 140
+ },
+ {
+ "epoch": 1.9633507853403143,
+ "grad_norm": 0.0286843404173851,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.350399971008301,
+ "step": 141
+ },
+ {
+ "epoch": 1.9773123909249564,
+ "grad_norm": 0.027145877480506897,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.33477783203125,
+ "step": 142
+ },
+ {
+ "epoch": 1.9912739965095985,
+ "grad_norm": 0.03530878946185112,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.322859764099121,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.028133150190114975,
+ "learning_rate": 0.000599778806120077,
+ "loss": 3.955288887023926,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7927101254463196,
+ "eval_runtime": 60.8372,
+ "eval_samples_per_second": 40.14,
+ "eval_steps_per_second": 0.641,
+ "step": 144
+ },
+ {
+ "epoch": 2.013961605584642,
+ "grad_norm": 0.03054557554423809,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.311975479125977,
+ "step": 145
+ },
+ {
+ "epoch": 2.0279232111692846,
+ "grad_norm": 0.05406459793448448,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.3030853271484375,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418848167539267,
+ "grad_norm": 0.10911019891500473,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.329056739807129,
+ "step": 147
+ },
+ {
+ "epoch": 2.055846422338569,
+ "grad_norm": 0.13948315382003784,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.362421035766602,
+ "step": 148
+ },
+ {
+ "epoch": 2.0698080279232114,
+ "grad_norm": 0.13176991045475006,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.3888139724731445,
+ "step": 149
+ },
+ {
+ "epoch": 2.0837696335078535,
+ "grad_norm": 0.09615885466337204,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.35154914855957,
+ "step": 150
+ },
+ {
+ "epoch": 2.0977312390924956,
+ "grad_norm": 0.06462561339139938,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.335790634155273,
+ "step": 151
+ },
+ {
+ "epoch": 2.1116928446771377,
+ "grad_norm": 0.05186621472239494,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.324396133422852,
+ "step": 152
+ },
+ {
+ "epoch": 2.1256544502617802,
+ "grad_norm": 0.06095981225371361,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.318549156188965,
+ "step": 153
+ },
+ {
+ "epoch": 2.1396160558464223,
+ "grad_norm": 0.036121442914009094,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.300607681274414,
+ "step": 154
+ },
+ {
+ "epoch": 2.1535776614310644,
+ "grad_norm": 0.055716633796691895,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.307400703430176,
+ "step": 155
+ },
+ {
+ "epoch": 2.167539267015707,
+ "grad_norm": 0.037626396864652634,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.287364959716797,
+ "step": 156
+ },
+ {
+ "epoch": 2.181500872600349,
+ "grad_norm": 0.04329098388552666,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.28244686126709,
+ "step": 157
+ },
+ {
+ "epoch": 2.195462478184991,
+ "grad_norm": 0.031868450343608856,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.2602410316467285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2094240837696333,
+ "grad_norm": 0.03647344186902046,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.263988971710205,
+ "step": 159
+ },
+ {
+ "epoch": 2.223385689354276,
+ "grad_norm": 0.03141765296459198,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.263205051422119,
+ "step": 160
+ },
+ {
+ "epoch": 2.237347294938918,
+ "grad_norm": 0.028253229334950447,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.253107070922852,
+ "step": 161
+ },
+ {
+ "epoch": 2.25130890052356,
+ "grad_norm": 0.03608971834182739,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.2571702003479,
+ "step": 162
+ },
+ {
+ "epoch": 2.2652705061082026,
+ "grad_norm": 0.02644292451441288,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.241732597351074,
+ "step": 163
+ },
+ {
+ "epoch": 2.2792321116928447,
+ "grad_norm": 0.028604455292224884,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.230289459228516,
+ "step": 164
+ },
+ {
+ "epoch": 2.2931937172774868,
+ "grad_norm": 0.023704614490270615,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.226101398468018,
+ "step": 165
+ },
+ {
+ "epoch": 2.3071553228621293,
+ "grad_norm": 0.020691027864813805,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.2096405029296875,
+ "step": 166
+ },
+ {
+ "epoch": 2.3211169284467714,
+ "grad_norm": 0.025598427280783653,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.22075080871582,
+ "step": 167
+ },
+ {
+ "epoch": 2.3350785340314135,
+ "grad_norm": 0.030546002089977264,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.216886520385742,
+ "step": 168
+ },
+ {
+ "epoch": 2.349040139616056,
+ "grad_norm": 0.026066245511174202,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.2022504806518555,
+ "step": 169
+ },
+ {
+ "epoch": 2.363001745200698,
+ "grad_norm": 0.023802239447832108,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.189509391784668,
+ "step": 170
+ },
+ {
+ "epoch": 2.3769633507853403,
+ "grad_norm": 0.029805002734065056,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.186277389526367,
+ "step": 171
+ },
+ {
+ "epoch": 2.3909249563699824,
+ "grad_norm": 0.043327830731868744,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.164876937866211,
+ "step": 172
+ },
+ {
+ "epoch": 2.404886561954625,
+ "grad_norm": 0.058787211775779724,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.179936408996582,
+ "step": 173
+ },
+ {
+ "epoch": 2.418848167539267,
+ "grad_norm": 0.0787397176027298,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.1830244064331055,
+ "step": 174
+ },
+ {
+ "epoch": 2.432809773123909,
+ "grad_norm": 0.0899854376912117,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.190343856811523,
+ "step": 175
+ },
+ {
+ "epoch": 2.4467713787085517,
+ "grad_norm": 0.06520707905292511,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.185057163238525,
+ "step": 176
+ },
+ {
+ "epoch": 2.4607329842931938,
+ "grad_norm": 0.1002504974603653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.171279430389404,
+ "step": 177
+ },
+ {
+ "epoch": 2.474694589877836,
+ "grad_norm": 0.09591004997491837,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.174382209777832,
+ "step": 178
+ },
+ {
+ "epoch": 2.488656195462478,
+ "grad_norm": 0.12127121537923813,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.185717582702637,
+ "step": 179
+ },
+ {
+ "epoch": 2.5026178010471205,
+ "grad_norm": 0.09640590846538544,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.196681976318359,
+ "step": 180
+ },
+ {
+ "epoch": 2.5165794066317626,
+ "grad_norm": 0.048341453075408936,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.158774375915527,
+ "step": 181
+ },
+ {
+ "epoch": 2.530541012216405,
+ "grad_norm": 0.08206334710121155,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.17445707321167,
+ "step": 182
+ },
+ {
+ "epoch": 2.5445026178010473,
+ "grad_norm": 0.038934968411922455,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.170527458190918,
+ "step": 183
+ },
+ {
+ "epoch": 2.5584642233856894,
+ "grad_norm": 0.060987457633018494,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.1501994132995605,
+ "step": 184
+ },
+ {
+ "epoch": 2.5724258289703315,
+ "grad_norm": 0.033269546926021576,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.144186019897461,
+ "step": 185
+ },
+ {
+ "epoch": 2.5863874345549736,
+ "grad_norm": 0.046137236058712006,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.1236677169799805,
+ "step": 186
+ },
+ {
+ "epoch": 2.600349040139616,
+ "grad_norm": 0.038265109062194824,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.121979236602783,
+ "step": 187
+ },
+ {
+ "epoch": 2.614310645724258,
+ "grad_norm": 0.039128001779317856,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.124156951904297,
+ "step": 188
+ },
+ {
+ "epoch": 2.6282722513089007,
+ "grad_norm": 0.025303540751338005,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.108036994934082,
+ "step": 189
+ },
+ {
+ "epoch": 2.642233856893543,
+ "grad_norm": 0.04441308230161667,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.1211161613464355,
+ "step": 190
+ },
+ {
+ "epoch": 2.656195462478185,
+ "grad_norm": 0.03083532303571701,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.094561576843262,
+ "step": 191
+ },
+ {
+ "epoch": 2.670157068062827,
+ "grad_norm": 0.02951296605169773,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.091608047485352,
+ "step": 192
+ },
+ {
+ "epoch": 2.6841186736474696,
+ "grad_norm": 0.02385987527668476,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.087080001831055,
+ "step": 193
+ },
+ {
+ "epoch": 2.6980802792321117,
+ "grad_norm": 0.03315780311822891,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.088132381439209,
+ "step": 194
+ },
+ {
+ "epoch": 2.712041884816754,
+ "grad_norm": 0.02410835586488247,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.081460952758789,
+ "step": 195
+ },
+ {
+ "epoch": 2.7260034904013963,
+ "grad_norm": 0.033069487661123276,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.0696539878845215,
+ "step": 196
+ },
+ {
+ "epoch": 2.7399650959860384,
+ "grad_norm": 0.031288471072912216,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.047612190246582,
+ "step": 197
+ },
+ {
+ "epoch": 2.7539267015706805,
+ "grad_norm": 0.03399677947163582,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.019026756286621,
+ "step": 198
+ },
+ {
+ "epoch": 2.7678883071553226,
+ "grad_norm": 0.031948331743478775,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.050133228302002,
+ "step": 199
+ },
+ {
+ "epoch": 2.781849912739965,
+ "grad_norm": 0.049206435680389404,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.045332431793213,
+ "step": 200
+ },
+ {
+ "epoch": 2.7958115183246073,
+ "grad_norm": 0.07528849691152573,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.059202194213867,
+ "step": 201
+ },
+ {
+ "epoch": 2.80977312390925,
+ "grad_norm": 0.11815470457077026,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.077486038208008,
+ "step": 202
+ },
+ {
+ "epoch": 2.823734729493892,
+ "grad_norm": 0.0867348462343216,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.045670032501221,
+ "step": 203
+ },
+ {
+ "epoch": 2.837696335078534,
+ "grad_norm": 0.05716705322265625,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.034373760223389,
+ "step": 204
+ },
+ {
+ "epoch": 2.851657940663176,
+ "grad_norm": 0.05944284051656723,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.049911975860596,
+ "step": 205
+ },
+ {
+ "epoch": 2.8656195462478182,
+ "grad_norm": 0.03220470994710922,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.039861679077148,
+ "step": 206
+ },
+ {
+ "epoch": 2.8795811518324608,
+ "grad_norm": 0.057081177830696106,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.010714530944824,
+ "step": 207
+ },
+ {
+ "epoch": 2.893542757417103,
+ "grad_norm": 0.04037516191601753,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.018877983093262,
+ "step": 208
+ },
+ {
+ "epoch": 2.9075043630017454,
+ "grad_norm": 0.036492880433797836,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.009493827819824,
+ "step": 209
+ },
+ {
+ "epoch": 2.9214659685863875,
+ "grad_norm": 0.04785023257136345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.0140862464904785,
+ "step": 210
+ },
+ {
+ "epoch": 2.9354275741710296,
+ "grad_norm": 0.07721036672592163,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.030359268188477,
+ "step": 211
+ },
+ {
+ "epoch": 2.9493891797556717,
+ "grad_norm": 0.14220190048217773,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.045989990234375,
+ "step": 212
+ },
+ {
+ "epoch": 2.9633507853403143,
+ "grad_norm": 0.10265158116817474,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.043968200683594,
+ "step": 213
+ },
+ {
+ "epoch": 2.9773123909249564,
+ "grad_norm": 0.051398951560258865,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.022278785705566,
+ "step": 214
+ },
+ {
+ "epoch": 2.9912739965095985,
+ "grad_norm": 0.06044618785381317,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.017296314239502,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.03284750506281853,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 3.7544798851013184,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7531244158744812,
+ "eval_runtime": 60.3986,
+ "eval_samples_per_second": 40.431,
+ "eval_steps_per_second": 0.646,
+ "step": 216
+ },
+ {
+ "epoch": 3.013961605584642,
+ "grad_norm": 0.03895840048789978,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 5.9985857009887695,
+ "step": 217
+ },
+ {
+ "epoch": 3.0279232111692846,
+ "grad_norm": 0.04554992541670799,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 5.984710693359375,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418848167539267,
+ "grad_norm": 0.03359116613864899,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 5.984078407287598,
+ "step": 219
+ },
+ {
+ "epoch": 3.055846422338569,
+ "grad_norm": 0.035171881318092346,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 5.979424476623535,
+ "step": 220
+ },
+ {
+ "epoch": 3.0698080279232114,
+ "grad_norm": 0.032774072140455246,
+ "learning_rate": 0.000598278794687186,
+ "loss": 5.94467830657959,
+ "step": 221
+ },
+ {
+ "epoch": 3.0837696335078535,
+ "grad_norm": 0.032842908054590225,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 5.95986270904541,
+ "step": 222
+ },
+ {
+ "epoch": 3.0977312390924956,
+ "grad_norm": 0.02870488539338112,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 5.951186656951904,
+ "step": 223
+ },
+ {
+ "epoch": 3.1116928446771377,
+ "grad_norm": 0.030777374282479286,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 5.942493438720703,
+ "step": 224
+ },
+ {
+ "epoch": 3.1256544502617802,
+ "grad_norm": 0.026439018547534943,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 5.938275337219238,
+ "step": 225
+ },
+ {
+ "epoch": 3.1396160558464223,
+ "grad_norm": 0.028689472004771233,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 5.931982040405273,
+ "step": 226
+ },
+ {
+ "epoch": 3.1535776614310644,
+ "grad_norm": 0.028100362047553062,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 5.919692039489746,
+ "step": 227
+ },
+ {
+ "epoch": 3.167539267015707,
+ "grad_norm": 0.02407177910208702,
+ "learning_rate": 0.00059807235435179,
+ "loss": 5.909029006958008,
+ "step": 228
+ },
+ {
+ "epoch": 3.181500872600349,
+ "grad_norm": 0.0237748846411705,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 5.9131598472595215,
+ "step": 229
+ },
+ {
+ "epoch": 3.195462478184991,
+ "grad_norm": 0.022418556734919548,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 5.909396171569824,
+ "step": 230
+ },
+ {
+ "epoch": 3.2094240837696333,
+ "grad_norm": 0.024656571447849274,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 5.885243892669678,
+ "step": 231
+ },
+ {
+ "epoch": 3.223385689354276,
+ "grad_norm": 0.02136118710041046,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 5.898958206176758,
+ "step": 232
+ },
+ {
+ "epoch": 3.237347294938918,
+ "grad_norm": 0.025290116667747498,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 5.886986255645752,
+ "step": 233
+ },
+ {
+ "epoch": 3.25130890052356,
+ "grad_norm": 0.02656404674053192,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 5.897647857666016,
+ "step": 234
+ },
+ {
+ "epoch": 3.2652705061082026,
+ "grad_norm": 0.019188793376088142,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 5.875659942626953,
+ "step": 235
+ },
+ {
+ "epoch": 3.2792321116928447,
+ "grad_norm": 0.02308926172554493,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 5.859380722045898,
+ "step": 236
+ },
+ {
+ "epoch": 3.2931937172774868,
+ "grad_norm": 0.029775725677609444,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 5.873961448669434,
+ "step": 237
+ },
+ {
+ "epoch": 3.3071553228621293,
+ "grad_norm": 0.03875259310007095,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 5.863171100616455,
+ "step": 238
+ },
+ {
+ "epoch": 3.3211169284467714,
+ "grad_norm": 0.052446842193603516,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 5.886547088623047,
+ "step": 239
+ },
+ {
+ "epoch": 3.3350785340314135,
+ "grad_norm": 0.05583837628364563,
+ "learning_rate": 0.000597691359743243,
+ "loss": 5.8849287033081055,
+ "step": 240
+ },
+ {
+ "epoch": 3.349040139616056,
+ "grad_norm": 0.034729067236185074,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 5.863419532775879,
+ "step": 241
+ },
+ {
+ "epoch": 3.363001745200698,
+ "grad_norm": 0.03590013459324837,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 5.852169990539551,
+ "step": 242
+ },
+ {
+ "epoch": 3.3769633507853403,
+ "grad_norm": 0.035041507333517075,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 5.836690902709961,
+ "step": 243
+ },
+ {
+ "epoch": 3.3909249563699824,
+ "grad_norm": 0.02857939340174198,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 5.836485862731934,
+ "step": 244
+ },
+ {
+ "epoch": 3.404886561954625,
+ "grad_norm": 0.03176642581820488,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 5.860719203948975,
+ "step": 245
+ },
+ {
+ "epoch": 3.418848167539267,
+ "grad_norm": 0.026105748489499092,
+ "learning_rate": 0.000597488038559129,
+ "loss": 5.8354949951171875,
+ "step": 246
+ },
+ {
+ "epoch": 3.432809773123909,
+ "grad_norm": 0.029094981029629707,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 5.846247673034668,
+ "step": 247
+ },
+ {
+ "epoch": 3.4467713787085517,
+ "grad_norm": 0.02388785220682621,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 5.837607383728027,
+ "step": 248
+ },
+ {
+ "epoch": 3.4607329842931938,
+ "grad_norm": 0.03094472736120224,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 5.832090377807617,
+ "step": 249
+ },
+ {
+ "epoch": 3.474694589877836,
+ "grad_norm": 0.03227658197283745,
+ "learning_rate": 0.000597347745466109,
+ "loss": 5.83722448348999,
+ "step": 250
+ },
+ {
+ "epoch": 3.488656195462478,
+ "grad_norm": 0.042813923209905624,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 5.851722717285156,
+ "step": 251
+ },
+ {
+ "epoch": 3.5026178010471205,
+ "grad_norm": 0.05825525149703026,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.795719146728516,
+ "step": 252
+ },
+ {
+ "epoch": 3.5165794066317626,
+ "grad_norm": 0.0867447778582573,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.813750267028809,
+ "step": 253
+ },
+ {
+ "epoch": 3.530541012216405,
+ "grad_norm": 0.09192710369825363,
+ "learning_rate": 0.000597203657966319,
+ "loss": 5.823488235473633,
+ "step": 254
+ },
+ {
+ "epoch": 3.5445026178010473,
+ "grad_norm": 0.06069248169660568,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 5.814342021942139,
+ "step": 255
+ },
+ {
+ "epoch": 3.5584642233856894,
+ "grad_norm": 0.06259952485561371,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 5.834272384643555,
+ "step": 256
+ },
+ {
+ "epoch": 3.5724258289703315,
+ "grad_norm": 0.0513639897108078,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.807951927185059,
+ "step": 257
+ },
+ {
+ "epoch": 3.5863874345549736,
+ "grad_norm": 0.03327798470854759,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.790238857269287,
+ "step": 258
+ },
+ {
+ "epoch": 3.600349040139616,
+ "grad_norm": 0.03963128477334976,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.7956695556640625,
+ "step": 259
+ },
+ {
+ "epoch": 3.614310645724258,
+ "grad_norm": 0.031153453513979912,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 5.7874531745910645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6282722513089007,
+ "grad_norm": 0.031419433653354645,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 5.779650688171387,
+ "step": 261
+ },
+ {
+ "epoch": 3.642233856893543,
+ "grad_norm": 0.0288684144616127,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.776577949523926,
+ "step": 262
+ },
+ {
+ "epoch": 3.656195462478185,
+ "grad_norm": 0.028315918520092964,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.778571128845215,
+ "step": 263
+ },
+ {
+ "epoch": 3.670157068062827,
+ "grad_norm": 0.024977795779705048,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 5.776470184326172,
+ "step": 264
+ },
+ {
+ "epoch": 3.6841186736474696,
+ "grad_norm": 0.026822632178664207,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.755108833312988,
+ "step": 265
+ },
+ {
+ "epoch": 3.6980802792321117,
+ "grad_norm": 0.02738630585372448,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.74041748046875,
+ "step": 266
+ },
+ {
+ "epoch": 3.712041884816754,
+ "grad_norm": 0.02361280657351017,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.764027118682861,
+ "step": 267
+ },
+ {
+ "epoch": 3.7260034904013963,
+ "grad_norm": 0.029926009476184845,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.780390739440918,
+ "step": 268
+ },
+ {
+ "epoch": 3.7399650959860384,
+ "grad_norm": 0.03519507870078087,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.758105278015137,
+ "step": 269
+ },
+ {
+ "epoch": 3.7539267015706805,
+ "grad_norm": 0.030114633962512016,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.757230758666992,
+ "step": 270
+ },
+ {
+ "epoch": 3.7678883071553226,
+ "grad_norm": 0.02467748522758484,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.7615461349487305,
+ "step": 271
+ },
+ {
+ "epoch": 3.781849912739965,
+ "grad_norm": 0.03425183147192001,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.749448776245117,
+ "step": 272
+ },
+ {
+ "epoch": 3.7958115183246073,
+ "grad_norm": 0.046307582408189774,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.7356977462768555,
+ "step": 273
+ },
+ {
+ "epoch": 3.80977312390925,
+ "grad_norm": 0.054102856665849686,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.741869926452637,
+ "step": 274
+ },
+ {
+ "epoch": 3.823734729493892,
+ "grad_norm": 0.07026689499616623,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.721103668212891,
+ "step": 275
+ },
+ {
+ "epoch": 3.837696335078534,
+ "grad_norm": 0.0733039602637291,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.731414794921875,
+ "step": 276
+ },
+ {
+ "epoch": 3.851657940663176,
+ "grad_norm": 0.05342920497059822,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.734922885894775,
+ "step": 277
+ },
+ {
+ "epoch": 3.8656195462478182,
+ "grad_norm": 0.0504680834710598,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.732746124267578,
+ "step": 278
+ },
+ {
+ "epoch": 3.8795811518324608,
+ "grad_norm": 0.038991592824459076,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.719473838806152,
+ "step": 279
+ },
+ {
+ "epoch": 3.893542757417103,
+ "grad_norm": 0.04434404894709587,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.709481239318848,
+ "step": 280
+ },
+ {
+ "epoch": 3.9075043630017454,
+ "grad_norm": 0.047285184264183044,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.693531036376953,
+ "step": 281
+ },
+ {
+ "epoch": 3.9214659685863875,
+ "grad_norm": 0.045081138610839844,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.728959083557129,
+ "step": 282
+ },
+ {
+ "epoch": 3.9354275741710296,
+ "grad_norm": 0.025133663788437843,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.719696998596191,
+ "step": 283
+ },
+ {
+ "epoch": 3.9493891797556717,
+ "grad_norm": 0.03353596478700638,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.700235366821289,
+ "step": 284
+ },
+ {
+ "epoch": 3.9633507853403143,
+ "grad_norm": 0.035162411630153656,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.684836387634277,
+ "step": 285
+ },
+ {
+ "epoch": 3.9773123909249564,
+ "grad_norm": 0.04025450348854065,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.692097187042236,
+ "step": 286
+ },
+ {
+ "epoch": 3.9912739965095985,
+ "grad_norm": 0.04995397478342056,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.6766862869262695,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.031619708985090256,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 3.5519747734069824,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.714687705039978,
+ "eval_runtime": 59.637,
+ "eval_samples_per_second": 40.948,
+ "eval_steps_per_second": 0.654,
+ "step": 288
+ },
+ {
+ "epoch": 4.013961605584642,
+ "grad_norm": 0.036328598856925964,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.666294097900391,
+ "step": 289
+ },
+ {
+ "epoch": 4.027923211169284,
+ "grad_norm": 0.07326336950063705,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.6879167556762695,
+ "step": 290
+ },
+ {
+ "epoch": 4.041884816753926,
+ "grad_norm": 0.07947526127099991,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.674883842468262,
+ "step": 291
+ },
+ {
+ "epoch": 4.055846422338569,
+ "grad_norm": 0.04599885269999504,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.688086986541748,
+ "step": 292
+ },
+ {
+ "epoch": 4.069808027923211,
+ "grad_norm": 0.0465150885283947,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.649483680725098,
+ "step": 293
+ },
+ {
+ "epoch": 4.0837696335078535,
+ "grad_norm": 0.038542453199625015,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.653055667877197,
+ "step": 294
+ },
+ {
+ "epoch": 4.097731239092496,
+ "grad_norm": 0.03659164905548096,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.65955924987793,
+ "step": 295
+ },
+ {
+ "epoch": 4.111692844677138,
+ "grad_norm": 0.03824038803577423,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.650853157043457,
+ "step": 296
+ },
+ {
+ "epoch": 4.12565445026178,
+ "grad_norm": 0.03591204807162285,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.643499374389648,
+ "step": 297
+ },
+ {
+ "epoch": 4.139616055846423,
+ "grad_norm": 0.03269682824611664,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.657600402832031,
+ "step": 298
+ },
+ {
+ "epoch": 4.153577661431065,
+ "grad_norm": 0.03343537449836731,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.6352081298828125,
+ "step": 299
+ },
+ {
+ "epoch": 4.167539267015707,
+ "grad_norm": 0.034283801913261414,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.626757621765137,
+ "step": 300
+ },
+ {
+ "epoch": 4.181500872600349,
+ "grad_norm": 0.034614723175764084,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.6316680908203125,
+ "step": 301
+ },
+ {
+ "epoch": 4.195462478184991,
+ "grad_norm": 0.0340128131210804,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.648160934448242,
+ "step": 302
+ },
+ {
+ "epoch": 4.209424083769633,
+ "grad_norm": 0.03068280965089798,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.619738578796387,
+ "step": 303
+ },
+ {
+ "epoch": 4.223385689354275,
+ "grad_norm": 0.030269058421254158,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.609238624572754,
+ "step": 304
+ },
+ {
+ "epoch": 4.237347294938918,
+ "grad_norm": 0.03612440824508667,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.602194786071777,
+ "step": 305
+ },
+ {
+ "epoch": 4.2513089005235605,
+ "grad_norm": 0.044292811304330826,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.61159610748291,
+ "step": 306
+ },
+ {
+ "epoch": 4.265270506108203,
+ "grad_norm": 0.04449291527271271,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.6274237632751465,
+ "step": 307
+ },
+ {
+ "epoch": 4.279232111692845,
+ "grad_norm": 0.05014493316411972,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.589034080505371,
+ "step": 308
+ },
+ {
+ "epoch": 4.293193717277487,
+ "grad_norm": 0.05286206305027008,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.600770950317383,
+ "step": 309
+ },
+ {
+ "epoch": 4.307155322862129,
+ "grad_norm": 0.048459846526384354,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.5796098709106445,
+ "step": 310
+ },
+ {
+ "epoch": 4.321116928446771,
+ "grad_norm": 0.05329663306474686,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.590407371520996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335078534031414,
+ "grad_norm": 0.052595119923353195,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.583693504333496,
+ "step": 312
+ },
+ {
+ "epoch": 4.349040139616056,
+ "grad_norm": 0.05164601653814316,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.603281021118164,
+ "step": 313
+ },
+ {
+ "epoch": 4.363001745200698,
+ "grad_norm": 0.04657137766480446,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.591344833374023,
+ "step": 314
+ },
+ {
+ "epoch": 4.37696335078534,
+ "grad_norm": 0.03804599866271019,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.580578804016113,
+ "step": 315
+ },
+ {
+ "epoch": 4.390924956369982,
+ "grad_norm": 0.03835320845246315,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.597836971282959,
+ "step": 316
+ },
+ {
+ "epoch": 4.4048865619546245,
+ "grad_norm": 0.034059133380651474,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.567502021789551,
+ "step": 317
+ },
+ {
+ "epoch": 4.418848167539267,
+ "grad_norm": 0.03350386768579483,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.552022933959961,
+ "step": 318
+ },
+ {
+ "epoch": 4.43280977312391,
+ "grad_norm": 0.032574303448200226,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.559177875518799,
+ "step": 319
+ },
+ {
+ "epoch": 4.446771378708552,
+ "grad_norm": 0.030893854796886444,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.549288272857666,
+ "step": 320
+ },
+ {
+ "epoch": 4.460732984293194,
+ "grad_norm": 0.03255036100745201,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.5400390625,
+ "step": 321
+ },
+ {
+ "epoch": 4.474694589877836,
+ "grad_norm": 0.03009258769452572,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.5328474044799805,
+ "step": 322
+ },
+ {
+ "epoch": 4.488656195462478,
+ "grad_norm": 0.03178320452570915,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.5594482421875,
+ "step": 323
+ },
+ {
+ "epoch": 4.50261780104712,
+ "grad_norm": 0.03360465541481972,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.5402727127075195,
+ "step": 324
+ },
+ {
+ "epoch": 4.516579406631763,
+ "grad_norm": 0.04275049269199371,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.534474849700928,
+ "step": 325
+ },
+ {
+ "epoch": 4.530541012216405,
+ "grad_norm": 0.04140641540288925,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.5324859619140625,
+ "step": 326
+ },
+ {
+ "epoch": 4.544502617801047,
+ "grad_norm": 0.04769204929471016,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.4919352531433105,
+ "step": 327
+ },
+ {
+ "epoch": 4.558464223385689,
+ "grad_norm": 0.04402285814285278,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.520858287811279,
+ "step": 328
+ },
+ {
+ "epoch": 4.5724258289703315,
+ "grad_norm": 0.038526229560375214,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.523778438568115,
+ "step": 329
+ },
+ {
+ "epoch": 4.5863874345549736,
+ "grad_norm": 0.05422377213835716,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.517973899841309,
+ "step": 330
+ },
+ {
+ "epoch": 4.600349040139616,
+ "grad_norm": 0.07236721366643906,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.548083782196045,
+ "step": 331
+ },
+ {
+ "epoch": 4.614310645724259,
+ "grad_norm": 0.06861061602830887,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.535577774047852,
+ "step": 332
+ },
+ {
+ "epoch": 4.628272251308901,
+ "grad_norm": 0.07148340344429016,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.54155158996582,
+ "step": 333
+ },
+ {
+ "epoch": 4.642233856893543,
+ "grad_norm": 0.07838302850723267,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.552852630615234,
+ "step": 334
+ },
+ {
+ "epoch": 4.656195462478185,
+ "grad_norm": 0.10946758091449738,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.555066108703613,
+ "step": 335
+ },
+ {
+ "epoch": 4.670157068062827,
+ "grad_norm": 0.10143530368804932,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.584932804107666,
+ "step": 336
+ },
+ {
+ "epoch": 4.684118673647469,
+ "grad_norm": 0.08926606178283691,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.570462226867676,
+ "step": 337
+ },
+ {
+ "epoch": 4.698080279232112,
+ "grad_norm": 0.07060826569795609,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.5716753005981445,
+ "step": 338
+ },
+ {
+ "epoch": 4.712041884816754,
+ "grad_norm": 0.0418456569314003,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.556496620178223,
+ "step": 339
+ },
+ {
+ "epoch": 4.726003490401396,
+ "grad_norm": 0.05952411890029907,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.567801475524902,
+ "step": 340
+ },
+ {
+ "epoch": 4.739965095986038,
+ "grad_norm": 0.042999885976314545,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.537403106689453,
+ "step": 341
+ },
+ {
+ "epoch": 4.7539267015706805,
+ "grad_norm": 0.04426449537277222,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.552387237548828,
+ "step": 342
+ },
+ {
+ "epoch": 4.767888307155323,
+ "grad_norm": 0.040835145860910416,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.53051233291626,
+ "step": 343
+ },
+ {
+ "epoch": 4.781849912739965,
+ "grad_norm": 0.03819291293621063,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.531174659729004,
+ "step": 344
+ },
+ {
+ "epoch": 4.795811518324607,
+ "grad_norm": 0.030088599771261215,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.499328136444092,
+ "step": 345
+ },
+ {
+ "epoch": 4.80977312390925,
+ "grad_norm": 0.029984362423419952,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.504540920257568,
+ "step": 346
+ },
+ {
+ "epoch": 4.823734729493892,
+ "grad_norm": 0.027402915060520172,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.50899076461792,
+ "step": 347
+ },
+ {
+ "epoch": 4.837696335078534,
+ "grad_norm": 0.025096213445067406,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.480479717254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.851657940663176,
+ "grad_norm": 0.026828132569789886,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.501729965209961,
+ "step": 349
+ },
+ {
+ "epoch": 4.865619546247818,
+ "grad_norm": 0.023715168237686157,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.466860771179199,
+ "step": 350
+ },
+ {
+ "epoch": 4.879581151832461,
+ "grad_norm": 0.024503640830516815,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.491827964782715,
+ "step": 351
+ },
+ {
+ "epoch": 4.893542757417103,
+ "grad_norm": 0.020891105756163597,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.478384971618652,
+ "step": 352
+ },
+ {
+ "epoch": 4.907504363001745,
+ "grad_norm": 0.021373765543103218,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.4612135887146,
+ "step": 353
+ },
+ {
+ "epoch": 4.9214659685863875,
+ "grad_norm": 0.021773340180516243,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.4547319412231445,
+ "step": 354
+ },
+ {
+ "epoch": 4.93542757417103,
+ "grad_norm": 0.019637586548924446,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.431741714477539,
+ "step": 355
+ },
+ {
+ "epoch": 4.949389179755672,
+ "grad_norm": 0.023108696565032005,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.448612213134766,
+ "step": 356
+ },
+ {
+ "epoch": 4.963350785340314,
+ "grad_norm": 0.021935103461146355,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.447112083435059,
+ "step": 357
+ },
+ {
+ "epoch": 4.977312390924956,
+ "grad_norm": 0.023260753601789474,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.436091423034668,
+ "step": 358
+ },
+ {
+ "epoch": 4.991273996509599,
+ "grad_norm": 0.021355057135224342,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.408421993255615,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.014992325566709042,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 3.4004223346710205,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6834986805915833,
+ "eval_runtime": 60.4148,
+ "eval_samples_per_second": 40.421,
+ "eval_steps_per_second": 0.646,
+ "step": 360
+ },
+ {
+ "epoch": 5.013961605584642,
+ "grad_norm": 0.02471218630671501,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.396238803863525,
+ "step": 361
+ },
+ {
+ "epoch": 5.027923211169284,
+ "grad_norm": 0.027672411873936653,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.4048919677734375,
+ "step": 362
+ },
+ {
+ "epoch": 5.041884816753926,
+ "grad_norm": 0.031560689210891724,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.399665832519531,
+ "step": 363
+ },
+ {
+ "epoch": 5.055846422338569,
+ "grad_norm": 0.03459061309695244,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.396360397338867,
+ "step": 364
+ },
+ {
+ "epoch": 5.069808027923211,
+ "grad_norm": 0.04040960222482681,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.405309677124023,
+ "step": 365
+ },
+ {
+ "epoch": 5.0837696335078535,
+ "grad_norm": 0.05444670096039772,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.408894062042236,
+ "step": 366
+ },
+ {
+ "epoch": 5.097731239092496,
+ "grad_norm": 0.05816745012998581,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.403566360473633,
+ "step": 367
+ },
+ {
+ "epoch": 5.111692844677138,
+ "grad_norm": 0.06330069154500961,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.395961761474609,
+ "step": 368
+ },
+ {
+ "epoch": 5.12565445026178,
+ "grad_norm": 0.07458917796611786,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.400968551635742,
+ "step": 369
+ },
+ {
+ "epoch": 5.139616055846423,
+ "grad_norm": 0.06231356039643288,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.412624359130859,
+ "step": 370
+ },
+ {
+ "epoch": 5.153577661431065,
+ "grad_norm": 0.05212346464395523,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.394794464111328,
+ "step": 371
+ },
+ {
+ "epoch": 5.167539267015707,
+ "grad_norm": 0.050003375858068466,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.380125045776367,
+ "step": 372
+ },
+ {
+ "epoch": 5.181500872600349,
+ "grad_norm": 0.048393815755844116,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.390806198120117,
+ "step": 373
+ },
+ {
+ "epoch": 5.195462478184991,
+ "grad_norm": 0.04200794920325279,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.387467384338379,
+ "step": 374
+ },
+ {
+ "epoch": 5.209424083769633,
+ "grad_norm": 0.04931448772549629,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.37552547454834,
+ "step": 375
+ },
+ {
+ "epoch": 5.223385689354275,
+ "grad_norm": 0.03851857781410217,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.373126983642578,
+ "step": 376
+ },
+ {
+ "epoch": 5.237347294938918,
+ "grad_norm": 0.03458717465400696,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.346826553344727,
+ "step": 377
+ },
+ {
+ "epoch": 5.2513089005235605,
+ "grad_norm": 0.031107677146792412,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.353341102600098,
+ "step": 378
+ },
+ {
+ "epoch": 5.265270506108203,
+ "grad_norm": 0.03077058307826519,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.358076095581055,
+ "step": 379
+ },
+ {
+ "epoch": 5.279232111692845,
+ "grad_norm": 0.028170401230454445,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.354019641876221,
+ "step": 380
+ },
+ {
+ "epoch": 5.293193717277487,
+ "grad_norm": 0.02871990203857422,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.368258476257324,
+ "step": 381
+ },
+ {
+ "epoch": 5.307155322862129,
+ "grad_norm": 0.03046584501862526,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.342505931854248,
+ "step": 382
+ },
+ {
+ "epoch": 5.321116928446771,
+ "grad_norm": 0.02781563811004162,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.339411735534668,
+ "step": 383
+ },
+ {
+ "epoch": 5.335078534031414,
+ "grad_norm": 0.028825288638472557,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.341948509216309,
+ "step": 384
+ },
+ {
+ "epoch": 5.349040139616056,
+ "grad_norm": 0.026684079319238663,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.340881824493408,
+ "step": 385
+ },
+ {
+ "epoch": 5.363001745200698,
+ "grad_norm": 0.02429167740046978,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.342733383178711,
+ "step": 386
+ },
+ {
+ "epoch": 5.37696335078534,
+ "grad_norm": 0.02855239436030388,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.313878536224365,
+ "step": 387
+ },
+ {
+ "epoch": 5.390924956369982,
+ "grad_norm": 0.029651548713445663,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.312665939331055,
+ "step": 388
+ },
+ {
+ "epoch": 5.4048865619546245,
+ "grad_norm": 0.027191132307052612,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3086838722229,
+ "step": 389
+ },
+ {
+ "epoch": 5.418848167539267,
+ "grad_norm": 0.026719525456428528,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.305359840393066,
+ "step": 390
+ },
+ {
+ "epoch": 5.43280977312391,
+ "grad_norm": 0.026763100177049637,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.3164381980896,
+ "step": 391
+ },
+ {
+ "epoch": 5.446771378708552,
+ "grad_norm": 0.03125109151005745,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.3083953857421875,
+ "step": 392
+ },
+ {
+ "epoch": 5.460732984293194,
+ "grad_norm": 0.0394827201962471,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.275996208190918,
+ "step": 393
+ },
+ {
+ "epoch": 5.474694589877836,
+ "grad_norm": 0.040579475462436676,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.292995929718018,
+ "step": 394
+ },
+ {
+ "epoch": 5.488656195462478,
+ "grad_norm": 0.037223003804683685,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.298526763916016,
+ "step": 395
+ },
+ {
+ "epoch": 5.50261780104712,
+ "grad_norm": 0.04207765311002731,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.287985801696777,
+ "step": 396
+ },
+ {
+ "epoch": 5.516579406631763,
+ "grad_norm": 0.048299964517354965,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.244608402252197,
+ "step": 397
+ },
+ {
+ "epoch": 5.530541012216405,
+ "grad_norm": 0.04650307446718216,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.285453796386719,
+ "step": 398
+ },
+ {
+ "epoch": 5.544502617801047,
+ "grad_norm": 0.045339711010456085,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.312187671661377,
+ "step": 399
+ },
+ {
+ "epoch": 5.558464223385689,
+ "grad_norm": 0.041067708283662796,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.292692184448242,
+ "step": 400
+ },
+ {
+ "epoch": 5.5724258289703315,
+ "grad_norm": 0.05455026403069496,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.253364562988281,
+ "step": 401
+ },
+ {
+ "epoch": 5.5863874345549736,
+ "grad_norm": 0.06660915911197662,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.262646675109863,
+ "step": 402
+ },
+ {
+ "epoch": 5.600349040139616,
+ "grad_norm": 0.06634793430566788,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.274322986602783,
+ "step": 403
+ },
+ {
+ "epoch": 5.614310645724259,
+ "grad_norm": 0.08996330946683884,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.287951469421387,
+ "step": 404
+ },
+ {
+ "epoch": 5.628272251308901,
+ "grad_norm": 0.08526470512151718,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.3007097244262695,
+ "step": 405
+ },
+ {
+ "epoch": 5.642233856893543,
+ "grad_norm": 0.08766836673021317,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.332538604736328,
+ "step": 406
+ },
+ {
+ "epoch": 5.656195462478185,
+ "grad_norm": 0.07055048644542694,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.296169281005859,
+ "step": 407
+ },
+ {
+ "epoch": 5.670157068062827,
+ "grad_norm": 0.07520575076341629,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.309683322906494,
+ "step": 408
+ },
+ {
+ "epoch": 5.684118673647469,
+ "grad_norm": 0.06115756183862686,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.311301231384277,
+ "step": 409
+ },
+ {
+ "epoch": 5.698080279232112,
+ "grad_norm": 0.05862593650817871,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.28066349029541,
+ "step": 410
+ },
+ {
+ "epoch": 5.712041884816754,
+ "grad_norm": 0.06169601157307625,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.286840438842773,
+ "step": 411
+ },
+ {
+ "epoch": 5.726003490401396,
+ "grad_norm": 0.05270719155669212,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.302594184875488,
+ "step": 412
+ },
+ {
+ "epoch": 5.739965095986038,
+ "grad_norm": 0.03912634402513504,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.262563705444336,
+ "step": 413
+ },
+ {
+ "epoch": 5.7539267015706805,
+ "grad_norm": 0.03215811401605606,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.269310474395752,
+ "step": 414
+ },
+ {
+ "epoch": 5.767888307155323,
+ "grad_norm": 0.03236516937613487,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.268766403198242,
+ "step": 415
+ },
+ {
+ "epoch": 5.781849912739965,
+ "grad_norm": 0.027429452165961266,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.283801555633545,
+ "step": 416
+ },
+ {
+ "epoch": 5.795811518324607,
+ "grad_norm": 0.02990538813173771,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.22291374206543,
+ "step": 417
+ },
+ {
+ "epoch": 5.80977312390925,
+ "grad_norm": 0.030410146340727806,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.2281599044799805,
+ "step": 418
+ },
+ {
+ "epoch": 5.823734729493892,
+ "grad_norm": 0.030068187043070793,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.250294208526611,
+ "step": 419
+ },
+ {
+ "epoch": 5.837696335078534,
+ "grad_norm": 0.02800513058900833,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.226508617401123,
+ "step": 420
+ },
+ {
+ "epoch": 5.851657940663176,
+ "grad_norm": 0.023823332041502,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.247838020324707,
+ "step": 421
+ },
+ {
+ "epoch": 5.865619546247818,
+ "grad_norm": 0.024974249303340912,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.2392473220825195,
+ "step": 422
+ },
+ {
+ "epoch": 5.879581151832461,
+ "grad_norm": 0.023320546373724937,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.209543704986572,
+ "step": 423
+ },
+ {
+ "epoch": 5.893542757417103,
+ "grad_norm": 0.024201491847634315,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.219725608825684,
+ "step": 424
+ },
+ {
+ "epoch": 5.907504363001745,
+ "grad_norm": 0.021649666130542755,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.203141212463379,
+ "step": 425
+ },
+ {
+ "epoch": 5.9214659685863875,
+ "grad_norm": 0.0248002577573061,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.179058074951172,
+ "step": 426
+ },
+ {
+ "epoch": 5.93542757417103,
+ "grad_norm": 0.028549587354063988,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.202126502990723,
+ "step": 427
+ },
+ {
+ "epoch": 5.949389179755672,
+ "grad_norm": 0.0331057533621788,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.208107948303223,
+ "step": 428
+ },
+ {
+ "epoch": 5.963350785340314,
+ "grad_norm": 0.031843364238739014,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.184683799743652,
+ "step": 429
+ },
+ {
+ "epoch": 5.977312390924956,
+ "grad_norm": 0.024927377700805664,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.180486679077148,
+ "step": 430
+ },
+ {
+ "epoch": 5.991273996509599,
+ "grad_norm": 0.026685267686843872,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.186746597290039,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.018545351922512054,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 3.217043876647949,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.653471827507019,
+ "eval_runtime": 61.1586,
+ "eval_samples_per_second": 39.929,
+ "eval_steps_per_second": 0.638,
+ "step": 432
+ },
+ {
+ "epoch": 6.013961605584642,
+ "grad_norm": 0.025291118770837784,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.156534194946289,
+ "step": 433
+ },
+ {
+ "epoch": 6.027923211169284,
+ "grad_norm": 0.03774752840399742,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.144041538238525,
+ "step": 434
+ },
+ {
+ "epoch": 6.041884816753926,
+ "grad_norm": 0.04244692996144295,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.158862113952637,
+ "step": 435
+ },
+ {
+ "epoch": 6.055846422338569,
+ "grad_norm": 0.039782389998435974,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.166165351867676,
+ "step": 436
+ },
+ {
+ "epoch": 6.069808027923211,
+ "grad_norm": 0.05489896610379219,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.155766487121582,
+ "step": 437
+ },
+ {
+ "epoch": 6.0837696335078535,
+ "grad_norm": 0.059844326227903366,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.153733253479004,
+ "step": 438
+ },
+ {
+ "epoch": 6.097731239092496,
+ "grad_norm": 0.05539938807487488,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.130112648010254,
+ "step": 439
+ },
+ {
+ "epoch": 6.111692844677138,
+ "grad_norm": 0.05283183231949806,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.136910915374756,
+ "step": 440
+ },
+ {
+ "epoch": 6.12565445026178,
+ "grad_norm": 0.054407261312007904,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.165216445922852,
+ "step": 441
+ },
+ {
+ "epoch": 6.139616055846423,
+ "grad_norm": 0.0509788915514946,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.140888214111328,
+ "step": 442
+ },
+ {
+ "epoch": 6.153577661431065,
+ "grad_norm": 0.05018748342990875,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.13115930557251,
+ "step": 443
+ },
+ {
+ "epoch": 6.167539267015707,
+ "grad_norm": 0.047012437134981155,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.146507263183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181500872600349,
+ "grad_norm": 0.04943335801362991,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.135608673095703,
+ "step": 445
+ },
+ {
+ "epoch": 6.195462478184991,
+ "grad_norm": 0.05304260924458504,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.136394023895264,
+ "step": 446
+ },
+ {
+ "epoch": 6.209424083769633,
+ "grad_norm": 0.04554047808051109,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.124661922454834,
+ "step": 447
+ },
+ {
+ "epoch": 6.223385689354275,
+ "grad_norm": 0.040672432631254196,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.129773139953613,
+ "step": 448
+ },
+ {
+ "epoch": 6.237347294938918,
+ "grad_norm": 0.03442879393696785,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.128028392791748,
+ "step": 449
+ },
+ {
+ "epoch": 6.2513089005235605,
+ "grad_norm": 0.03927934914827347,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.092485427856445,
+ "step": 450
+ },
+ {
+ "epoch": 6.265270506108203,
+ "grad_norm": 0.029380977153778076,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.077669143676758,
+ "step": 451
+ },
+ {
+ "epoch": 6.279232111692845,
+ "grad_norm": 0.03331821784377098,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.138857841491699,
+ "step": 452
+ },
+ {
+ "epoch": 6.293193717277487,
+ "grad_norm": 0.02939118817448616,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.082937240600586,
+ "step": 453
+ },
+ {
+ "epoch": 6.307155322862129,
+ "grad_norm": 0.02730412967503071,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.093183517456055,
+ "step": 454
+ },
+ {
+ "epoch": 6.321116928446771,
+ "grad_norm": 0.02964780665934086,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.1028151512146,
+ "step": 455
+ },
+ {
+ "epoch": 6.335078534031414,
+ "grad_norm": 0.026350129395723343,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.066362380981445,
+ "step": 456
+ },
+ {
+ "epoch": 6.349040139616056,
+ "grad_norm": 0.027010295540094376,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.070037841796875,
+ "step": 457
+ },
+ {
+ "epoch": 6.363001745200698,
+ "grad_norm": 0.02906128577888012,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.107708930969238,
+ "step": 458
+ },
+ {
+ "epoch": 6.37696335078534,
+ "grad_norm": 0.028619443997740746,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.073182582855225,
+ "step": 459
+ },
+ {
+ "epoch": 6.390924956369982,
+ "grad_norm": 0.029981834813952446,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.076602935791016,
+ "step": 460
+ },
+ {
+ "epoch": 6.4048865619546245,
+ "grad_norm": 0.03255423903465271,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.063301086425781,
+ "step": 461
+ },
+ {
+ "epoch": 6.418848167539267,
+ "grad_norm": 0.03863866627216339,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.041812896728516,
+ "step": 462
+ },
+ {
+ "epoch": 6.43280977312391,
+ "grad_norm": 0.03860306739807129,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.06298828125,
+ "step": 463
+ },
+ {
+ "epoch": 6.446771378708552,
+ "grad_norm": 0.03378435596823692,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.0434370040893555,
+ "step": 464
+ },
+ {
+ "epoch": 6.460732984293194,
+ "grad_norm": 0.03276759013533592,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.036187171936035,
+ "step": 465
+ },
+ {
+ "epoch": 6.474694589877836,
+ "grad_norm": 0.04130513221025467,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.039098739624023,
+ "step": 466
+ },
+ {
+ "epoch": 6.488656195462478,
+ "grad_norm": 0.05389940366148949,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.062798976898193,
+ "step": 467
+ },
+ {
+ "epoch": 6.50261780104712,
+ "grad_norm": 0.05881704390048981,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.067028522491455,
+ "step": 468
+ },
+ {
+ "epoch": 6.516579406631763,
+ "grad_norm": 0.061897799372673035,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.056947231292725,
+ "step": 469
+ },
+ {
+ "epoch": 6.530541012216405,
+ "grad_norm": 0.06549323350191116,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.089262008666992,
+ "step": 470
+ },
+ {
+ "epoch": 6.544502617801047,
+ "grad_norm": 0.048082590103149414,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.055979251861572,
+ "step": 471
+ },
+ {
+ "epoch": 6.558464223385689,
+ "grad_norm": 0.03768027946352959,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.0660600662231445,
+ "step": 472
+ },
+ {
+ "epoch": 6.5724258289703315,
+ "grad_norm": 0.043922603130340576,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 5.036745071411133,
+ "step": 473
+ },
+ {
+ "epoch": 6.5863874345549736,
+ "grad_norm": 0.044848836958408356,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.073650360107422,
+ "step": 474
+ },
+ {
+ "epoch": 6.600349040139616,
+ "grad_norm": 0.05841026455163956,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.053045272827148,
+ "step": 475
+ },
+ {
+ "epoch": 6.614310645724259,
+ "grad_norm": 0.0650017261505127,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 5.052206039428711,
+ "step": 476
+ },
+ {
+ "epoch": 6.628272251308901,
+ "grad_norm": 0.048979002982378006,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 5.054277420043945,
+ "step": 477
+ },
+ {
+ "epoch": 6.642233856893543,
+ "grad_norm": 0.05085799843072891,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 5.063844680786133,
+ "step": 478
+ },
+ {
+ "epoch": 6.656195462478185,
+ "grad_norm": 0.04336017370223999,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 5.045474052429199,
+ "step": 479
+ },
+ {
+ "epoch": 6.670157068062827,
+ "grad_norm": 0.04113380238413811,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 5.031338214874268,
+ "step": 480
+ },
+ {
+ "epoch": 6.684118673647469,
+ "grad_norm": 0.03199538215994835,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 5.030025959014893,
+ "step": 481
+ },
+ {
+ "epoch": 6.698080279232112,
+ "grad_norm": 0.03043113648891449,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 5.033163070678711,
+ "step": 482
+ },
+ {
+ "epoch": 6.712041884816754,
+ "grad_norm": 0.031694378703832626,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 5.004739761352539,
+ "step": 483
+ },
+ {
+ "epoch": 6.726003490401396,
+ "grad_norm": 0.030243348330259323,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 5.012566089630127,
+ "step": 484
+ },
+ {
+ "epoch": 6.739965095986038,
+ "grad_norm": 0.02937416359782219,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 5.039803504943848,
+ "step": 485
+ },
+ {
+ "epoch": 6.7539267015706805,
+ "grad_norm": 0.029616443440318108,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 5.026117324829102,
+ "step": 486
+ },
+ {
+ "epoch": 6.767888307155323,
+ "grad_norm": 0.029567131772637367,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 5.006569862365723,
+ "step": 487
+ },
+ {
+ "epoch": 6.781849912739965,
+ "grad_norm": 0.029443370178341866,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.9941487312316895,
+ "step": 488
+ },
+ {
+ "epoch": 6.795811518324607,
+ "grad_norm": 0.03434352949261665,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.976431846618652,
+ "step": 489
+ },
+ {
+ "epoch": 6.80977312390925,
+ "grad_norm": 0.03845234215259552,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.993168354034424,
+ "step": 490
+ },
+ {
+ "epoch": 6.823734729493892,
+ "grad_norm": 0.05058208107948303,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.030162334442139,
+ "step": 491
+ },
+ {
+ "epoch": 6.837696335078534,
+ "grad_norm": 0.05961710214614868,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 5.007053375244141,
+ "step": 492
+ },
+ {
+ "epoch": 6.851657940663176,
+ "grad_norm": 0.056259628385305405,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 5.020356178283691,
+ "step": 493
+ },
+ {
+ "epoch": 6.865619546247818,
+ "grad_norm": 0.048628658056259155,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 5.037652969360352,
+ "step": 494
+ },
+ {
+ "epoch": 6.879581151832461,
+ "grad_norm": 0.040836453437805176,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.990390777587891,
+ "step": 495
+ },
+ {
+ "epoch": 6.893542757417103,
+ "grad_norm": 0.04112066701054573,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 5.017822742462158,
+ "step": 496
+ },
+ {
+ "epoch": 6.907504363001745,
+ "grad_norm": 0.03956829756498337,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.993971824645996,
+ "step": 497
+ },
+ {
+ "epoch": 6.9214659685863875,
+ "grad_norm": 0.03887917473912239,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.982889175415039,
+ "step": 498
+ },
+ {
+ "epoch": 6.93542757417103,
+ "grad_norm": 0.031012587249279022,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.961717128753662,
+ "step": 499
+ },
+ {
+ "epoch": 6.949389179755672,
+ "grad_norm": 0.03443489596247673,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.9904046058654785,
+ "step": 500
+ },
+ {
+ "epoch": 6.963350785340314,
+ "grad_norm": 0.035175010561943054,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.983619213104248,
+ "step": 501
+ },
+ {
+ "epoch": 6.977312390924956,
+ "grad_norm": 0.031724050641059875,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 5.000522613525391,
+ "step": 502
+ },
+ {
+ "epoch": 6.991273996509599,
+ "grad_norm": 0.032037995755672455,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.947064399719238,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.01883271522819996,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 3.112166166305542,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6296830177307129,
+ "eval_runtime": 59.5495,
+ "eval_samples_per_second": 41.008,
+ "eval_steps_per_second": 0.655,
+ "step": 504
+ },
+ {
+ "epoch": 7.013961605584642,
+ "grad_norm": 0.031091563403606415,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.919820785522461,
+ "step": 505
+ },
+ {
+ "epoch": 7.027923211169284,
+ "grad_norm": 0.04074421525001526,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.945765972137451,
+ "step": 506
+ },
+ {
+ "epoch": 7.041884816753926,
+ "grad_norm": 0.04225052148103714,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.935208797454834,
+ "step": 507
+ },
+ {
+ "epoch": 7.055846422338569,
+ "grad_norm": 0.0399051196873188,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.951314926147461,
+ "step": 508
+ },
+ {
+ "epoch": 7.069808027923211,
+ "grad_norm": 0.04785890504717827,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.92843770980835,
+ "step": 509
+ },
+ {
+ "epoch": 7.0837696335078535,
+ "grad_norm": 0.04922301694750786,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.901020050048828,
+ "step": 510
+ },
+ {
+ "epoch": 7.097731239092496,
+ "grad_norm": 0.05307295173406601,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.961151599884033,
+ "step": 511
+ },
+ {
+ "epoch": 7.111692844677138,
+ "grad_norm": 0.07003362476825714,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.9427876472473145,
+ "step": 512
+ },
+ {
+ "epoch": 7.12565445026178,
+ "grad_norm": 0.06498776376247406,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.947838306427002,
+ "step": 513
+ },
+ {
+ "epoch": 7.139616055846423,
+ "grad_norm": 0.058723628520965576,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.938859939575195,
+ "step": 514
+ },
+ {
+ "epoch": 7.153577661431065,
+ "grad_norm": 0.06590746343135834,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.962221145629883,
+ "step": 515
+ },
+ {
+ "epoch": 7.167539267015707,
+ "grad_norm": 0.052714601159095764,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.931011199951172,
+ "step": 516
+ },
+ {
+ "epoch": 7.181500872600349,
+ "grad_norm": 0.04296804964542389,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.93586540222168,
+ "step": 517
+ },
+ {
+ "epoch": 7.195462478184991,
+ "grad_norm": 0.03699173778295517,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.910763263702393,
+ "step": 518
+ },
+ {
+ "epoch": 7.209424083769633,
+ "grad_norm": 0.0317913293838501,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.9256439208984375,
+ "step": 519
+ },
+ {
+ "epoch": 7.223385689354275,
+ "grad_norm": 0.030313925817608833,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.938329696655273,
+ "step": 520
+ },
+ {
+ "epoch": 7.237347294938918,
+ "grad_norm": 0.02879229187965393,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.924966812133789,
+ "step": 521
+ },
+ {
+ "epoch": 7.2513089005235605,
+ "grad_norm": 0.02689451351761818,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.9079742431640625,
+ "step": 522
+ },
+ {
+ "epoch": 7.265270506108203,
+ "grad_norm": 0.028403930366039276,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.913238525390625,
+ "step": 523
+ },
+ {
+ "epoch": 7.279232111692845,
+ "grad_norm": 0.025912854820489883,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.896944046020508,
+ "step": 524
+ },
+ {
+ "epoch": 7.293193717277487,
+ "grad_norm": 0.024899760261178017,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.884920120239258,
+ "step": 525
+ },
+ {
+ "epoch": 7.307155322862129,
+ "grad_norm": 0.026912573724985123,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.9181060791015625,
+ "step": 526
+ },
+ {
+ "epoch": 7.321116928446771,
+ "grad_norm": 0.029539721086621284,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.914867401123047,
+ "step": 527
+ },
+ {
+ "epoch": 7.335078534031414,
+ "grad_norm": 0.028527872636914253,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.907217979431152,
+ "step": 528
+ },
+ {
+ "epoch": 7.349040139616056,
+ "grad_norm": 0.025879016146063805,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.905783176422119,
+ "step": 529
+ },
+ {
+ "epoch": 7.363001745200698,
+ "grad_norm": 0.026395199820399284,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.880505561828613,
+ "step": 530
+ },
+ {
+ "epoch": 7.37696335078534,
+ "grad_norm": 0.025072351098060608,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.870449066162109,
+ "step": 531
+ },
+ {
+ "epoch": 7.390924956369982,
+ "grad_norm": 0.02751019410789013,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.888952255249023,
+ "step": 532
+ },
+ {
+ "epoch": 7.4048865619546245,
+ "grad_norm": 0.025810658931732178,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.894552230834961,
+ "step": 533
+ },
+ {
+ "epoch": 7.418848167539267,
+ "grad_norm": 0.02399870567023754,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.8768229484558105,
+ "step": 534
+ },
+ {
+ "epoch": 7.43280977312391,
+ "grad_norm": 0.024639368057250977,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.9030914306640625,
+ "step": 535
+ },
+ {
+ "epoch": 7.446771378708552,
+ "grad_norm": 0.02543291077017784,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.86467170715332,
+ "step": 536
+ },
+ {
+ "epoch": 7.460732984293194,
+ "grad_norm": 0.02702544629573822,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.888283729553223,
+ "step": 537
+ },
+ {
+ "epoch": 7.474694589877836,
+ "grad_norm": 0.028904743492603302,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.866081237792969,
+ "step": 538
+ },
+ {
+ "epoch": 7.488656195462478,
+ "grad_norm": 0.03055848367512226,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.890739440917969,
+ "step": 539
+ },
+ {
+ "epoch": 7.50261780104712,
+ "grad_norm": 0.03410084918141365,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.862918853759766,
+ "step": 540
+ },
+ {
+ "epoch": 7.516579406631763,
+ "grad_norm": 0.03738495334982872,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.864728927612305,
+ "step": 541
+ },
+ {
+ "epoch": 7.530541012216405,
+ "grad_norm": 0.037646882236003876,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.865185737609863,
+ "step": 542
+ },
+ {
+ "epoch": 7.544502617801047,
+ "grad_norm": 0.03418618440628052,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.866114139556885,
+ "step": 543
+ },
+ {
+ "epoch": 7.558464223385689,
+ "grad_norm": 0.036924004554748535,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.883186340332031,
+ "step": 544
+ },
+ {
+ "epoch": 7.5724258289703315,
+ "grad_norm": 0.03590598329901695,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.863725662231445,
+ "step": 545
+ },
+ {
+ "epoch": 7.5863874345549736,
+ "grad_norm": 0.037363454699516296,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.841030120849609,
+ "step": 546
+ },
+ {
+ "epoch": 7.600349040139616,
+ "grad_norm": 0.046537864953279495,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.873837471008301,
+ "step": 547
+ },
+ {
+ "epoch": 7.614310645724259,
+ "grad_norm": 0.057226769626140594,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.865679740905762,
+ "step": 548
+ },
+ {
+ "epoch": 7.628272251308901,
+ "grad_norm": 0.06175961345434189,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.888645648956299,
+ "step": 549
+ },
+ {
+ "epoch": 7.642233856893543,
+ "grad_norm": 0.05063694715499878,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.877828598022461,
+ "step": 550
+ },
+ {
+ "epoch": 7.656195462478185,
+ "grad_norm": 0.03950628638267517,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.88637638092041,
+ "step": 551
+ },
+ {
+ "epoch": 7.670157068062827,
+ "grad_norm": 0.04183337092399597,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.8597917556762695,
+ "step": 552
+ },
+ {
+ "epoch": 7.684118673647469,
+ "grad_norm": 0.04642999917268753,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.880380153656006,
+ "step": 553
+ },
+ {
+ "epoch": 7.698080279232112,
+ "grad_norm": 0.05333106964826584,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.853932857513428,
+ "step": 554
+ },
+ {
+ "epoch": 7.712041884816754,
+ "grad_norm": 0.04764975979924202,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.870235443115234,
+ "step": 555
+ },
+ {
+ "epoch": 7.726003490401396,
+ "grad_norm": 0.043072231113910675,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.822697639465332,
+ "step": 556
+ },
+ {
+ "epoch": 7.739965095986038,
+ "grad_norm": 0.03913857042789459,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.849332809448242,
+ "step": 557
+ },
+ {
+ "epoch": 7.7539267015706805,
+ "grad_norm": 0.034836407750844955,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.842280387878418,
+ "step": 558
+ },
+ {
+ "epoch": 7.767888307155323,
+ "grad_norm": 0.033344294875860214,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.837811470031738,
+ "step": 559
+ },
+ {
+ "epoch": 7.781849912739965,
+ "grad_norm": 0.03342701122164726,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.876040458679199,
+ "step": 560
+ },
+ {
+ "epoch": 7.795811518324607,
+ "grad_norm": 0.034638747572898865,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.830340385437012,
+ "step": 561
+ },
+ {
+ "epoch": 7.80977312390925,
+ "grad_norm": 0.03358135372400284,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.828607559204102,
+ "step": 562
+ },
+ {
+ "epoch": 7.823734729493892,
+ "grad_norm": 0.036001041531562805,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.827905654907227,
+ "step": 563
+ },
+ {
+ "epoch": 7.837696335078534,
+ "grad_norm": 0.038079630583524704,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.866955280303955,
+ "step": 564
+ },
+ {
+ "epoch": 7.851657940663176,
+ "grad_norm": 0.03728470206260681,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.828039169311523,
+ "step": 565
+ },
+ {
+ "epoch": 7.865619546247818,
+ "grad_norm": 0.03920375928282738,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.841631889343262,
+ "step": 566
+ },
+ {
+ "epoch": 7.879581151832461,
+ "grad_norm": 0.032353926450014114,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.861110210418701,
+ "step": 567
+ },
+ {
+ "epoch": 7.893542757417103,
+ "grad_norm": 0.03241422027349472,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.818441390991211,
+ "step": 568
+ },
+ {
+ "epoch": 7.907504363001745,
+ "grad_norm": 0.03879080340266228,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.823749542236328,
+ "step": 569
+ },
+ {
+ "epoch": 7.9214659685863875,
+ "grad_norm": 0.03486498072743416,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.830697059631348,
+ "step": 570
+ },
+ {
+ "epoch": 7.93542757417103,
+ "grad_norm": 0.03279713913798332,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.815736770629883,
+ "step": 571
+ },
+ {
+ "epoch": 7.949389179755672,
+ "grad_norm": 0.03313925117254257,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.8364033699035645,
+ "step": 572
+ },
+ {
+ "epoch": 7.963350785340314,
+ "grad_norm": 0.03405062481760979,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.816010475158691,
+ "step": 573
+ },
+ {
+ "epoch": 7.977312390924956,
+ "grad_norm": 0.03259236365556717,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.813413619995117,
+ "step": 574
+ },
+ {
+ "epoch": 7.991273996509599,
+ "grad_norm": 0.030408985912799835,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.833840370178223,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.022680889815092087,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 2.99565052986145,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6126086115837097,
+ "eval_runtime": 60.2662,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 576
+ },
+ {
+ "epoch": 8.013961605584642,
+ "grad_norm": 0.029732126742601395,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.7816314697265625,
+ "step": 577
+ },
+ {
+ "epoch": 8.027923211169284,
+ "grad_norm": 0.04357453063130379,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.7575225830078125,
+ "step": 578
+ },
+ {
+ "epoch": 8.041884816753926,
+ "grad_norm": 0.051903221756219864,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.802519798278809,
+ "step": 579
+ },
+ {
+ "epoch": 8.055846422338568,
+ "grad_norm": 0.0539076030254364,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.799004554748535,
+ "step": 580
+ },
+ {
+ "epoch": 8.06980802792321,
+ "grad_norm": 0.043642807751894,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.7896528244018555,
+ "step": 581
+ },
+ {
+ "epoch": 8.083769633507853,
+ "grad_norm": 0.0469554178416729,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.765430927276611,
+ "step": 582
+ },
+ {
+ "epoch": 8.097731239092496,
+ "grad_norm": 0.042060062289237976,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.784132957458496,
+ "step": 583
+ },
+ {
+ "epoch": 8.111692844677139,
+ "grad_norm": 0.035101912915706635,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.767234802246094,
+ "step": 584
+ },
+ {
+ "epoch": 8.12565445026178,
+ "grad_norm": 0.042576879262924194,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.762052536010742,
+ "step": 585
+ },
+ {
+ "epoch": 8.139616055846423,
+ "grad_norm": 0.04231419041752815,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.752150535583496,
+ "step": 586
+ },
+ {
+ "epoch": 8.153577661431065,
+ "grad_norm": 0.046912599354982376,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.771881103515625,
+ "step": 587
+ },
+ {
+ "epoch": 8.167539267015707,
+ "grad_norm": 0.051271144300699234,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.773284912109375,
+ "step": 588
+ },
+ {
+ "epoch": 8.181500872600349,
+ "grad_norm": 0.04655958339571953,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.769917964935303,
+ "step": 589
+ },
+ {
+ "epoch": 8.195462478184991,
+ "grad_norm": 0.04463687539100647,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.77938175201416,
+ "step": 590
+ },
+ {
+ "epoch": 8.209424083769633,
+ "grad_norm": 0.047959886491298676,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.785962104797363,
+ "step": 591
+ },
+ {
+ "epoch": 8.223385689354275,
+ "grad_norm": 0.04414986073970795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.786281585693359,
+ "step": 592
+ },
+ {
+ "epoch": 8.237347294938917,
+ "grad_norm": 0.044150348752737045,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.777070045471191,
+ "step": 593
+ },
+ {
+ "epoch": 8.25130890052356,
+ "grad_norm": 0.037002332508563995,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.77547550201416,
+ "step": 594
+ },
+ {
+ "epoch": 8.265270506108202,
+ "grad_norm": 0.03417367860674858,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.782101154327393,
+ "step": 595
+ },
+ {
+ "epoch": 8.279232111692846,
+ "grad_norm": 0.03069210797548294,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.745595455169678,
+ "step": 596
+ },
+ {
+ "epoch": 8.293193717277488,
+ "grad_norm": 0.03228718787431717,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.7663655281066895,
+ "step": 597
+ },
+ {
+ "epoch": 8.30715532286213,
+ "grad_norm": 0.028603695333003998,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.767274856567383,
+ "step": 598
+ },
+ {
+ "epoch": 8.321116928446772,
+ "grad_norm": 0.03098377026617527,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.7853684425354,
+ "step": 599
+ },
+ {
+ "epoch": 8.335078534031414,
+ "grad_norm": 0.029242439195513725,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.738029956817627,
+ "step": 600
+ },
+ {
+ "epoch": 8.349040139616056,
+ "grad_norm": 0.02751081995666027,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.76801872253418,
+ "step": 601
+ },
+ {
+ "epoch": 8.363001745200698,
+ "grad_norm": 0.02838125079870224,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.743884086608887,
+ "step": 602
+ },
+ {
+ "epoch": 8.37696335078534,
+ "grad_norm": 0.029203062877058983,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.746668815612793,
+ "step": 603
+ },
+ {
+ "epoch": 8.390924956369982,
+ "grad_norm": 0.02679944597184658,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.760764122009277,
+ "step": 604
+ },
+ {
+ "epoch": 8.404886561954624,
+ "grad_norm": 0.028955848887562752,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.7353668212890625,
+ "step": 605
+ },
+ {
+ "epoch": 8.418848167539267,
+ "grad_norm": 0.02880851738154888,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.766837120056152,
+ "step": 606
+ },
+ {
+ "epoch": 8.432809773123909,
+ "grad_norm": 0.02495497092604637,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.78959321975708,
+ "step": 607
+ },
+ {
+ "epoch": 8.44677137870855,
+ "grad_norm": 0.026064753532409668,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.736433029174805,
+ "step": 608
+ },
+ {
+ "epoch": 8.460732984293193,
+ "grad_norm": 0.027566388249397278,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.782382965087891,
+ "step": 609
+ },
+ {
+ "epoch": 8.474694589877837,
+ "grad_norm": 0.02967359498143196,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.735286712646484,
+ "step": 610
+ },
+ {
+ "epoch": 8.488656195462479,
+ "grad_norm": 0.03079676441848278,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.725196838378906,
+ "step": 611
+ },
+ {
+ "epoch": 8.502617801047121,
+ "grad_norm": 0.032531846314668655,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.719333648681641,
+ "step": 612
+ },
+ {
+ "epoch": 8.516579406631763,
+ "grad_norm": 0.03120352514088154,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.723845958709717,
+ "step": 613
+ },
+ {
+ "epoch": 8.530541012216405,
+ "grad_norm": 0.03328424692153931,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.7669782638549805,
+ "step": 614
+ },
+ {
+ "epoch": 8.544502617801047,
+ "grad_norm": 0.035601649433374405,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.714925765991211,
+ "step": 615
+ },
+ {
+ "epoch": 8.55846422338569,
+ "grad_norm": 0.03675796464085579,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.715317726135254,
+ "step": 616
+ },
+ {
+ "epoch": 8.572425828970331,
+ "grad_norm": 0.03701988235116005,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.724606990814209,
+ "step": 617
+ },
+ {
+ "epoch": 8.586387434554974,
+ "grad_norm": 0.036719705909490585,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.746044635772705,
+ "step": 618
+ },
+ {
+ "epoch": 8.600349040139616,
+ "grad_norm": 0.034756604582071304,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.74885368347168,
+ "step": 619
+ },
+ {
+ "epoch": 8.614310645724258,
+ "grad_norm": 0.03237922117114067,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.721858501434326,
+ "step": 620
+ },
+ {
+ "epoch": 8.6282722513089,
+ "grad_norm": 0.028420142829418182,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.733814239501953,
+ "step": 621
+ },
+ {
+ "epoch": 8.642233856893542,
+ "grad_norm": 0.026203038170933723,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.726088523864746,
+ "step": 622
+ },
+ {
+ "epoch": 8.656195462478184,
+ "grad_norm": 0.02627231739461422,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.727700233459473,
+ "step": 623
+ },
+ {
+ "epoch": 8.670157068062828,
+ "grad_norm": 0.02920261025428772,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.731454849243164,
+ "step": 624
+ },
+ {
+ "epoch": 8.68411867364747,
+ "grad_norm": 0.02792907878756523,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.734154224395752,
+ "step": 625
+ },
+ {
+ "epoch": 8.698080279232112,
+ "grad_norm": 0.026770325377583504,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.715337753295898,
+ "step": 626
+ },
+ {
+ "epoch": 8.712041884816754,
+ "grad_norm": 0.027346540242433548,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.716512680053711,
+ "step": 627
+ },
+ {
+ "epoch": 8.726003490401396,
+ "grad_norm": 0.031965773552656174,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.7072038650512695,
+ "step": 628
+ },
+ {
+ "epoch": 8.739965095986038,
+ "grad_norm": 0.03708025813102722,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.749785900115967,
+ "step": 629
+ },
+ {
+ "epoch": 8.75392670157068,
+ "grad_norm": 0.04846490919589996,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.735586166381836,
+ "step": 630
+ },
+ {
+ "epoch": 8.767888307155323,
+ "grad_norm": 0.053352758288383484,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.720524787902832,
+ "step": 631
+ },
+ {
+ "epoch": 8.781849912739965,
+ "grad_norm": 0.040559202432632446,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.719901084899902,
+ "step": 632
+ },
+ {
+ "epoch": 8.795811518324607,
+ "grad_norm": 0.03333032503724098,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.708139419555664,
+ "step": 633
+ },
+ {
+ "epoch": 8.809773123909249,
+ "grad_norm": 0.030532941222190857,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.728184223175049,
+ "step": 634
+ },
+ {
+ "epoch": 8.823734729493891,
+ "grad_norm": 0.03337649628520012,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.726020336151123,
+ "step": 635
+ },
+ {
+ "epoch": 8.837696335078533,
+ "grad_norm": 0.03352045267820358,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.7196855545043945,
+ "step": 636
+ },
+ {
+ "epoch": 8.851657940663177,
+ "grad_norm": 0.03662872314453125,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.732823371887207,
+ "step": 637
+ },
+ {
+ "epoch": 8.86561954624782,
+ "grad_norm": 0.037085745483636856,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.728443145751953,
+ "step": 638
+ },
+ {
+ "epoch": 8.879581151832461,
+ "grad_norm": 0.036635689437389374,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.713988780975342,
+ "step": 639
+ },
+ {
+ "epoch": 8.893542757417103,
+ "grad_norm": 0.03719659522175789,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.715950965881348,
+ "step": 640
+ },
+ {
+ "epoch": 8.907504363001745,
+ "grad_norm": 0.03719159588217735,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.7124924659729,
+ "step": 641
+ },
+ {
+ "epoch": 8.921465968586388,
+ "grad_norm": 0.04275721684098244,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.708404064178467,
+ "step": 642
+ },
+ {
+ "epoch": 8.93542757417103,
+ "grad_norm": 0.04204598814249039,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.715372085571289,
+ "step": 643
+ },
+ {
+ "epoch": 8.949389179755672,
+ "grad_norm": 0.03938259929418564,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.729745864868164,
+ "step": 644
+ },
+ {
+ "epoch": 8.963350785340314,
+ "grad_norm": 0.036226850003004074,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.733749866485596,
+ "step": 645
+ },
+ {
+ "epoch": 8.977312390924956,
+ "grad_norm": 0.03542874753475189,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.706825256347656,
+ "step": 646
+ },
+ {
+ "epoch": 8.991273996509598,
+ "grad_norm": 0.029055381193757057,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.718567848205566,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.01990380510687828,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 2.943183183670044,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.601699709892273,
+ "eval_runtime": 60.0221,
+ "eval_samples_per_second": 40.685,
+ "eval_steps_per_second": 0.65,
+ "step": 648
+ },
+ {
+ "epoch": 9.013961605584642,
+ "grad_norm": 0.030904127284884453,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.660327911376953,
+ "step": 649
+ },
+ {
+ "epoch": 9.027923211169284,
+ "grad_norm": 0.0441313236951828,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.701565742492676,
+ "step": 650
+ },
+ {
+ "epoch": 9.041884816753926,
+ "grad_norm": 0.047547727823257446,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.666440486907959,
+ "step": 651
+ },
+ {
+ "epoch": 9.055846422338568,
+ "grad_norm": 0.04549463838338852,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.660548210144043,
+ "step": 652
+ },
+ {
+ "epoch": 9.06980802792321,
+ "grad_norm": 0.050346389412879944,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.640229225158691,
+ "step": 653
+ },
+ {
+ "epoch": 9.083769633507853,
+ "grad_norm": 0.058505453169345856,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.673881530761719,
+ "step": 654
+ },
+ {
+ "epoch": 9.097731239092496,
+ "grad_norm": 0.051989227533340454,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.693943977355957,
+ "step": 655
+ },
+ {
+ "epoch": 9.111692844677139,
+ "grad_norm": 0.052564945071935654,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.644315242767334,
+ "step": 656
+ },
+ {
+ "epoch": 9.12565445026178,
+ "grad_norm": 0.05647841840982437,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.681768417358398,
+ "step": 657
+ },
+ {
+ "epoch": 9.139616055846423,
+ "grad_norm": 0.058640867471694946,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.676457405090332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153577661431065,
+ "grad_norm": 0.05386025458574295,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.659588813781738,
+ "step": 659
+ },
+ {
+ "epoch": 9.167539267015707,
+ "grad_norm": 0.048699021339416504,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.680031776428223,
+ "step": 660
+ },
+ {
+ "epoch": 9.181500872600349,
+ "grad_norm": 0.04374176263809204,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.672671318054199,
+ "step": 661
+ },
+ {
+ "epoch": 9.195462478184991,
+ "grad_norm": 0.042337436228990555,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.673431873321533,
+ "step": 662
+ },
+ {
+ "epoch": 9.209424083769633,
+ "grad_norm": 0.04564661160111427,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.632595062255859,
+ "step": 663
+ },
+ {
+ "epoch": 9.223385689354275,
+ "grad_norm": 0.04404835030436516,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.678891181945801,
+ "step": 664
+ },
+ {
+ "epoch": 9.237347294938917,
+ "grad_norm": 0.040642790496349335,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.633418560028076,
+ "step": 665
+ },
+ {
+ "epoch": 9.25130890052356,
+ "grad_norm": 0.036492060869932175,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.670109748840332,
+ "step": 666
+ },
+ {
+ "epoch": 9.265270506108202,
+ "grad_norm": 0.03735879808664322,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.608570098876953,
+ "step": 667
+ },
+ {
+ "epoch": 9.279232111692846,
+ "grad_norm": 0.034563641995191574,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.632505416870117,
+ "step": 668
+ },
+ {
+ "epoch": 9.293193717277488,
+ "grad_norm": 0.03522484749555588,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.6475629806518555,
+ "step": 669
+ },
+ {
+ "epoch": 9.30715532286213,
+ "grad_norm": 0.03216472268104553,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.627873420715332,
+ "step": 670
+ },
+ {
+ "epoch": 9.321116928446772,
+ "grad_norm": 0.029155759140849113,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.652706146240234,
+ "step": 671
+ },
+ {
+ "epoch": 9.335078534031414,
+ "grad_norm": 0.031060541048645973,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.644545078277588,
+ "step": 672
+ },
+ {
+ "epoch": 9.349040139616056,
+ "grad_norm": 0.034979186952114105,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.642645359039307,
+ "step": 673
+ },
+ {
+ "epoch": 9.363001745200698,
+ "grad_norm": 0.035258810967206955,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.647383689880371,
+ "step": 674
+ },
+ {
+ "epoch": 9.37696335078534,
+ "grad_norm": 0.031552575528621674,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.667779445648193,
+ "step": 675
+ },
+ {
+ "epoch": 9.390924956369982,
+ "grad_norm": 0.03193951025605202,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.63625431060791,
+ "step": 676
+ },
+ {
+ "epoch": 9.404886561954624,
+ "grad_norm": 0.029811711981892586,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.658000946044922,
+ "step": 677
+ },
+ {
+ "epoch": 9.418848167539267,
+ "grad_norm": 0.028464138507843018,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.6545515060424805,
+ "step": 678
+ },
+ {
+ "epoch": 9.432809773123909,
+ "grad_norm": 0.0315735749900341,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.641153335571289,
+ "step": 679
+ },
+ {
+ "epoch": 9.44677137870855,
+ "grad_norm": 0.033412106335163116,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.626931667327881,
+ "step": 680
+ },
+ {
+ "epoch": 9.460732984293193,
+ "grad_norm": 0.03757626190781593,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.671254634857178,
+ "step": 681
+ },
+ {
+ "epoch": 9.474694589877837,
+ "grad_norm": 0.03910793736577034,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.651954174041748,
+ "step": 682
+ },
+ {
+ "epoch": 9.488656195462479,
+ "grad_norm": 0.04015549272298813,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.613324165344238,
+ "step": 683
+ },
+ {
+ "epoch": 9.502617801047121,
+ "grad_norm": 0.040689144283533096,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.639243125915527,
+ "step": 684
+ },
+ {
+ "epoch": 9.516579406631763,
+ "grad_norm": 0.04075361415743828,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.649250030517578,
+ "step": 685
+ },
+ {
+ "epoch": 9.530541012216405,
+ "grad_norm": 0.03914607688784599,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.66444730758667,
+ "step": 686
+ },
+ {
+ "epoch": 9.544502617801047,
+ "grad_norm": 0.033253733068704605,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.647335052490234,
+ "step": 687
+ },
+ {
+ "epoch": 9.55846422338569,
+ "grad_norm": 0.03179008886218071,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.6727800369262695,
+ "step": 688
+ },
+ {
+ "epoch": 9.572425828970331,
+ "grad_norm": 0.03477025404572487,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.655778884887695,
+ "step": 689
+ },
+ {
+ "epoch": 9.586387434554974,
+ "grad_norm": 0.030094800516963005,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.680851936340332,
+ "step": 690
+ },
+ {
+ "epoch": 9.600349040139616,
+ "grad_norm": 0.030153103172779083,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.654638290405273,
+ "step": 691
+ },
+ {
+ "epoch": 9.614310645724258,
+ "grad_norm": 0.029009278863668442,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.625916481018066,
+ "step": 692
+ },
+ {
+ "epoch": 9.6282722513089,
+ "grad_norm": 0.027248440310359,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.656370162963867,
+ "step": 693
+ },
+ {
+ "epoch": 9.642233856893542,
+ "grad_norm": 0.027727462351322174,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.631473541259766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656195462478184,
+ "grad_norm": 0.026710307225584984,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.636524200439453,
+ "step": 695
+ },
+ {
+ "epoch": 9.670157068062828,
+ "grad_norm": 0.02726181596517563,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.629701614379883,
+ "step": 696
+ },
+ {
+ "epoch": 9.68411867364747,
+ "grad_norm": 0.027146652340888977,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.640240669250488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698080279232112,
+ "grad_norm": 0.027165887877345085,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.64462947845459,
+ "step": 698
+ },
+ {
+ "epoch": 9.712041884816754,
+ "grad_norm": 0.03025982715189457,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.6271820068359375,
+ "step": 699
+ },
+ {
+ "epoch": 9.726003490401396,
+ "grad_norm": 0.02837025746703148,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.629697799682617,
+ "step": 700
+ },
+ {
+ "epoch": 9.739965095986038,
+ "grad_norm": 0.026777995750308037,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.646146774291992,
+ "step": 701
+ },
+ {
+ "epoch": 9.75392670157068,
+ "grad_norm": 0.024626802653074265,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.641671657562256,
+ "step": 702
+ },
+ {
+ "epoch": 9.767888307155323,
+ "grad_norm": 0.025329895317554474,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.626345157623291,
+ "step": 703
+ },
+ {
+ "epoch": 9.781849912739965,
+ "grad_norm": 0.027804844081401825,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.621457099914551,
+ "step": 704
+ },
+ {
+ "epoch": 9.795811518324607,
+ "grad_norm": 0.028763437643647194,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.6124114990234375,
+ "step": 705
+ },
+ {
+ "epoch": 9.809773123909249,
+ "grad_norm": 0.030769916251301765,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.604399681091309,
+ "step": 706
+ },
+ {
+ "epoch": 9.823734729493891,
+ "grad_norm": 0.03060949593782425,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.640804290771484,
+ "step": 707
+ },
+ {
+ "epoch": 9.837696335078533,
+ "grad_norm": 0.029832778498530388,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.593376159667969,
+ "step": 708
+ },
+ {
+ "epoch": 9.851657940663177,
+ "grad_norm": 0.02952776849269867,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.619533538818359,
+ "step": 709
+ },
+ {
+ "epoch": 9.86561954624782,
+ "grad_norm": 0.02894212119281292,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.640516757965088,
+ "step": 710
+ },
+ {
+ "epoch": 9.879581151832461,
+ "grad_norm": 0.026435624808073044,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.6342997550964355,
+ "step": 711
+ },
+ {
+ "epoch": 9.893542757417103,
+ "grad_norm": 0.02591666206717491,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.630884170532227,
+ "step": 712
+ },
+ {
+ "epoch": 9.907504363001745,
+ "grad_norm": 0.02702309936285019,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.632085800170898,
+ "step": 713
+ },
+ {
+ "epoch": 9.921465968586388,
+ "grad_norm": 0.026766657829284668,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.6231231689453125,
+ "step": 714
+ },
+ {
+ "epoch": 9.93542757417103,
+ "grad_norm": 0.026666559278964996,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.601619243621826,
+ "step": 715
+ },
+ {
+ "epoch": 9.949389179755672,
+ "grad_norm": 0.026801064610481262,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.641724109649658,
+ "step": 716
+ },
+ {
+ "epoch": 9.963350785340314,
+ "grad_norm": 0.03152798116207123,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.642511367797852,
+ "step": 717
+ },
+ {
+ "epoch": 9.977312390924956,
+ "grad_norm": 0.035406943410634995,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.635779857635498,
+ "step": 718
+ },
+ {
+ "epoch": 9.991273996509598,
+ "grad_norm": 0.037504684180021286,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.629463195800781,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02263174019753933,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 2.876767635345459,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.5942904949188232,
+ "eval_runtime": 60.2086,
+ "eval_samples_per_second": 40.559,
+ "eval_steps_per_second": 0.648,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03707852587103844,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.5765533447265625,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.05454680323600769,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.576363563537598,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.04832114651799202,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.586915969848633,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.050469402223825455,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.5265092849731445,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0622565895318985,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.592833518981934,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0640949010848999,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.5855560302734375,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.06502196192741394,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.591236114501953,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.06167149916291237,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.56747579574585,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.06747374683618546,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.60452938079834,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.07210271060466766,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.561588764190674,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.0733448714017868,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.610766410827637,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.0731528028845787,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.589570045471191,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.0822535902261734,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.631077289581299,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.06921929866075516,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.628293991088867,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.07443586736917496,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.627321243286133,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.07153813540935516,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.641068935394287,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.069196417927742,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.615875720977783,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06483697891235352,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.6165313720703125,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.05482150986790657,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.599890232086182,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.05364019051194191,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.61989164352417,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.053216829895973206,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.564305305480957,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05740506947040558,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.582683563232422,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.05587657541036606,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.61264705657959,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.050252825021743774,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.648050308227539,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.050955042243003845,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.615367889404297,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04903746768832207,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.617378234863281,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04252410680055618,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.609130859375,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.03801053389906883,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.625149250030518,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.03422372043132782,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.606106281280518,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.030507879331707954,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.6125993728637695,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.0340375155210495,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.5690202713012695,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.030507374554872513,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.6042985916137695,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027383938431739807,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.585570812225342,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.028792303055524826,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.566786766052246,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02866457775235176,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.566376686096191,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.028657952323555946,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.581750869750977,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.028227943927049637,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.576788902282715,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.024757765233516693,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.59373664855957,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.027442635968327522,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.570978164672852,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.02590034157037735,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.571286201477051,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.025696570053696632,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.542209625244141,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.025478003546595573,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.5900421142578125,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.02563696913421154,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.544609546661377,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.025326019152998924,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.5808515548706055,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.024502627551555634,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.560579776763916,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.02687518671154976,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.567154884338379,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.025280235335230827,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.560011863708496,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.026266315951943398,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.5758442878723145,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.026058891788125038,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.583309173583984,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.02620471641421318,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.547737121582031,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02626362442970276,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.556955337524414,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.024503158405423164,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.580735206604004,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02580532431602478,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.564000129699707,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.024437161162495613,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.545594215393066,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.026664750650525093,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.56415319442749,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.02499818615615368,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.550689697265625,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.027514580637216568,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.5548248291015625,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.026799403131008148,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.548295021057129,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.02685476280748844,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.5660600662231445,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.026104001328349113,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.556082248687744,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.027479661628603935,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.5580244064331055,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.02638247236609459,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.550073623657227,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.025853842496871948,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.578088760375977,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.02636682614684105,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.58807373046875,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.02684398740530014,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.561017036437988,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.028538964688777924,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.548762321472168,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.028327809646725655,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.548969745635986,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.025131452828645706,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.564047813415527,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.026498543098568916,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.553997039794922,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.030081963166594505,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.5787787437438965,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.03201213851571083,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.566424369812012,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.020518779754638672,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8421616554260254,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.5899484157562256,
+ "eval_runtime": 60.2671,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03692612797021866,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.492715835571289,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.050258126109838486,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.480323791503906,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.04179152473807335,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.484371185302734,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.04160204157233238,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.4905524253845215,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.03974860534071922,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.503803253173828,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.037031810730695724,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.494629859924316,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.038762833923101425,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.495111465454102,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.0412563718855381,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.490113735198975,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.041371624916791916,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.490643501281738,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04151557758450508,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.49881649017334,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04037577286362648,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.492178916931152,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.03676290065050125,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.477460861206055,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.038055796176195145,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.509150981903076,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.041152238845825195,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.47133731842041,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.03728092834353447,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.508072376251221,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.035338856279850006,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.505619049072266,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.037813492119312286,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.509489059448242,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.03789111226797104,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.508646488189697,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.03880828619003296,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.52567720413208,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.03611127287149429,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.498889446258545,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.036903638392686844,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.498601913452148,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03620576485991478,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.490840911865234,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03555903211236,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.506166458129883,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.03441924974322319,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.498001575469971,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.0334339402616024,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.491082191467285,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.031418606638908386,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.527408599853516,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03073921985924244,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.514107704162598,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.029589885845780373,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.48323917388916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.02974501997232437,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.51338529586792,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.029247552156448364,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.504205226898193,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.03132032975554466,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.505674362182617,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03088037669658661,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.477198600769043,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.031054604798555374,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.498721122741699,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03293661028146744,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.527826309204102,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.03181406110525131,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.495999336242676,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.029380083084106445,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.530351638793945,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.029296347871422768,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.492367744445801,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.02958036959171295,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.493037700653076,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03078395500779152,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.47504186630249,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03280661255121231,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.518409729003906,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.03323954716324806,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.460753917694092,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.032254140824079514,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.474285125732422,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.03018898330628872,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.5162353515625,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.02802245132625103,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.538684844970703,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.02821732498705387,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.4850287437438965,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.029189350083470345,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.50770378112793,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.02851196378469467,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.499988555908203,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.02741030789911747,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.499852180480957,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.028708625584840775,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.465211868286133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.028519680723547935,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.502464294433594,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.027846841141581535,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.503248691558838,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030456233769655228,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.490391731262207,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029952052980661392,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.50056266784668,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.02925778552889824,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.49215030670166,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.029679695144295692,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.523841857910156,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.031178541481494904,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.499871730804443,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.031264498829841614,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.50546932220459,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03482109308242798,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.494480133056641,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.03341105580329895,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.532899856567383,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028623664751648903,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.5016279220581055,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028600146993994713,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.511589050292969,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.029127901419997215,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.502058982849121,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.027898237109184265,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.493182182312012,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.028683282434940338,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.519793510437012,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.02887023612856865,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.457289695739746,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.02811296097934246,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.49398946762085,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.028485111892223358,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.478344440460205,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02914245054125786,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.487666130065918,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.028030497953295708,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.518488883972168,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.027973609045147896,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.515271186828613,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.027458522468805313,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.490856647491455,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.01909293793141842,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.822120189666748,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5866912603378296,
+ "eval_runtime": 61.3185,
+ "eval_samples_per_second": 39.825,
+ "eval_steps_per_second": 0.636,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03683992475271225,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.438363075256348,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.05042923241853714,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.443560600280762,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.055914171040058136,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.422282695770264,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.057007886469364166,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.442644119262695,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.05999710038304329,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.44780969619751,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06472347676753998,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.485074996948242,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.06923902779817581,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.457515716552734,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.0710320770740509,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.465476989746094,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.06942574679851532,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.426226615905762,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.06867868453264236,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.438092231750488,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.0688796266913414,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.446664810180664,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.0690719485282898,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.456482887268066,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.06068912148475647,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.4730730056762695,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.05411486327648163,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.4596452713012695,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.056623928248882294,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.418197154998779,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.05634097009897232,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.482892036437988,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.05056057870388031,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.457589149475098,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.04460986703634262,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.451439380645752,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.045022401958703995,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.452855110168457,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.04054102674126625,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.417323112487793,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.03682882338762283,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.465713977813721,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.03528819605708122,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.4502339363098145,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.03568802773952484,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.422399520874023,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.0363088883459568,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.447678089141846,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.03460463136434555,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.437664985656738,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.034751344472169876,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.489151954650879,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.033658601343631744,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.4471635818481445,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.03433239087462425,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.455332279205322,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.032074104994535446,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.435157775878906,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.032168325036764145,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.435118675231934,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.03007678873836994,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.42781925201416,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03094375506043434,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.433105945587158,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.032186444848775864,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.383000373840332,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.030610304325819016,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.473257064819336,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.030257435515522957,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.453041076660156,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.0301323514431715,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.396578788757324,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.030934512615203857,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.445223808288574,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.029907869175076485,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.442287445068359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03163841366767883,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.434650421142578,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.029627414420247078,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.454374313354492,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.031026875600218773,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.413784027099609,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03099491260945797,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.458438873291016,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.02981731668114662,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.438560962677002,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.02927570603787899,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.44493293762207,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.029779717326164246,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.445121765136719,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.02909068390727043,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.4736127853393555,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.029217200353741646,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.434604644775391,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.029872911050915718,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.454591751098633,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.028742222115397453,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.468222141265869,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.030313383787870407,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.45759391784668,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.03014662116765976,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.446331977844238,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029481127858161926,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.478209495544434,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.028485266491770744,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.444445610046387,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.02857700176537037,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.430604934692383,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.029038602486252785,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.43759298324585,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.029576119035482407,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.474579811096191,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.02897879108786583,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.440483093261719,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.029799485579133034,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.4839277267456055,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.030710609629750252,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.466696739196777,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.030270954594016075,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.455197334289551,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.03170182928442955,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.454664707183838,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030684730038046837,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.437619209289551,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.030972573906183243,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.431796073913574,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028914963826537132,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.460618019104004,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.029810264706611633,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.407312393188477,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.029806939885020256,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.466520309448242,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.03020886518061161,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.427349090576172,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030598826706409454,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.427891731262207,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03192012757062912,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.448113918304443,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.030492989346385002,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.450060844421387,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03188254311680794,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.43007230758667,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02081947773694992,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.782536506652832,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5860040187835693,
+ "eval_runtime": 60.6725,
+ "eval_samples_per_second": 40.249,
+ "eval_steps_per_second": 0.643,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.036721356213092804,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.358441352844238,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.056633755564689636,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.361527919769287,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.06175022944808006,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.3827924728393555,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05159056559205055,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.354856491088867,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.04885871335864067,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.359966278076172,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.04407693073153496,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.364612579345703,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.04891050234436989,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.344021320343018,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.04643657058477402,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.366066932678223,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.05241481214761734,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.3390607833862305,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.05837438255548477,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.390261650085449,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.058482322841882706,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.3856201171875,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.06507553160190582,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.414223670959473,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.0575365275144577,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.402576446533203,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05055592209100723,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.389563083648682,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.047167956829071045,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.402750015258789,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.04473405331373215,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.358315467834473,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.042440690100193024,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.392551898956299,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04465719312429428,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.373558044433594,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.04337528720498085,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.37944221496582,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.04322915896773338,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.38493013381958,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.039873406291007996,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.389793395996094,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.0413178876042366,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.427964210510254,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.04567280039191246,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.404172897338867,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04154926910996437,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.391359329223633,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.03774876892566681,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.381457328796387,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03588850423693657,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.392526626586914,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.03757261857390404,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.39273738861084,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03849692642688751,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.409083366394043,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03787294402718544,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.3776044845581055,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.034496866166591644,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.413462162017822,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.03973205015063286,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.359318733215332,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.036192040890455246,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.392844200134277,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.03436632081866264,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.421384811401367,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.035658642649650574,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.386757850646973,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.03434896469116211,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.364978313446045,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03642290085554123,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.385326385498047,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.03427736461162567,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.406352996826172,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.034845270216464996,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.370953559875488,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03390975296497345,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.3799943923950195,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.03134961053729057,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.419156551361084,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.03512260690331459,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.404660701751709,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03319067507982254,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.388535022735596,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03113507106900215,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.403522968292236,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03559132665395737,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.382323265075684,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.03579750657081604,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.37620210647583,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.033675212413072586,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.411443710327148,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.03220067173242569,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.443470001220703,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031610481441020966,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.406038284301758,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.02958764135837555,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.41604471206665,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.0318237729370594,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4253387451171875,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03260074555873871,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.370792388916016,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.03359333053231239,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.397639274597168,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032516930252313614,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.4140777587890625,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.03203396871685982,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.408872604370117,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.03201935440301895,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.402975082397461,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.03330211713910103,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.394018173217773,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.0348181314766407,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.380587100982666,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.03174913674592972,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.372378349304199,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.029864462092518806,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.42018985748291,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.033312153071165085,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.385973930358887,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03203144669532776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.395167827606201,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.031578030437231064,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.3982062339782715,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.030885562300682068,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.400731086730957,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03235667943954468,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.389830589294434,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.032664328813552856,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.401277542114258,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.030354196205735207,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.422520637512207,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.031162403523921967,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.393032073974609,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.032187070697546005,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.464164733886719,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.03445067256689072,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.368991851806641,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03398612141609192,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.418752670288086,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.029730670154094696,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.41429328918457,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022440090775489807,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7315595149993896,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5853753089904785,
+ "eval_runtime": 60.7174,
+ "eval_samples_per_second": 40.219,
+ "eval_steps_per_second": 0.642,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.03754177689552307,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.304028511047363,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.051616180688142776,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.34659481048584,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055258627980947495,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.310445785522461,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.049489911645650864,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3416619300842285,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05377533286809921,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.3143310546875,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.05443088337779045,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.320600986480713,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.050222691148519516,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.334789752960205,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.051904670894145966,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.316756248474121,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.05376426503062248,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.274116516113281,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.053313180804252625,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.321177005767822,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.058234598487615585,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.3169264793396,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05730101093649864,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.307013511657715,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.058140840381383896,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.332961082458496,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05654815956950188,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.315462112426758,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.05129384249448776,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.338476181030273,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.047652702778577805,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.321232795715332,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.050912003964185715,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.318971633911133,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.0514022521674633,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.3334150314331055,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.05457981303334236,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.325802803039551,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04678885266184807,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.307281494140625,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.05048161372542381,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.299831390380859,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.05334261804819107,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.313231468200684,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.05055771768093109,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.36859130859375,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04955561086535454,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.301530361175537,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.04548008739948273,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.339517593383789,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.046585116535425186,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.333036422729492,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.0503716766834259,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.364710807800293,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.040569406002759933,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.369089126586914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.03754054754972458,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.3349432945251465,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03775954991579056,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.333134174346924,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.04066266492009163,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.354358673095703,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.040490977466106415,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.360981464385986,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.0426657609641552,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.357726097106934,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.0415605828166008,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.363457679748535,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.04020509123802185,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.372447967529297,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.04101516678929329,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.292452812194824,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.03546427935361862,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.33892297744751,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.03661105036735535,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.329778671264648,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03565476834774017,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.351012229919434,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.03591311350464821,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.376980781555176,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.03665630891919136,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.354767799377441,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.0371236577630043,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.374916076660156,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.03464849665760994,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.3713059425354,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03345152735710144,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.350670337677002,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.03357568010687828,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.3573713302612305,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03503486514091492,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.362038612365723,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.03455005958676338,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.3263773918151855,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.03173859417438507,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.351345062255859,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03301836550235748,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.398364067077637,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.033733855932950974,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.335268020629883,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034080155193805695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.371492385864258,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.034396376460790634,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.382076740264893,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.03303401917219162,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.36623477935791,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03297369182109833,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3148393630981445,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.034819211810827255,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.358110427856445,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.033389780670404434,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.367427825927734,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.032363444566726685,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.3813676834106445,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03509577363729477,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.386520862579346,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.034685730934143066,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.33738374710083,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.034755002707242966,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.360980987548828,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03461688384413719,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.396650314331055,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03392560034990311,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.356130123138428,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.03342147544026375,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.359145164489746,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.032852694392204285,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.3796162605285645,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03472859039902687,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.361110687255859,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03277098760008812,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.361811637878418,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.03341882303357124,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.369524955749512,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.032347049564123154,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.369096755981445,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03161193057894707,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.339844703674316,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.03322301432490349,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.376675605773926,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03419092297554016,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.346864223480225,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02461826801300049,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7236275672912598,
+ "step": 1080
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 4.608441369545933e+17,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1080/training_args.bin b/runs/l2r90-wd0033-ssl/checkpoint-1080/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..313c41733ff55284c06034a59bb885b30e40ffc8
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1080/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7ddd0bdddba68d3e19ab33477000f3fc684fe4ee8e3580eeb8c02646e5fc27ac
+size 4856
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1440/chat_template.jinja b/runs/l2r90-wd0033-ssl/checkpoint-1440/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1440/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1440/config.json b/runs/l2r90-wd0033-ssl/checkpoint-1440/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1440/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1440/generation_config.json b/runs/l2r90-wd0033-ssl/checkpoint-1440/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1440/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1440/model.safetensors b/runs/l2r90-wd0033-ssl/checkpoint-1440/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..a764a746c9021aad98a05860593a47537b7bcdb3
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1440/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:cd633253a827bdaec99558a8b2562803d5c3c185b776e3f996f725fa9d530bd2
+size 583356232
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1440/optimizer.pt b/runs/l2r90-wd0033-ssl/checkpoint-1440/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..2e4664440eb255d319e7185023c191a8afd20cc6
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1440/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4b809a1d3449aa3d023f710f5c6f7a5a3e7b82e75e82d434730509f0fb57fbb9
+size 1166825338
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1440/rng_state_0.pth b/runs/l2r90-wd0033-ssl/checkpoint-1440/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..aef5c379b1e7bf7cc59526983c1472f4b0d7adc5
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1440/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:080d651ea8a26085b7aabaad02f9a561809bc425226d0a35b3b91c62f17a7b2b
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1440/rng_state_1.pth b/runs/l2r90-wd0033-ssl/checkpoint-1440/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..235fb514f02235435ff61ba15b79fbd8a1111e1d
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1440/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f69206403a46fd4968a11322d4859dc6c1089247abc5792063e8ab4b78218928
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1440/scheduler.pt b/runs/l2r90-wd0033-ssl/checkpoint-1440/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..59b8db4a485766e1d20d11a905a37774da0cec09
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1440/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2fe49e98b5d706cd9174dc7cf5e5cd5039ca03537426b9020792a5363720022f
+size 1064
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1440/tokenizer.json b/runs/l2r90-wd0033-ssl/checkpoint-1440/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1440/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1440/tokenizer_config.json b/runs/l2r90-wd0033-ssl/checkpoint-1440/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1440/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1440/trainer_state.json b/runs/l2r90-wd0033-ssl/checkpoint-1440/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..1eaf0a674a17a1519f678d250d7a3d4c99a26b71
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1440/trainer_state.json
@@ -0,0 +1,10266 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 20.0,
+ "eval_steps": 500,
+ "global_step": 1440,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013961605584642234,
+ "grad_norm": 0.26733073592185974,
+ "learning_rate": 0.0,
+ "loss": 7.990724086761475,
+ "step": 1
+ },
+ {
+ "epoch": 0.027923211169284468,
+ "grad_norm": 0.26712340116500854,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 7.990110874176025,
+ "step": 2
+ },
+ {
+ "epoch": 0.041884816753926704,
+ "grad_norm": 0.2646932601928711,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 7.956935405731201,
+ "step": 3
+ },
+ {
+ "epoch": 0.055846422338568937,
+ "grad_norm": 0.24684545397758484,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 7.890539646148682,
+ "step": 4
+ },
+ {
+ "epoch": 0.06980802792321117,
+ "grad_norm": 0.21016643941402435,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 7.8146138191223145,
+ "step": 5
+ },
+ {
+ "epoch": 0.08376963350785341,
+ "grad_norm": 0.17992742359638214,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 7.724307060241699,
+ "step": 6
+ },
+ {
+ "epoch": 0.09773123909249563,
+ "grad_norm": 0.14356893301010132,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 7.645662307739258,
+ "step": 7
+ },
+ {
+ "epoch": 0.11169284467713787,
+ "grad_norm": 0.10830964893102646,
+ "learning_rate": 4.2e-05,
+ "loss": 7.582019329071045,
+ "step": 8
+ },
+ {
+ "epoch": 0.1256544502617801,
+ "grad_norm": 0.08265161514282227,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 7.5230793952941895,
+ "step": 9
+ },
+ {
+ "epoch": 0.13961605584642234,
+ "grad_norm": 0.06423710286617279,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 7.473812103271484,
+ "step": 10
+ },
+ {
+ "epoch": 0.15357766143106458,
+ "grad_norm": 0.05132246017456055,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 7.440786838531494,
+ "step": 11
+ },
+ {
+ "epoch": 0.16753926701570682,
+ "grad_norm": 0.040572404861450195,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 7.41290283203125,
+ "step": 12
+ },
+ {
+ "epoch": 0.18150087260034903,
+ "grad_norm": 0.03138705715537071,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 7.39749002456665,
+ "step": 13
+ },
+ {
+ "epoch": 0.19546247818499127,
+ "grad_norm": 0.02553008496761322,
+ "learning_rate": 7.8e-05,
+ "loss": 7.387092590332031,
+ "step": 14
+ },
+ {
+ "epoch": 0.2094240837696335,
+ "grad_norm": 0.022178977727890015,
+ "learning_rate": 8.4e-05,
+ "loss": 7.378993511199951,
+ "step": 15
+ },
+ {
+ "epoch": 0.22338568935427575,
+ "grad_norm": 0.019915293902158737,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 7.37220573425293,
+ "step": 16
+ },
+ {
+ "epoch": 0.23734729493891799,
+ "grad_norm": 0.017390629276633263,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 7.366759300231934,
+ "step": 17
+ },
+ {
+ "epoch": 0.2513089005235602,
+ "grad_norm": 0.015472771599888802,
+ "learning_rate": 0.000102,
+ "loss": 7.3652663230896,
+ "step": 18
+ },
+ {
+ "epoch": 0.26527050610820246,
+ "grad_norm": 0.013599214144051075,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 7.359714984893799,
+ "step": 19
+ },
+ {
+ "epoch": 0.2792321116928447,
+ "grad_norm": 0.012442861683666706,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 7.353302001953125,
+ "step": 20
+ },
+ {
+ "epoch": 0.2931937172774869,
+ "grad_norm": 0.010471965186297894,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 7.357417583465576,
+ "step": 21
+ },
+ {
+ "epoch": 0.30715532286212915,
+ "grad_norm": 0.009647224098443985,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 7.351763725280762,
+ "step": 22
+ },
+ {
+ "epoch": 0.32111692844677137,
+ "grad_norm": 0.008228210732340813,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 7.345434188842773,
+ "step": 23
+ },
+ {
+ "epoch": 0.33507853403141363,
+ "grad_norm": 0.007454514969140291,
+ "learning_rate": 0.000138,
+ "loss": 7.342520236968994,
+ "step": 24
+ },
+ {
+ "epoch": 0.34904013961605584,
+ "grad_norm": 0.0069856527261435986,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 7.338870048522949,
+ "step": 25
+ },
+ {
+ "epoch": 0.36300174520069806,
+ "grad_norm": 0.006091665010899305,
+ "learning_rate": 0.00015,
+ "loss": 7.328121662139893,
+ "step": 26
+ },
+ {
+ "epoch": 0.3769633507853403,
+ "grad_norm": 0.005525320768356323,
+ "learning_rate": 0.000156,
+ "loss": 7.335029602050781,
+ "step": 27
+ },
+ {
+ "epoch": 0.39092495636998253,
+ "grad_norm": 0.0061133443377912045,
+ "learning_rate": 0.000162,
+ "loss": 7.319417953491211,
+ "step": 28
+ },
+ {
+ "epoch": 0.4048865619546248,
+ "grad_norm": 0.006040909793227911,
+ "learning_rate": 0.000168,
+ "loss": 7.326376438140869,
+ "step": 29
+ },
+ {
+ "epoch": 0.418848167539267,
+ "grad_norm": 0.006003216840326786,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 7.319552421569824,
+ "step": 30
+ },
+ {
+ "epoch": 0.4328097731239092,
+ "grad_norm": 0.007096003741025925,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 7.323486328125,
+ "step": 31
+ },
+ {
+ "epoch": 0.4467713787085515,
+ "grad_norm": 0.007725459989160299,
+ "learning_rate": 0.000186,
+ "loss": 7.3190717697143555,
+ "step": 32
+ },
+ {
+ "epoch": 0.4607329842931937,
+ "grad_norm": 0.007338670082390308,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 7.317154884338379,
+ "step": 33
+ },
+ {
+ "epoch": 0.47469458987783597,
+ "grad_norm": 0.007709544617682695,
+ "learning_rate": 0.000198,
+ "loss": 7.311826705932617,
+ "step": 34
+ },
+ {
+ "epoch": 0.4886561954624782,
+ "grad_norm": 0.009249784983694553,
+ "learning_rate": 0.000204,
+ "loss": 7.322498798370361,
+ "step": 35
+ },
+ {
+ "epoch": 0.5026178010471204,
+ "grad_norm": 0.006950767710804939,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 7.314571380615234,
+ "step": 36
+ },
+ {
+ "epoch": 0.5165794066317626,
+ "grad_norm": 0.007583985570818186,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 7.309540748596191,
+ "step": 37
+ },
+ {
+ "epoch": 0.5305410122164049,
+ "grad_norm": 0.014731820672750473,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 7.307004928588867,
+ "step": 38
+ },
+ {
+ "epoch": 0.5445026178010471,
+ "grad_norm": 0.016276473179459572,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 7.299642562866211,
+ "step": 39
+ },
+ {
+ "epoch": 0.5584642233856894,
+ "grad_norm": 0.020018689334392548,
+ "learning_rate": 0.000234,
+ "loss": 7.298466682434082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5724258289703316,
+ "grad_norm": 0.014941536821424961,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 7.289968967437744,
+ "step": 41
+ },
+ {
+ "epoch": 0.5863874345549738,
+ "grad_norm": 0.018299812451004982,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 7.276564598083496,
+ "step": 42
+ },
+ {
+ "epoch": 0.6003490401396161,
+ "grad_norm": 0.01615075021982193,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.2575554847717285,
+ "step": 43
+ },
+ {
+ "epoch": 0.6143106457242583,
+ "grad_norm": 0.013063831254839897,
+ "learning_rate": 0.000258,
+ "loss": 7.253885269165039,
+ "step": 44
+ },
+ {
+ "epoch": 0.6282722513089005,
+ "grad_norm": 0.015211676247417927,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.245779037475586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6422338568935427,
+ "grad_norm": 0.02093500830233097,
+ "learning_rate": 0.00027,
+ "loss": 7.225130081176758,
+ "step": 46
+ },
+ {
+ "epoch": 0.6561954624781849,
+ "grad_norm": 0.01808053068816662,
+ "learning_rate": 0.000276,
+ "loss": 7.209654331207275,
+ "step": 47
+ },
+ {
+ "epoch": 0.6701570680628273,
+ "grad_norm": 0.02209644205868244,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.197124481201172,
+ "step": 48
+ },
+ {
+ "epoch": 0.6841186736474695,
+ "grad_norm": 0.023524878546595573,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.1841936111450195,
+ "step": 49
+ },
+ {
+ "epoch": 0.6980802792321117,
+ "grad_norm": 0.031148087233304977,
+ "learning_rate": 0.000294,
+ "loss": 7.188040733337402,
+ "step": 50
+ },
+ {
+ "epoch": 0.7120418848167539,
+ "grad_norm": 0.02931929938495159,
+ "learning_rate": 0.0003,
+ "loss": 7.17371940612793,
+ "step": 51
+ },
+ {
+ "epoch": 0.7260034904013961,
+ "grad_norm": 0.01611483097076416,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.1714324951171875,
+ "step": 52
+ },
+ {
+ "epoch": 0.7399650959860384,
+ "grad_norm": 0.02619907818734646,
+ "learning_rate": 0.000312,
+ "loss": 7.158261299133301,
+ "step": 53
+ },
+ {
+ "epoch": 0.7539267015706806,
+ "grad_norm": 0.049462560564279556,
+ "learning_rate": 0.000318,
+ "loss": 7.167076110839844,
+ "step": 54
+ },
+ {
+ "epoch": 0.7678883071553229,
+ "grad_norm": 0.09835070371627808,
+ "learning_rate": 0.000324,
+ "loss": 7.154292106628418,
+ "step": 55
+ },
+ {
+ "epoch": 0.7818499127399651,
+ "grad_norm": 0.11406510323286057,
+ "learning_rate": 0.00033,
+ "loss": 7.162016868591309,
+ "step": 56
+ },
+ {
+ "epoch": 0.7958115183246073,
+ "grad_norm": 0.027026303112506866,
+ "learning_rate": 0.000336,
+ "loss": 7.141416549682617,
+ "step": 57
+ },
+ {
+ "epoch": 0.8097731239092496,
+ "grad_norm": 0.06690579652786255,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.13959264755249,
+ "step": 58
+ },
+ {
+ "epoch": 0.8237347294938918,
+ "grad_norm": 0.020919431000947952,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.126513481140137,
+ "step": 59
+ },
+ {
+ "epoch": 0.837696335078534,
+ "grad_norm": 0.05618130415678024,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.124345779418945,
+ "step": 60
+ },
+ {
+ "epoch": 0.8516579406631762,
+ "grad_norm": 0.03402302414178848,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.100615501403809,
+ "step": 61
+ },
+ {
+ "epoch": 0.8656195462478184,
+ "grad_norm": 0.03646906837821007,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.116507530212402,
+ "step": 62
+ },
+ {
+ "epoch": 0.8795811518324608,
+ "grad_norm": 0.03076397068798542,
+ "learning_rate": 0.000372,
+ "loss": 7.097659111022949,
+ "step": 63
+ },
+ {
+ "epoch": 0.893542757417103,
+ "grad_norm": 0.03329646959900856,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.097290992736816,
+ "step": 64
+ },
+ {
+ "epoch": 0.9075043630017452,
+ "grad_norm": 0.01998630352318287,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.092268943786621,
+ "step": 65
+ },
+ {
+ "epoch": 0.9214659685863874,
+ "grad_norm": 0.03278940171003342,
+ "learning_rate": 0.00039,
+ "loss": 7.088292121887207,
+ "step": 66
+ },
+ {
+ "epoch": 0.9354275741710296,
+ "grad_norm": 0.024620421230793,
+ "learning_rate": 0.000396,
+ "loss": 7.078650951385498,
+ "step": 67
+ },
+ {
+ "epoch": 0.9493891797556719,
+ "grad_norm": 0.03750383481383324,
+ "learning_rate": 0.000402,
+ "loss": 7.052778244018555,
+ "step": 68
+ },
+ {
+ "epoch": 0.9633507853403142,
+ "grad_norm": 0.07558075338602066,
+ "learning_rate": 0.000408,
+ "loss": 7.045991897583008,
+ "step": 69
+ },
+ {
+ "epoch": 0.9773123909249564,
+ "grad_norm": 0.16021482646465302,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.023152828216553,
+ "step": 70
+ },
+ {
+ "epoch": 0.9912739965095986,
+ "grad_norm": 0.24981893599033356,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.067540645599365,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.10255520790815353,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 4.425640106201172,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8861758708953857,
+ "eval_runtime": 58.2699,
+ "eval_samples_per_second": 41.908,
+ "eval_steps_per_second": 0.669,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139616055846423,
+ "grad_norm": 0.13374797999858856,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.073373317718506,
+ "step": 73
+ },
+ {
+ "epoch": 1.0279232111692844,
+ "grad_norm": 0.05079387128353119,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.0243306159973145,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418848167539267,
+ "grad_norm": 0.10056068748235703,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.014476299285889,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558464223385688,
+ "grad_norm": 0.06995758414268494,
+ "learning_rate": 0.00045,
+ "loss": 7.011931419372559,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698080279232112,
+ "grad_norm": 0.0483892522752285,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 6.9822797775268555,
+ "step": 77
+ },
+ {
+ "epoch": 1.0837696335078535,
+ "grad_norm": 0.08408170193433762,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 6.964348793029785,
+ "step": 78
+ },
+ {
+ "epoch": 1.0977312390924956,
+ "grad_norm": 0.08027710020542145,
+ "learning_rate": 0.000468,
+ "loss": 6.96852445602417,
+ "step": 79
+ },
+ {
+ "epoch": 1.111692844677138,
+ "grad_norm": 0.05647159367799759,
+ "learning_rate": 0.000474,
+ "loss": 6.944484710693359,
+ "step": 80
+ },
+ {
+ "epoch": 1.12565445026178,
+ "grad_norm": 0.07464330643415451,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 6.9246745109558105,
+ "step": 81
+ },
+ {
+ "epoch": 1.1396160558464223,
+ "grad_norm": 0.04451674595475197,
+ "learning_rate": 0.000486,
+ "loss": 6.895403861999512,
+ "step": 82
+ },
+ {
+ "epoch": 1.1535776614310647,
+ "grad_norm": 0.05820256471633911,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 6.887338638305664,
+ "step": 83
+ },
+ {
+ "epoch": 1.1675392670157068,
+ "grad_norm": 0.02925296127796173,
+ "learning_rate": 0.000498,
+ "loss": 6.857797622680664,
+ "step": 84
+ },
+ {
+ "epoch": 1.181500872600349,
+ "grad_norm": 0.05096987634897232,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 6.854822158813477,
+ "step": 85
+ },
+ {
+ "epoch": 1.1954624781849912,
+ "grad_norm": 0.03483197093009949,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 6.840914726257324,
+ "step": 86
+ },
+ {
+ "epoch": 1.2094240837696335,
+ "grad_norm": 0.03323264792561531,
+ "learning_rate": 0.000516,
+ "loss": 6.834497451782227,
+ "step": 87
+ },
+ {
+ "epoch": 1.2233856893542758,
+ "grad_norm": 0.03290868178009987,
+ "learning_rate": 0.000522,
+ "loss": 6.820746898651123,
+ "step": 88
+ },
+ {
+ "epoch": 1.237347294938918,
+ "grad_norm": 0.025398844853043556,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 6.795746803283691,
+ "step": 89
+ },
+ {
+ "epoch": 1.2513089005235603,
+ "grad_norm": 0.030700596049427986,
+ "learning_rate": 0.000534,
+ "loss": 6.7867889404296875,
+ "step": 90
+ },
+ {
+ "epoch": 1.2652705061082026,
+ "grad_norm": 0.019589420408010483,
+ "learning_rate": 0.00054,
+ "loss": 6.77443790435791,
+ "step": 91
+ },
+ {
+ "epoch": 1.2792321116928447,
+ "grad_norm": 0.03313920646905899,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 6.762271881103516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2931937172774868,
+ "grad_norm": 0.025389468297362328,
+ "learning_rate": 0.000552,
+ "loss": 6.750258445739746,
+ "step": 93
+ },
+ {
+ "epoch": 1.307155322862129,
+ "grad_norm": 0.040296655148267746,
+ "learning_rate": 0.000558,
+ "loss": 6.729808807373047,
+ "step": 94
+ },
+ {
+ "epoch": 1.3211169284467714,
+ "grad_norm": 0.04735933244228363,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 6.737982749938965,
+ "step": 95
+ },
+ {
+ "epoch": 1.3350785340314135,
+ "grad_norm": 0.05529680848121643,
+ "learning_rate": 0.00057,
+ "loss": 6.735714912414551,
+ "step": 96
+ },
+ {
+ "epoch": 1.3490401396160558,
+ "grad_norm": 0.07148554176092148,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 6.708859920501709,
+ "step": 97
+ },
+ {
+ "epoch": 1.3630017452006982,
+ "grad_norm": 0.059708479791879654,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 6.708860397338867,
+ "step": 98
+ },
+ {
+ "epoch": 1.3769633507853403,
+ "grad_norm": 0.027111195027828217,
+ "learning_rate": 0.000588,
+ "loss": 6.686685085296631,
+ "step": 99
+ },
+ {
+ "epoch": 1.3909249563699826,
+ "grad_norm": 0.0500967763364315,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 6.676325798034668,
+ "step": 100
+ },
+ {
+ "epoch": 1.404886561954625,
+ "grad_norm": 0.044565796852111816,
+ "learning_rate": 0.0006,
+ "loss": 6.674530029296875,
+ "step": 101
+ },
+ {
+ "epoch": 1.418848167539267,
+ "grad_norm": 0.03608180209994316,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 6.653741359710693,
+ "step": 102
+ },
+ {
+ "epoch": 1.4328097731239091,
+ "grad_norm": 0.030394982546567917,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 6.643152236938477,
+ "step": 103
+ },
+ {
+ "epoch": 1.4467713787085514,
+ "grad_norm": 0.040185656398534775,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 6.628686904907227,
+ "step": 104
+ },
+ {
+ "epoch": 1.4607329842931938,
+ "grad_norm": 0.021049322560429573,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 6.610864639282227,
+ "step": 105
+ },
+ {
+ "epoch": 1.4746945898778359,
+ "grad_norm": 0.04164277762174606,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 6.602880954742432,
+ "step": 106
+ },
+ {
+ "epoch": 1.4886561954624782,
+ "grad_norm": 0.0389864556491375,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 6.599456787109375,
+ "step": 107
+ },
+ {
+ "epoch": 1.5026178010471205,
+ "grad_norm": 0.05805973336100578,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 6.588700294494629,
+ "step": 108
+ },
+ {
+ "epoch": 1.5165794066317626,
+ "grad_norm": 0.08825036883354187,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 6.587496757507324,
+ "step": 109
+ },
+ {
+ "epoch": 1.530541012216405,
+ "grad_norm": 0.12521523237228394,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 6.626501083374023,
+ "step": 110
+ },
+ {
+ "epoch": 1.5445026178010473,
+ "grad_norm": 0.07429645210504532,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 6.590241432189941,
+ "step": 111
+ },
+ {
+ "epoch": 1.5584642233856894,
+ "grad_norm": 0.06463009119033813,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 6.578714370727539,
+ "step": 112
+ },
+ {
+ "epoch": 1.5724258289703315,
+ "grad_norm": 0.07508203387260437,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 6.561270713806152,
+ "step": 113
+ },
+ {
+ "epoch": 1.5863874345549738,
+ "grad_norm": 0.04125197231769562,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 6.558821678161621,
+ "step": 114
+ },
+ {
+ "epoch": 1.600349040139616,
+ "grad_norm": 0.05457386001944542,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.541611671447754,
+ "step": 115
+ },
+ {
+ "epoch": 1.6143106457242582,
+ "grad_norm": 0.03684385493397713,
+ "learning_rate": 0.000599973080557108,
+ "loss": 6.547738075256348,
+ "step": 116
+ },
+ {
+ "epoch": 1.6282722513089005,
+ "grad_norm": 0.05888746306300163,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.529491424560547,
+ "step": 117
+ },
+ {
+ "epoch": 1.6422338568935428,
+ "grad_norm": 0.056668370962142944,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.519444942474365,
+ "step": 118
+ },
+ {
+ "epoch": 1.656195462478185,
+ "grad_norm": 0.09123896062374115,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.522984981536865,
+ "step": 119
+ },
+ {
+ "epoch": 1.6701570680628273,
+ "grad_norm": 0.12380023300647736,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.515375137329102,
+ "step": 120
+ },
+ {
+ "epoch": 1.6841186736474696,
+ "grad_norm": 0.06953322887420654,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.500491142272949,
+ "step": 121
+ },
+ {
+ "epoch": 1.6980802792321117,
+ "grad_norm": 0.08613869547843933,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.5097198486328125,
+ "step": 122
+ },
+ {
+ "epoch": 1.7120418848167538,
+ "grad_norm": 0.08296383172273636,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.487521171569824,
+ "step": 123
+ },
+ {
+ "epoch": 1.7260034904013961,
+ "grad_norm": 0.05744783207774162,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.485134124755859,
+ "step": 124
+ },
+ {
+ "epoch": 1.7399650959860384,
+ "grad_norm": 0.04108566418290138,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.466182708740234,
+ "step": 125
+ },
+ {
+ "epoch": 1.7539267015706805,
+ "grad_norm": 0.05374931916594505,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.479011535644531,
+ "step": 126
+ },
+ {
+ "epoch": 1.7678883071553229,
+ "grad_norm": 0.03867914527654648,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.465902328491211,
+ "step": 127
+ },
+ {
+ "epoch": 1.7818499127399652,
+ "grad_norm": 0.05477374792098999,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.453082084655762,
+ "step": 128
+ },
+ {
+ "epoch": 1.7958115183246073,
+ "grad_norm": 0.06508298963308334,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.446089744567871,
+ "step": 129
+ },
+ {
+ "epoch": 1.8097731239092496,
+ "grad_norm": 0.06388624757528305,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.451748847961426,
+ "step": 130
+ },
+ {
+ "epoch": 1.823734729493892,
+ "grad_norm": 0.04223188757896423,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.4492387771606445,
+ "step": 131
+ },
+ {
+ "epoch": 1.837696335078534,
+ "grad_norm": 0.04528658092021942,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.409390926361084,
+ "step": 132
+ },
+ {
+ "epoch": 1.8516579406631761,
+ "grad_norm": 0.04962967336177826,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.415774345397949,
+ "step": 133
+ },
+ {
+ "epoch": 1.8656195462478184,
+ "grad_norm": 0.031944390386343,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.411528587341309,
+ "step": 134
+ },
+ {
+ "epoch": 1.8795811518324608,
+ "grad_norm": 0.028411149978637695,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.392329216003418,
+ "step": 135
+ },
+ {
+ "epoch": 1.8935427574171029,
+ "grad_norm": 0.03516148403286934,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.392238616943359,
+ "step": 136
+ },
+ {
+ "epoch": 1.9075043630017452,
+ "grad_norm": 0.027800120413303375,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.39179801940918,
+ "step": 137
+ },
+ {
+ "epoch": 1.9214659685863875,
+ "grad_norm": 0.02362987957894802,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.378960609436035,
+ "step": 138
+ },
+ {
+ "epoch": 1.9354275741710296,
+ "grad_norm": 0.033999308943748474,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.36802339553833,
+ "step": 139
+ },
+ {
+ "epoch": 1.949389179755672,
+ "grad_norm": 0.03290879726409912,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.366308689117432,
+ "step": 140
+ },
+ {
+ "epoch": 1.9633507853403143,
+ "grad_norm": 0.0286843404173851,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.350399971008301,
+ "step": 141
+ },
+ {
+ "epoch": 1.9773123909249564,
+ "grad_norm": 0.027145877480506897,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.33477783203125,
+ "step": 142
+ },
+ {
+ "epoch": 1.9912739965095985,
+ "grad_norm": 0.03530878946185112,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.322859764099121,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.028133150190114975,
+ "learning_rate": 0.000599778806120077,
+ "loss": 3.955288887023926,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7927101254463196,
+ "eval_runtime": 60.8372,
+ "eval_samples_per_second": 40.14,
+ "eval_steps_per_second": 0.641,
+ "step": 144
+ },
+ {
+ "epoch": 2.013961605584642,
+ "grad_norm": 0.03054557554423809,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.311975479125977,
+ "step": 145
+ },
+ {
+ "epoch": 2.0279232111692846,
+ "grad_norm": 0.05406459793448448,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.3030853271484375,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418848167539267,
+ "grad_norm": 0.10911019891500473,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.329056739807129,
+ "step": 147
+ },
+ {
+ "epoch": 2.055846422338569,
+ "grad_norm": 0.13948315382003784,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.362421035766602,
+ "step": 148
+ },
+ {
+ "epoch": 2.0698080279232114,
+ "grad_norm": 0.13176991045475006,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.3888139724731445,
+ "step": 149
+ },
+ {
+ "epoch": 2.0837696335078535,
+ "grad_norm": 0.09615885466337204,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.35154914855957,
+ "step": 150
+ },
+ {
+ "epoch": 2.0977312390924956,
+ "grad_norm": 0.06462561339139938,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.335790634155273,
+ "step": 151
+ },
+ {
+ "epoch": 2.1116928446771377,
+ "grad_norm": 0.05186621472239494,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.324396133422852,
+ "step": 152
+ },
+ {
+ "epoch": 2.1256544502617802,
+ "grad_norm": 0.06095981225371361,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.318549156188965,
+ "step": 153
+ },
+ {
+ "epoch": 2.1396160558464223,
+ "grad_norm": 0.036121442914009094,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.300607681274414,
+ "step": 154
+ },
+ {
+ "epoch": 2.1535776614310644,
+ "grad_norm": 0.055716633796691895,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.307400703430176,
+ "step": 155
+ },
+ {
+ "epoch": 2.167539267015707,
+ "grad_norm": 0.037626396864652634,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.287364959716797,
+ "step": 156
+ },
+ {
+ "epoch": 2.181500872600349,
+ "grad_norm": 0.04329098388552666,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.28244686126709,
+ "step": 157
+ },
+ {
+ "epoch": 2.195462478184991,
+ "grad_norm": 0.031868450343608856,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.2602410316467285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2094240837696333,
+ "grad_norm": 0.03647344186902046,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.263988971710205,
+ "step": 159
+ },
+ {
+ "epoch": 2.223385689354276,
+ "grad_norm": 0.03141765296459198,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.263205051422119,
+ "step": 160
+ },
+ {
+ "epoch": 2.237347294938918,
+ "grad_norm": 0.028253229334950447,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.253107070922852,
+ "step": 161
+ },
+ {
+ "epoch": 2.25130890052356,
+ "grad_norm": 0.03608971834182739,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.2571702003479,
+ "step": 162
+ },
+ {
+ "epoch": 2.2652705061082026,
+ "grad_norm": 0.02644292451441288,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.241732597351074,
+ "step": 163
+ },
+ {
+ "epoch": 2.2792321116928447,
+ "grad_norm": 0.028604455292224884,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.230289459228516,
+ "step": 164
+ },
+ {
+ "epoch": 2.2931937172774868,
+ "grad_norm": 0.023704614490270615,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.226101398468018,
+ "step": 165
+ },
+ {
+ "epoch": 2.3071553228621293,
+ "grad_norm": 0.020691027864813805,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.2096405029296875,
+ "step": 166
+ },
+ {
+ "epoch": 2.3211169284467714,
+ "grad_norm": 0.025598427280783653,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.22075080871582,
+ "step": 167
+ },
+ {
+ "epoch": 2.3350785340314135,
+ "grad_norm": 0.030546002089977264,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.216886520385742,
+ "step": 168
+ },
+ {
+ "epoch": 2.349040139616056,
+ "grad_norm": 0.026066245511174202,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.2022504806518555,
+ "step": 169
+ },
+ {
+ "epoch": 2.363001745200698,
+ "grad_norm": 0.023802239447832108,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.189509391784668,
+ "step": 170
+ },
+ {
+ "epoch": 2.3769633507853403,
+ "grad_norm": 0.029805002734065056,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.186277389526367,
+ "step": 171
+ },
+ {
+ "epoch": 2.3909249563699824,
+ "grad_norm": 0.043327830731868744,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.164876937866211,
+ "step": 172
+ },
+ {
+ "epoch": 2.404886561954625,
+ "grad_norm": 0.058787211775779724,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.179936408996582,
+ "step": 173
+ },
+ {
+ "epoch": 2.418848167539267,
+ "grad_norm": 0.0787397176027298,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.1830244064331055,
+ "step": 174
+ },
+ {
+ "epoch": 2.432809773123909,
+ "grad_norm": 0.0899854376912117,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.190343856811523,
+ "step": 175
+ },
+ {
+ "epoch": 2.4467713787085517,
+ "grad_norm": 0.06520707905292511,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.185057163238525,
+ "step": 176
+ },
+ {
+ "epoch": 2.4607329842931938,
+ "grad_norm": 0.1002504974603653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.171279430389404,
+ "step": 177
+ },
+ {
+ "epoch": 2.474694589877836,
+ "grad_norm": 0.09591004997491837,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.174382209777832,
+ "step": 178
+ },
+ {
+ "epoch": 2.488656195462478,
+ "grad_norm": 0.12127121537923813,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.185717582702637,
+ "step": 179
+ },
+ {
+ "epoch": 2.5026178010471205,
+ "grad_norm": 0.09640590846538544,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.196681976318359,
+ "step": 180
+ },
+ {
+ "epoch": 2.5165794066317626,
+ "grad_norm": 0.048341453075408936,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.158774375915527,
+ "step": 181
+ },
+ {
+ "epoch": 2.530541012216405,
+ "grad_norm": 0.08206334710121155,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.17445707321167,
+ "step": 182
+ },
+ {
+ "epoch": 2.5445026178010473,
+ "grad_norm": 0.038934968411922455,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.170527458190918,
+ "step": 183
+ },
+ {
+ "epoch": 2.5584642233856894,
+ "grad_norm": 0.060987457633018494,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.1501994132995605,
+ "step": 184
+ },
+ {
+ "epoch": 2.5724258289703315,
+ "grad_norm": 0.033269546926021576,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.144186019897461,
+ "step": 185
+ },
+ {
+ "epoch": 2.5863874345549736,
+ "grad_norm": 0.046137236058712006,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.1236677169799805,
+ "step": 186
+ },
+ {
+ "epoch": 2.600349040139616,
+ "grad_norm": 0.038265109062194824,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.121979236602783,
+ "step": 187
+ },
+ {
+ "epoch": 2.614310645724258,
+ "grad_norm": 0.039128001779317856,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.124156951904297,
+ "step": 188
+ },
+ {
+ "epoch": 2.6282722513089007,
+ "grad_norm": 0.025303540751338005,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.108036994934082,
+ "step": 189
+ },
+ {
+ "epoch": 2.642233856893543,
+ "grad_norm": 0.04441308230161667,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.1211161613464355,
+ "step": 190
+ },
+ {
+ "epoch": 2.656195462478185,
+ "grad_norm": 0.03083532303571701,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.094561576843262,
+ "step": 191
+ },
+ {
+ "epoch": 2.670157068062827,
+ "grad_norm": 0.02951296605169773,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.091608047485352,
+ "step": 192
+ },
+ {
+ "epoch": 2.6841186736474696,
+ "grad_norm": 0.02385987527668476,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.087080001831055,
+ "step": 193
+ },
+ {
+ "epoch": 2.6980802792321117,
+ "grad_norm": 0.03315780311822891,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.088132381439209,
+ "step": 194
+ },
+ {
+ "epoch": 2.712041884816754,
+ "grad_norm": 0.02410835586488247,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.081460952758789,
+ "step": 195
+ },
+ {
+ "epoch": 2.7260034904013963,
+ "grad_norm": 0.033069487661123276,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.0696539878845215,
+ "step": 196
+ },
+ {
+ "epoch": 2.7399650959860384,
+ "grad_norm": 0.031288471072912216,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.047612190246582,
+ "step": 197
+ },
+ {
+ "epoch": 2.7539267015706805,
+ "grad_norm": 0.03399677947163582,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.019026756286621,
+ "step": 198
+ },
+ {
+ "epoch": 2.7678883071553226,
+ "grad_norm": 0.031948331743478775,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.050133228302002,
+ "step": 199
+ },
+ {
+ "epoch": 2.781849912739965,
+ "grad_norm": 0.049206435680389404,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.045332431793213,
+ "step": 200
+ },
+ {
+ "epoch": 2.7958115183246073,
+ "grad_norm": 0.07528849691152573,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.059202194213867,
+ "step": 201
+ },
+ {
+ "epoch": 2.80977312390925,
+ "grad_norm": 0.11815470457077026,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.077486038208008,
+ "step": 202
+ },
+ {
+ "epoch": 2.823734729493892,
+ "grad_norm": 0.0867348462343216,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.045670032501221,
+ "step": 203
+ },
+ {
+ "epoch": 2.837696335078534,
+ "grad_norm": 0.05716705322265625,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.034373760223389,
+ "step": 204
+ },
+ {
+ "epoch": 2.851657940663176,
+ "grad_norm": 0.05944284051656723,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.049911975860596,
+ "step": 205
+ },
+ {
+ "epoch": 2.8656195462478182,
+ "grad_norm": 0.03220470994710922,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.039861679077148,
+ "step": 206
+ },
+ {
+ "epoch": 2.8795811518324608,
+ "grad_norm": 0.057081177830696106,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.010714530944824,
+ "step": 207
+ },
+ {
+ "epoch": 2.893542757417103,
+ "grad_norm": 0.04037516191601753,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.018877983093262,
+ "step": 208
+ },
+ {
+ "epoch": 2.9075043630017454,
+ "grad_norm": 0.036492880433797836,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.009493827819824,
+ "step": 209
+ },
+ {
+ "epoch": 2.9214659685863875,
+ "grad_norm": 0.04785023257136345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.0140862464904785,
+ "step": 210
+ },
+ {
+ "epoch": 2.9354275741710296,
+ "grad_norm": 0.07721036672592163,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.030359268188477,
+ "step": 211
+ },
+ {
+ "epoch": 2.9493891797556717,
+ "grad_norm": 0.14220190048217773,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.045989990234375,
+ "step": 212
+ },
+ {
+ "epoch": 2.9633507853403143,
+ "grad_norm": 0.10265158116817474,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.043968200683594,
+ "step": 213
+ },
+ {
+ "epoch": 2.9773123909249564,
+ "grad_norm": 0.051398951560258865,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.022278785705566,
+ "step": 214
+ },
+ {
+ "epoch": 2.9912739965095985,
+ "grad_norm": 0.06044618785381317,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.017296314239502,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.03284750506281853,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 3.7544798851013184,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7531244158744812,
+ "eval_runtime": 60.3986,
+ "eval_samples_per_second": 40.431,
+ "eval_steps_per_second": 0.646,
+ "step": 216
+ },
+ {
+ "epoch": 3.013961605584642,
+ "grad_norm": 0.03895840048789978,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 5.9985857009887695,
+ "step": 217
+ },
+ {
+ "epoch": 3.0279232111692846,
+ "grad_norm": 0.04554992541670799,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 5.984710693359375,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418848167539267,
+ "grad_norm": 0.03359116613864899,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 5.984078407287598,
+ "step": 219
+ },
+ {
+ "epoch": 3.055846422338569,
+ "grad_norm": 0.035171881318092346,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 5.979424476623535,
+ "step": 220
+ },
+ {
+ "epoch": 3.0698080279232114,
+ "grad_norm": 0.032774072140455246,
+ "learning_rate": 0.000598278794687186,
+ "loss": 5.94467830657959,
+ "step": 221
+ },
+ {
+ "epoch": 3.0837696335078535,
+ "grad_norm": 0.032842908054590225,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 5.95986270904541,
+ "step": 222
+ },
+ {
+ "epoch": 3.0977312390924956,
+ "grad_norm": 0.02870488539338112,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 5.951186656951904,
+ "step": 223
+ },
+ {
+ "epoch": 3.1116928446771377,
+ "grad_norm": 0.030777374282479286,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 5.942493438720703,
+ "step": 224
+ },
+ {
+ "epoch": 3.1256544502617802,
+ "grad_norm": 0.026439018547534943,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 5.938275337219238,
+ "step": 225
+ },
+ {
+ "epoch": 3.1396160558464223,
+ "grad_norm": 0.028689472004771233,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 5.931982040405273,
+ "step": 226
+ },
+ {
+ "epoch": 3.1535776614310644,
+ "grad_norm": 0.028100362047553062,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 5.919692039489746,
+ "step": 227
+ },
+ {
+ "epoch": 3.167539267015707,
+ "grad_norm": 0.02407177910208702,
+ "learning_rate": 0.00059807235435179,
+ "loss": 5.909029006958008,
+ "step": 228
+ },
+ {
+ "epoch": 3.181500872600349,
+ "grad_norm": 0.0237748846411705,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 5.9131598472595215,
+ "step": 229
+ },
+ {
+ "epoch": 3.195462478184991,
+ "grad_norm": 0.022418556734919548,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 5.909396171569824,
+ "step": 230
+ },
+ {
+ "epoch": 3.2094240837696333,
+ "grad_norm": 0.024656571447849274,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 5.885243892669678,
+ "step": 231
+ },
+ {
+ "epoch": 3.223385689354276,
+ "grad_norm": 0.02136118710041046,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 5.898958206176758,
+ "step": 232
+ },
+ {
+ "epoch": 3.237347294938918,
+ "grad_norm": 0.025290116667747498,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 5.886986255645752,
+ "step": 233
+ },
+ {
+ "epoch": 3.25130890052356,
+ "grad_norm": 0.02656404674053192,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 5.897647857666016,
+ "step": 234
+ },
+ {
+ "epoch": 3.2652705061082026,
+ "grad_norm": 0.019188793376088142,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 5.875659942626953,
+ "step": 235
+ },
+ {
+ "epoch": 3.2792321116928447,
+ "grad_norm": 0.02308926172554493,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 5.859380722045898,
+ "step": 236
+ },
+ {
+ "epoch": 3.2931937172774868,
+ "grad_norm": 0.029775725677609444,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 5.873961448669434,
+ "step": 237
+ },
+ {
+ "epoch": 3.3071553228621293,
+ "grad_norm": 0.03875259310007095,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 5.863171100616455,
+ "step": 238
+ },
+ {
+ "epoch": 3.3211169284467714,
+ "grad_norm": 0.052446842193603516,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 5.886547088623047,
+ "step": 239
+ },
+ {
+ "epoch": 3.3350785340314135,
+ "grad_norm": 0.05583837628364563,
+ "learning_rate": 0.000597691359743243,
+ "loss": 5.8849287033081055,
+ "step": 240
+ },
+ {
+ "epoch": 3.349040139616056,
+ "grad_norm": 0.034729067236185074,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 5.863419532775879,
+ "step": 241
+ },
+ {
+ "epoch": 3.363001745200698,
+ "grad_norm": 0.03590013459324837,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 5.852169990539551,
+ "step": 242
+ },
+ {
+ "epoch": 3.3769633507853403,
+ "grad_norm": 0.035041507333517075,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 5.836690902709961,
+ "step": 243
+ },
+ {
+ "epoch": 3.3909249563699824,
+ "grad_norm": 0.02857939340174198,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 5.836485862731934,
+ "step": 244
+ },
+ {
+ "epoch": 3.404886561954625,
+ "grad_norm": 0.03176642581820488,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 5.860719203948975,
+ "step": 245
+ },
+ {
+ "epoch": 3.418848167539267,
+ "grad_norm": 0.026105748489499092,
+ "learning_rate": 0.000597488038559129,
+ "loss": 5.8354949951171875,
+ "step": 246
+ },
+ {
+ "epoch": 3.432809773123909,
+ "grad_norm": 0.029094981029629707,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 5.846247673034668,
+ "step": 247
+ },
+ {
+ "epoch": 3.4467713787085517,
+ "grad_norm": 0.02388785220682621,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 5.837607383728027,
+ "step": 248
+ },
+ {
+ "epoch": 3.4607329842931938,
+ "grad_norm": 0.03094472736120224,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 5.832090377807617,
+ "step": 249
+ },
+ {
+ "epoch": 3.474694589877836,
+ "grad_norm": 0.03227658197283745,
+ "learning_rate": 0.000597347745466109,
+ "loss": 5.83722448348999,
+ "step": 250
+ },
+ {
+ "epoch": 3.488656195462478,
+ "grad_norm": 0.042813923209905624,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 5.851722717285156,
+ "step": 251
+ },
+ {
+ "epoch": 3.5026178010471205,
+ "grad_norm": 0.05825525149703026,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.795719146728516,
+ "step": 252
+ },
+ {
+ "epoch": 3.5165794066317626,
+ "grad_norm": 0.0867447778582573,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.813750267028809,
+ "step": 253
+ },
+ {
+ "epoch": 3.530541012216405,
+ "grad_norm": 0.09192710369825363,
+ "learning_rate": 0.000597203657966319,
+ "loss": 5.823488235473633,
+ "step": 254
+ },
+ {
+ "epoch": 3.5445026178010473,
+ "grad_norm": 0.06069248169660568,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 5.814342021942139,
+ "step": 255
+ },
+ {
+ "epoch": 3.5584642233856894,
+ "grad_norm": 0.06259952485561371,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 5.834272384643555,
+ "step": 256
+ },
+ {
+ "epoch": 3.5724258289703315,
+ "grad_norm": 0.0513639897108078,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.807951927185059,
+ "step": 257
+ },
+ {
+ "epoch": 3.5863874345549736,
+ "grad_norm": 0.03327798470854759,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.790238857269287,
+ "step": 258
+ },
+ {
+ "epoch": 3.600349040139616,
+ "grad_norm": 0.03963128477334976,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.7956695556640625,
+ "step": 259
+ },
+ {
+ "epoch": 3.614310645724258,
+ "grad_norm": 0.031153453513979912,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 5.7874531745910645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6282722513089007,
+ "grad_norm": 0.031419433653354645,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 5.779650688171387,
+ "step": 261
+ },
+ {
+ "epoch": 3.642233856893543,
+ "grad_norm": 0.0288684144616127,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.776577949523926,
+ "step": 262
+ },
+ {
+ "epoch": 3.656195462478185,
+ "grad_norm": 0.028315918520092964,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.778571128845215,
+ "step": 263
+ },
+ {
+ "epoch": 3.670157068062827,
+ "grad_norm": 0.024977795779705048,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 5.776470184326172,
+ "step": 264
+ },
+ {
+ "epoch": 3.6841186736474696,
+ "grad_norm": 0.026822632178664207,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.755108833312988,
+ "step": 265
+ },
+ {
+ "epoch": 3.6980802792321117,
+ "grad_norm": 0.02738630585372448,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.74041748046875,
+ "step": 266
+ },
+ {
+ "epoch": 3.712041884816754,
+ "grad_norm": 0.02361280657351017,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.764027118682861,
+ "step": 267
+ },
+ {
+ "epoch": 3.7260034904013963,
+ "grad_norm": 0.029926009476184845,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.780390739440918,
+ "step": 268
+ },
+ {
+ "epoch": 3.7399650959860384,
+ "grad_norm": 0.03519507870078087,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.758105278015137,
+ "step": 269
+ },
+ {
+ "epoch": 3.7539267015706805,
+ "grad_norm": 0.030114633962512016,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.757230758666992,
+ "step": 270
+ },
+ {
+ "epoch": 3.7678883071553226,
+ "grad_norm": 0.02467748522758484,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.7615461349487305,
+ "step": 271
+ },
+ {
+ "epoch": 3.781849912739965,
+ "grad_norm": 0.03425183147192001,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.749448776245117,
+ "step": 272
+ },
+ {
+ "epoch": 3.7958115183246073,
+ "grad_norm": 0.046307582408189774,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.7356977462768555,
+ "step": 273
+ },
+ {
+ "epoch": 3.80977312390925,
+ "grad_norm": 0.054102856665849686,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.741869926452637,
+ "step": 274
+ },
+ {
+ "epoch": 3.823734729493892,
+ "grad_norm": 0.07026689499616623,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.721103668212891,
+ "step": 275
+ },
+ {
+ "epoch": 3.837696335078534,
+ "grad_norm": 0.0733039602637291,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.731414794921875,
+ "step": 276
+ },
+ {
+ "epoch": 3.851657940663176,
+ "grad_norm": 0.05342920497059822,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.734922885894775,
+ "step": 277
+ },
+ {
+ "epoch": 3.8656195462478182,
+ "grad_norm": 0.0504680834710598,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.732746124267578,
+ "step": 278
+ },
+ {
+ "epoch": 3.8795811518324608,
+ "grad_norm": 0.038991592824459076,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.719473838806152,
+ "step": 279
+ },
+ {
+ "epoch": 3.893542757417103,
+ "grad_norm": 0.04434404894709587,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.709481239318848,
+ "step": 280
+ },
+ {
+ "epoch": 3.9075043630017454,
+ "grad_norm": 0.047285184264183044,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.693531036376953,
+ "step": 281
+ },
+ {
+ "epoch": 3.9214659685863875,
+ "grad_norm": 0.045081138610839844,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.728959083557129,
+ "step": 282
+ },
+ {
+ "epoch": 3.9354275741710296,
+ "grad_norm": 0.025133663788437843,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.719696998596191,
+ "step": 283
+ },
+ {
+ "epoch": 3.9493891797556717,
+ "grad_norm": 0.03353596478700638,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.700235366821289,
+ "step": 284
+ },
+ {
+ "epoch": 3.9633507853403143,
+ "grad_norm": 0.035162411630153656,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.684836387634277,
+ "step": 285
+ },
+ {
+ "epoch": 3.9773123909249564,
+ "grad_norm": 0.04025450348854065,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.692097187042236,
+ "step": 286
+ },
+ {
+ "epoch": 3.9912739965095985,
+ "grad_norm": 0.04995397478342056,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.6766862869262695,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.031619708985090256,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 3.5519747734069824,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.714687705039978,
+ "eval_runtime": 59.637,
+ "eval_samples_per_second": 40.948,
+ "eval_steps_per_second": 0.654,
+ "step": 288
+ },
+ {
+ "epoch": 4.013961605584642,
+ "grad_norm": 0.036328598856925964,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.666294097900391,
+ "step": 289
+ },
+ {
+ "epoch": 4.027923211169284,
+ "grad_norm": 0.07326336950063705,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.6879167556762695,
+ "step": 290
+ },
+ {
+ "epoch": 4.041884816753926,
+ "grad_norm": 0.07947526127099991,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.674883842468262,
+ "step": 291
+ },
+ {
+ "epoch": 4.055846422338569,
+ "grad_norm": 0.04599885269999504,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.688086986541748,
+ "step": 292
+ },
+ {
+ "epoch": 4.069808027923211,
+ "grad_norm": 0.0465150885283947,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.649483680725098,
+ "step": 293
+ },
+ {
+ "epoch": 4.0837696335078535,
+ "grad_norm": 0.038542453199625015,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.653055667877197,
+ "step": 294
+ },
+ {
+ "epoch": 4.097731239092496,
+ "grad_norm": 0.03659164905548096,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.65955924987793,
+ "step": 295
+ },
+ {
+ "epoch": 4.111692844677138,
+ "grad_norm": 0.03824038803577423,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.650853157043457,
+ "step": 296
+ },
+ {
+ "epoch": 4.12565445026178,
+ "grad_norm": 0.03591204807162285,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.643499374389648,
+ "step": 297
+ },
+ {
+ "epoch": 4.139616055846423,
+ "grad_norm": 0.03269682824611664,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.657600402832031,
+ "step": 298
+ },
+ {
+ "epoch": 4.153577661431065,
+ "grad_norm": 0.03343537449836731,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.6352081298828125,
+ "step": 299
+ },
+ {
+ "epoch": 4.167539267015707,
+ "grad_norm": 0.034283801913261414,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.626757621765137,
+ "step": 300
+ },
+ {
+ "epoch": 4.181500872600349,
+ "grad_norm": 0.034614723175764084,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.6316680908203125,
+ "step": 301
+ },
+ {
+ "epoch": 4.195462478184991,
+ "grad_norm": 0.0340128131210804,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.648160934448242,
+ "step": 302
+ },
+ {
+ "epoch": 4.209424083769633,
+ "grad_norm": 0.03068280965089798,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.619738578796387,
+ "step": 303
+ },
+ {
+ "epoch": 4.223385689354275,
+ "grad_norm": 0.030269058421254158,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.609238624572754,
+ "step": 304
+ },
+ {
+ "epoch": 4.237347294938918,
+ "grad_norm": 0.03612440824508667,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.602194786071777,
+ "step": 305
+ },
+ {
+ "epoch": 4.2513089005235605,
+ "grad_norm": 0.044292811304330826,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.61159610748291,
+ "step": 306
+ },
+ {
+ "epoch": 4.265270506108203,
+ "grad_norm": 0.04449291527271271,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.6274237632751465,
+ "step": 307
+ },
+ {
+ "epoch": 4.279232111692845,
+ "grad_norm": 0.05014493316411972,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.589034080505371,
+ "step": 308
+ },
+ {
+ "epoch": 4.293193717277487,
+ "grad_norm": 0.05286206305027008,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.600770950317383,
+ "step": 309
+ },
+ {
+ "epoch": 4.307155322862129,
+ "grad_norm": 0.048459846526384354,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.5796098709106445,
+ "step": 310
+ },
+ {
+ "epoch": 4.321116928446771,
+ "grad_norm": 0.05329663306474686,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.590407371520996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335078534031414,
+ "grad_norm": 0.052595119923353195,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.583693504333496,
+ "step": 312
+ },
+ {
+ "epoch": 4.349040139616056,
+ "grad_norm": 0.05164601653814316,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.603281021118164,
+ "step": 313
+ },
+ {
+ "epoch": 4.363001745200698,
+ "grad_norm": 0.04657137766480446,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.591344833374023,
+ "step": 314
+ },
+ {
+ "epoch": 4.37696335078534,
+ "grad_norm": 0.03804599866271019,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.580578804016113,
+ "step": 315
+ },
+ {
+ "epoch": 4.390924956369982,
+ "grad_norm": 0.03835320845246315,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.597836971282959,
+ "step": 316
+ },
+ {
+ "epoch": 4.4048865619546245,
+ "grad_norm": 0.034059133380651474,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.567502021789551,
+ "step": 317
+ },
+ {
+ "epoch": 4.418848167539267,
+ "grad_norm": 0.03350386768579483,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.552022933959961,
+ "step": 318
+ },
+ {
+ "epoch": 4.43280977312391,
+ "grad_norm": 0.032574303448200226,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.559177875518799,
+ "step": 319
+ },
+ {
+ "epoch": 4.446771378708552,
+ "grad_norm": 0.030893854796886444,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.549288272857666,
+ "step": 320
+ },
+ {
+ "epoch": 4.460732984293194,
+ "grad_norm": 0.03255036100745201,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.5400390625,
+ "step": 321
+ },
+ {
+ "epoch": 4.474694589877836,
+ "grad_norm": 0.03009258769452572,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.5328474044799805,
+ "step": 322
+ },
+ {
+ "epoch": 4.488656195462478,
+ "grad_norm": 0.03178320452570915,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.5594482421875,
+ "step": 323
+ },
+ {
+ "epoch": 4.50261780104712,
+ "grad_norm": 0.03360465541481972,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.5402727127075195,
+ "step": 324
+ },
+ {
+ "epoch": 4.516579406631763,
+ "grad_norm": 0.04275049269199371,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.534474849700928,
+ "step": 325
+ },
+ {
+ "epoch": 4.530541012216405,
+ "grad_norm": 0.04140641540288925,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.5324859619140625,
+ "step": 326
+ },
+ {
+ "epoch": 4.544502617801047,
+ "grad_norm": 0.04769204929471016,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.4919352531433105,
+ "step": 327
+ },
+ {
+ "epoch": 4.558464223385689,
+ "grad_norm": 0.04402285814285278,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.520858287811279,
+ "step": 328
+ },
+ {
+ "epoch": 4.5724258289703315,
+ "grad_norm": 0.038526229560375214,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.523778438568115,
+ "step": 329
+ },
+ {
+ "epoch": 4.5863874345549736,
+ "grad_norm": 0.05422377213835716,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.517973899841309,
+ "step": 330
+ },
+ {
+ "epoch": 4.600349040139616,
+ "grad_norm": 0.07236721366643906,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.548083782196045,
+ "step": 331
+ },
+ {
+ "epoch": 4.614310645724259,
+ "grad_norm": 0.06861061602830887,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.535577774047852,
+ "step": 332
+ },
+ {
+ "epoch": 4.628272251308901,
+ "grad_norm": 0.07148340344429016,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.54155158996582,
+ "step": 333
+ },
+ {
+ "epoch": 4.642233856893543,
+ "grad_norm": 0.07838302850723267,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.552852630615234,
+ "step": 334
+ },
+ {
+ "epoch": 4.656195462478185,
+ "grad_norm": 0.10946758091449738,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.555066108703613,
+ "step": 335
+ },
+ {
+ "epoch": 4.670157068062827,
+ "grad_norm": 0.10143530368804932,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.584932804107666,
+ "step": 336
+ },
+ {
+ "epoch": 4.684118673647469,
+ "grad_norm": 0.08926606178283691,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.570462226867676,
+ "step": 337
+ },
+ {
+ "epoch": 4.698080279232112,
+ "grad_norm": 0.07060826569795609,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.5716753005981445,
+ "step": 338
+ },
+ {
+ "epoch": 4.712041884816754,
+ "grad_norm": 0.0418456569314003,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.556496620178223,
+ "step": 339
+ },
+ {
+ "epoch": 4.726003490401396,
+ "grad_norm": 0.05952411890029907,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.567801475524902,
+ "step": 340
+ },
+ {
+ "epoch": 4.739965095986038,
+ "grad_norm": 0.042999885976314545,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.537403106689453,
+ "step": 341
+ },
+ {
+ "epoch": 4.7539267015706805,
+ "grad_norm": 0.04426449537277222,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.552387237548828,
+ "step": 342
+ },
+ {
+ "epoch": 4.767888307155323,
+ "grad_norm": 0.040835145860910416,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.53051233291626,
+ "step": 343
+ },
+ {
+ "epoch": 4.781849912739965,
+ "grad_norm": 0.03819291293621063,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.531174659729004,
+ "step": 344
+ },
+ {
+ "epoch": 4.795811518324607,
+ "grad_norm": 0.030088599771261215,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.499328136444092,
+ "step": 345
+ },
+ {
+ "epoch": 4.80977312390925,
+ "grad_norm": 0.029984362423419952,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.504540920257568,
+ "step": 346
+ },
+ {
+ "epoch": 4.823734729493892,
+ "grad_norm": 0.027402915060520172,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.50899076461792,
+ "step": 347
+ },
+ {
+ "epoch": 4.837696335078534,
+ "grad_norm": 0.025096213445067406,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.480479717254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.851657940663176,
+ "grad_norm": 0.026828132569789886,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.501729965209961,
+ "step": 349
+ },
+ {
+ "epoch": 4.865619546247818,
+ "grad_norm": 0.023715168237686157,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.466860771179199,
+ "step": 350
+ },
+ {
+ "epoch": 4.879581151832461,
+ "grad_norm": 0.024503640830516815,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.491827964782715,
+ "step": 351
+ },
+ {
+ "epoch": 4.893542757417103,
+ "grad_norm": 0.020891105756163597,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.478384971618652,
+ "step": 352
+ },
+ {
+ "epoch": 4.907504363001745,
+ "grad_norm": 0.021373765543103218,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.4612135887146,
+ "step": 353
+ },
+ {
+ "epoch": 4.9214659685863875,
+ "grad_norm": 0.021773340180516243,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.4547319412231445,
+ "step": 354
+ },
+ {
+ "epoch": 4.93542757417103,
+ "grad_norm": 0.019637586548924446,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.431741714477539,
+ "step": 355
+ },
+ {
+ "epoch": 4.949389179755672,
+ "grad_norm": 0.023108696565032005,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.448612213134766,
+ "step": 356
+ },
+ {
+ "epoch": 4.963350785340314,
+ "grad_norm": 0.021935103461146355,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.447112083435059,
+ "step": 357
+ },
+ {
+ "epoch": 4.977312390924956,
+ "grad_norm": 0.023260753601789474,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.436091423034668,
+ "step": 358
+ },
+ {
+ "epoch": 4.991273996509599,
+ "grad_norm": 0.021355057135224342,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.408421993255615,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.014992325566709042,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 3.4004223346710205,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6834986805915833,
+ "eval_runtime": 60.4148,
+ "eval_samples_per_second": 40.421,
+ "eval_steps_per_second": 0.646,
+ "step": 360
+ },
+ {
+ "epoch": 5.013961605584642,
+ "grad_norm": 0.02471218630671501,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.396238803863525,
+ "step": 361
+ },
+ {
+ "epoch": 5.027923211169284,
+ "grad_norm": 0.027672411873936653,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.4048919677734375,
+ "step": 362
+ },
+ {
+ "epoch": 5.041884816753926,
+ "grad_norm": 0.031560689210891724,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.399665832519531,
+ "step": 363
+ },
+ {
+ "epoch": 5.055846422338569,
+ "grad_norm": 0.03459061309695244,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.396360397338867,
+ "step": 364
+ },
+ {
+ "epoch": 5.069808027923211,
+ "grad_norm": 0.04040960222482681,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.405309677124023,
+ "step": 365
+ },
+ {
+ "epoch": 5.0837696335078535,
+ "grad_norm": 0.05444670096039772,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.408894062042236,
+ "step": 366
+ },
+ {
+ "epoch": 5.097731239092496,
+ "grad_norm": 0.05816745012998581,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.403566360473633,
+ "step": 367
+ },
+ {
+ "epoch": 5.111692844677138,
+ "grad_norm": 0.06330069154500961,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.395961761474609,
+ "step": 368
+ },
+ {
+ "epoch": 5.12565445026178,
+ "grad_norm": 0.07458917796611786,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.400968551635742,
+ "step": 369
+ },
+ {
+ "epoch": 5.139616055846423,
+ "grad_norm": 0.06231356039643288,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.412624359130859,
+ "step": 370
+ },
+ {
+ "epoch": 5.153577661431065,
+ "grad_norm": 0.05212346464395523,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.394794464111328,
+ "step": 371
+ },
+ {
+ "epoch": 5.167539267015707,
+ "grad_norm": 0.050003375858068466,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.380125045776367,
+ "step": 372
+ },
+ {
+ "epoch": 5.181500872600349,
+ "grad_norm": 0.048393815755844116,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.390806198120117,
+ "step": 373
+ },
+ {
+ "epoch": 5.195462478184991,
+ "grad_norm": 0.04200794920325279,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.387467384338379,
+ "step": 374
+ },
+ {
+ "epoch": 5.209424083769633,
+ "grad_norm": 0.04931448772549629,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.37552547454834,
+ "step": 375
+ },
+ {
+ "epoch": 5.223385689354275,
+ "grad_norm": 0.03851857781410217,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.373126983642578,
+ "step": 376
+ },
+ {
+ "epoch": 5.237347294938918,
+ "grad_norm": 0.03458717465400696,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.346826553344727,
+ "step": 377
+ },
+ {
+ "epoch": 5.2513089005235605,
+ "grad_norm": 0.031107677146792412,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.353341102600098,
+ "step": 378
+ },
+ {
+ "epoch": 5.265270506108203,
+ "grad_norm": 0.03077058307826519,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.358076095581055,
+ "step": 379
+ },
+ {
+ "epoch": 5.279232111692845,
+ "grad_norm": 0.028170401230454445,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.354019641876221,
+ "step": 380
+ },
+ {
+ "epoch": 5.293193717277487,
+ "grad_norm": 0.02871990203857422,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.368258476257324,
+ "step": 381
+ },
+ {
+ "epoch": 5.307155322862129,
+ "grad_norm": 0.03046584501862526,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.342505931854248,
+ "step": 382
+ },
+ {
+ "epoch": 5.321116928446771,
+ "grad_norm": 0.02781563811004162,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.339411735534668,
+ "step": 383
+ },
+ {
+ "epoch": 5.335078534031414,
+ "grad_norm": 0.028825288638472557,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.341948509216309,
+ "step": 384
+ },
+ {
+ "epoch": 5.349040139616056,
+ "grad_norm": 0.026684079319238663,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.340881824493408,
+ "step": 385
+ },
+ {
+ "epoch": 5.363001745200698,
+ "grad_norm": 0.02429167740046978,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.342733383178711,
+ "step": 386
+ },
+ {
+ "epoch": 5.37696335078534,
+ "grad_norm": 0.02855239436030388,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.313878536224365,
+ "step": 387
+ },
+ {
+ "epoch": 5.390924956369982,
+ "grad_norm": 0.029651548713445663,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.312665939331055,
+ "step": 388
+ },
+ {
+ "epoch": 5.4048865619546245,
+ "grad_norm": 0.027191132307052612,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3086838722229,
+ "step": 389
+ },
+ {
+ "epoch": 5.418848167539267,
+ "grad_norm": 0.026719525456428528,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.305359840393066,
+ "step": 390
+ },
+ {
+ "epoch": 5.43280977312391,
+ "grad_norm": 0.026763100177049637,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.3164381980896,
+ "step": 391
+ },
+ {
+ "epoch": 5.446771378708552,
+ "grad_norm": 0.03125109151005745,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.3083953857421875,
+ "step": 392
+ },
+ {
+ "epoch": 5.460732984293194,
+ "grad_norm": 0.0394827201962471,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.275996208190918,
+ "step": 393
+ },
+ {
+ "epoch": 5.474694589877836,
+ "grad_norm": 0.040579475462436676,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.292995929718018,
+ "step": 394
+ },
+ {
+ "epoch": 5.488656195462478,
+ "grad_norm": 0.037223003804683685,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.298526763916016,
+ "step": 395
+ },
+ {
+ "epoch": 5.50261780104712,
+ "grad_norm": 0.04207765311002731,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.287985801696777,
+ "step": 396
+ },
+ {
+ "epoch": 5.516579406631763,
+ "grad_norm": 0.048299964517354965,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.244608402252197,
+ "step": 397
+ },
+ {
+ "epoch": 5.530541012216405,
+ "grad_norm": 0.04650307446718216,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.285453796386719,
+ "step": 398
+ },
+ {
+ "epoch": 5.544502617801047,
+ "grad_norm": 0.045339711010456085,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.312187671661377,
+ "step": 399
+ },
+ {
+ "epoch": 5.558464223385689,
+ "grad_norm": 0.041067708283662796,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.292692184448242,
+ "step": 400
+ },
+ {
+ "epoch": 5.5724258289703315,
+ "grad_norm": 0.05455026403069496,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.253364562988281,
+ "step": 401
+ },
+ {
+ "epoch": 5.5863874345549736,
+ "grad_norm": 0.06660915911197662,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.262646675109863,
+ "step": 402
+ },
+ {
+ "epoch": 5.600349040139616,
+ "grad_norm": 0.06634793430566788,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.274322986602783,
+ "step": 403
+ },
+ {
+ "epoch": 5.614310645724259,
+ "grad_norm": 0.08996330946683884,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.287951469421387,
+ "step": 404
+ },
+ {
+ "epoch": 5.628272251308901,
+ "grad_norm": 0.08526470512151718,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.3007097244262695,
+ "step": 405
+ },
+ {
+ "epoch": 5.642233856893543,
+ "grad_norm": 0.08766836673021317,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.332538604736328,
+ "step": 406
+ },
+ {
+ "epoch": 5.656195462478185,
+ "grad_norm": 0.07055048644542694,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.296169281005859,
+ "step": 407
+ },
+ {
+ "epoch": 5.670157068062827,
+ "grad_norm": 0.07520575076341629,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.309683322906494,
+ "step": 408
+ },
+ {
+ "epoch": 5.684118673647469,
+ "grad_norm": 0.06115756183862686,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.311301231384277,
+ "step": 409
+ },
+ {
+ "epoch": 5.698080279232112,
+ "grad_norm": 0.05862593650817871,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.28066349029541,
+ "step": 410
+ },
+ {
+ "epoch": 5.712041884816754,
+ "grad_norm": 0.06169601157307625,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.286840438842773,
+ "step": 411
+ },
+ {
+ "epoch": 5.726003490401396,
+ "grad_norm": 0.05270719155669212,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.302594184875488,
+ "step": 412
+ },
+ {
+ "epoch": 5.739965095986038,
+ "grad_norm": 0.03912634402513504,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.262563705444336,
+ "step": 413
+ },
+ {
+ "epoch": 5.7539267015706805,
+ "grad_norm": 0.03215811401605606,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.269310474395752,
+ "step": 414
+ },
+ {
+ "epoch": 5.767888307155323,
+ "grad_norm": 0.03236516937613487,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.268766403198242,
+ "step": 415
+ },
+ {
+ "epoch": 5.781849912739965,
+ "grad_norm": 0.027429452165961266,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.283801555633545,
+ "step": 416
+ },
+ {
+ "epoch": 5.795811518324607,
+ "grad_norm": 0.02990538813173771,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.22291374206543,
+ "step": 417
+ },
+ {
+ "epoch": 5.80977312390925,
+ "grad_norm": 0.030410146340727806,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.2281599044799805,
+ "step": 418
+ },
+ {
+ "epoch": 5.823734729493892,
+ "grad_norm": 0.030068187043070793,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.250294208526611,
+ "step": 419
+ },
+ {
+ "epoch": 5.837696335078534,
+ "grad_norm": 0.02800513058900833,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.226508617401123,
+ "step": 420
+ },
+ {
+ "epoch": 5.851657940663176,
+ "grad_norm": 0.023823332041502,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.247838020324707,
+ "step": 421
+ },
+ {
+ "epoch": 5.865619546247818,
+ "grad_norm": 0.024974249303340912,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.2392473220825195,
+ "step": 422
+ },
+ {
+ "epoch": 5.879581151832461,
+ "grad_norm": 0.023320546373724937,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.209543704986572,
+ "step": 423
+ },
+ {
+ "epoch": 5.893542757417103,
+ "grad_norm": 0.024201491847634315,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.219725608825684,
+ "step": 424
+ },
+ {
+ "epoch": 5.907504363001745,
+ "grad_norm": 0.021649666130542755,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.203141212463379,
+ "step": 425
+ },
+ {
+ "epoch": 5.9214659685863875,
+ "grad_norm": 0.0248002577573061,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.179058074951172,
+ "step": 426
+ },
+ {
+ "epoch": 5.93542757417103,
+ "grad_norm": 0.028549587354063988,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.202126502990723,
+ "step": 427
+ },
+ {
+ "epoch": 5.949389179755672,
+ "grad_norm": 0.0331057533621788,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.208107948303223,
+ "step": 428
+ },
+ {
+ "epoch": 5.963350785340314,
+ "grad_norm": 0.031843364238739014,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.184683799743652,
+ "step": 429
+ },
+ {
+ "epoch": 5.977312390924956,
+ "grad_norm": 0.024927377700805664,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.180486679077148,
+ "step": 430
+ },
+ {
+ "epoch": 5.991273996509599,
+ "grad_norm": 0.026685267686843872,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.186746597290039,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.018545351922512054,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 3.217043876647949,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.653471827507019,
+ "eval_runtime": 61.1586,
+ "eval_samples_per_second": 39.929,
+ "eval_steps_per_second": 0.638,
+ "step": 432
+ },
+ {
+ "epoch": 6.013961605584642,
+ "grad_norm": 0.025291118770837784,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.156534194946289,
+ "step": 433
+ },
+ {
+ "epoch": 6.027923211169284,
+ "grad_norm": 0.03774752840399742,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.144041538238525,
+ "step": 434
+ },
+ {
+ "epoch": 6.041884816753926,
+ "grad_norm": 0.04244692996144295,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.158862113952637,
+ "step": 435
+ },
+ {
+ "epoch": 6.055846422338569,
+ "grad_norm": 0.039782389998435974,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.166165351867676,
+ "step": 436
+ },
+ {
+ "epoch": 6.069808027923211,
+ "grad_norm": 0.05489896610379219,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.155766487121582,
+ "step": 437
+ },
+ {
+ "epoch": 6.0837696335078535,
+ "grad_norm": 0.059844326227903366,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.153733253479004,
+ "step": 438
+ },
+ {
+ "epoch": 6.097731239092496,
+ "grad_norm": 0.05539938807487488,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.130112648010254,
+ "step": 439
+ },
+ {
+ "epoch": 6.111692844677138,
+ "grad_norm": 0.05283183231949806,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.136910915374756,
+ "step": 440
+ },
+ {
+ "epoch": 6.12565445026178,
+ "grad_norm": 0.054407261312007904,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.165216445922852,
+ "step": 441
+ },
+ {
+ "epoch": 6.139616055846423,
+ "grad_norm": 0.0509788915514946,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.140888214111328,
+ "step": 442
+ },
+ {
+ "epoch": 6.153577661431065,
+ "grad_norm": 0.05018748342990875,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.13115930557251,
+ "step": 443
+ },
+ {
+ "epoch": 6.167539267015707,
+ "grad_norm": 0.047012437134981155,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.146507263183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181500872600349,
+ "grad_norm": 0.04943335801362991,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.135608673095703,
+ "step": 445
+ },
+ {
+ "epoch": 6.195462478184991,
+ "grad_norm": 0.05304260924458504,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.136394023895264,
+ "step": 446
+ },
+ {
+ "epoch": 6.209424083769633,
+ "grad_norm": 0.04554047808051109,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.124661922454834,
+ "step": 447
+ },
+ {
+ "epoch": 6.223385689354275,
+ "grad_norm": 0.040672432631254196,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.129773139953613,
+ "step": 448
+ },
+ {
+ "epoch": 6.237347294938918,
+ "grad_norm": 0.03442879393696785,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.128028392791748,
+ "step": 449
+ },
+ {
+ "epoch": 6.2513089005235605,
+ "grad_norm": 0.03927934914827347,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.092485427856445,
+ "step": 450
+ },
+ {
+ "epoch": 6.265270506108203,
+ "grad_norm": 0.029380977153778076,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.077669143676758,
+ "step": 451
+ },
+ {
+ "epoch": 6.279232111692845,
+ "grad_norm": 0.03331821784377098,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.138857841491699,
+ "step": 452
+ },
+ {
+ "epoch": 6.293193717277487,
+ "grad_norm": 0.02939118817448616,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.082937240600586,
+ "step": 453
+ },
+ {
+ "epoch": 6.307155322862129,
+ "grad_norm": 0.02730412967503071,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.093183517456055,
+ "step": 454
+ },
+ {
+ "epoch": 6.321116928446771,
+ "grad_norm": 0.02964780665934086,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.1028151512146,
+ "step": 455
+ },
+ {
+ "epoch": 6.335078534031414,
+ "grad_norm": 0.026350129395723343,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.066362380981445,
+ "step": 456
+ },
+ {
+ "epoch": 6.349040139616056,
+ "grad_norm": 0.027010295540094376,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.070037841796875,
+ "step": 457
+ },
+ {
+ "epoch": 6.363001745200698,
+ "grad_norm": 0.02906128577888012,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.107708930969238,
+ "step": 458
+ },
+ {
+ "epoch": 6.37696335078534,
+ "grad_norm": 0.028619443997740746,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.073182582855225,
+ "step": 459
+ },
+ {
+ "epoch": 6.390924956369982,
+ "grad_norm": 0.029981834813952446,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.076602935791016,
+ "step": 460
+ },
+ {
+ "epoch": 6.4048865619546245,
+ "grad_norm": 0.03255423903465271,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.063301086425781,
+ "step": 461
+ },
+ {
+ "epoch": 6.418848167539267,
+ "grad_norm": 0.03863866627216339,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.041812896728516,
+ "step": 462
+ },
+ {
+ "epoch": 6.43280977312391,
+ "grad_norm": 0.03860306739807129,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.06298828125,
+ "step": 463
+ },
+ {
+ "epoch": 6.446771378708552,
+ "grad_norm": 0.03378435596823692,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.0434370040893555,
+ "step": 464
+ },
+ {
+ "epoch": 6.460732984293194,
+ "grad_norm": 0.03276759013533592,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.036187171936035,
+ "step": 465
+ },
+ {
+ "epoch": 6.474694589877836,
+ "grad_norm": 0.04130513221025467,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.039098739624023,
+ "step": 466
+ },
+ {
+ "epoch": 6.488656195462478,
+ "grad_norm": 0.05389940366148949,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.062798976898193,
+ "step": 467
+ },
+ {
+ "epoch": 6.50261780104712,
+ "grad_norm": 0.05881704390048981,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.067028522491455,
+ "step": 468
+ },
+ {
+ "epoch": 6.516579406631763,
+ "grad_norm": 0.061897799372673035,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.056947231292725,
+ "step": 469
+ },
+ {
+ "epoch": 6.530541012216405,
+ "grad_norm": 0.06549323350191116,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.089262008666992,
+ "step": 470
+ },
+ {
+ "epoch": 6.544502617801047,
+ "grad_norm": 0.048082590103149414,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.055979251861572,
+ "step": 471
+ },
+ {
+ "epoch": 6.558464223385689,
+ "grad_norm": 0.03768027946352959,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.0660600662231445,
+ "step": 472
+ },
+ {
+ "epoch": 6.5724258289703315,
+ "grad_norm": 0.043922603130340576,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 5.036745071411133,
+ "step": 473
+ },
+ {
+ "epoch": 6.5863874345549736,
+ "grad_norm": 0.044848836958408356,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.073650360107422,
+ "step": 474
+ },
+ {
+ "epoch": 6.600349040139616,
+ "grad_norm": 0.05841026455163956,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.053045272827148,
+ "step": 475
+ },
+ {
+ "epoch": 6.614310645724259,
+ "grad_norm": 0.0650017261505127,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 5.052206039428711,
+ "step": 476
+ },
+ {
+ "epoch": 6.628272251308901,
+ "grad_norm": 0.048979002982378006,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 5.054277420043945,
+ "step": 477
+ },
+ {
+ "epoch": 6.642233856893543,
+ "grad_norm": 0.05085799843072891,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 5.063844680786133,
+ "step": 478
+ },
+ {
+ "epoch": 6.656195462478185,
+ "grad_norm": 0.04336017370223999,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 5.045474052429199,
+ "step": 479
+ },
+ {
+ "epoch": 6.670157068062827,
+ "grad_norm": 0.04113380238413811,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 5.031338214874268,
+ "step": 480
+ },
+ {
+ "epoch": 6.684118673647469,
+ "grad_norm": 0.03199538215994835,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 5.030025959014893,
+ "step": 481
+ },
+ {
+ "epoch": 6.698080279232112,
+ "grad_norm": 0.03043113648891449,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 5.033163070678711,
+ "step": 482
+ },
+ {
+ "epoch": 6.712041884816754,
+ "grad_norm": 0.031694378703832626,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 5.004739761352539,
+ "step": 483
+ },
+ {
+ "epoch": 6.726003490401396,
+ "grad_norm": 0.030243348330259323,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 5.012566089630127,
+ "step": 484
+ },
+ {
+ "epoch": 6.739965095986038,
+ "grad_norm": 0.02937416359782219,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 5.039803504943848,
+ "step": 485
+ },
+ {
+ "epoch": 6.7539267015706805,
+ "grad_norm": 0.029616443440318108,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 5.026117324829102,
+ "step": 486
+ },
+ {
+ "epoch": 6.767888307155323,
+ "grad_norm": 0.029567131772637367,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 5.006569862365723,
+ "step": 487
+ },
+ {
+ "epoch": 6.781849912739965,
+ "grad_norm": 0.029443370178341866,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.9941487312316895,
+ "step": 488
+ },
+ {
+ "epoch": 6.795811518324607,
+ "grad_norm": 0.03434352949261665,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.976431846618652,
+ "step": 489
+ },
+ {
+ "epoch": 6.80977312390925,
+ "grad_norm": 0.03845234215259552,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.993168354034424,
+ "step": 490
+ },
+ {
+ "epoch": 6.823734729493892,
+ "grad_norm": 0.05058208107948303,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.030162334442139,
+ "step": 491
+ },
+ {
+ "epoch": 6.837696335078534,
+ "grad_norm": 0.05961710214614868,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 5.007053375244141,
+ "step": 492
+ },
+ {
+ "epoch": 6.851657940663176,
+ "grad_norm": 0.056259628385305405,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 5.020356178283691,
+ "step": 493
+ },
+ {
+ "epoch": 6.865619546247818,
+ "grad_norm": 0.048628658056259155,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 5.037652969360352,
+ "step": 494
+ },
+ {
+ "epoch": 6.879581151832461,
+ "grad_norm": 0.040836453437805176,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.990390777587891,
+ "step": 495
+ },
+ {
+ "epoch": 6.893542757417103,
+ "grad_norm": 0.04112066701054573,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 5.017822742462158,
+ "step": 496
+ },
+ {
+ "epoch": 6.907504363001745,
+ "grad_norm": 0.03956829756498337,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.993971824645996,
+ "step": 497
+ },
+ {
+ "epoch": 6.9214659685863875,
+ "grad_norm": 0.03887917473912239,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.982889175415039,
+ "step": 498
+ },
+ {
+ "epoch": 6.93542757417103,
+ "grad_norm": 0.031012587249279022,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.961717128753662,
+ "step": 499
+ },
+ {
+ "epoch": 6.949389179755672,
+ "grad_norm": 0.03443489596247673,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.9904046058654785,
+ "step": 500
+ },
+ {
+ "epoch": 6.963350785340314,
+ "grad_norm": 0.035175010561943054,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.983619213104248,
+ "step": 501
+ },
+ {
+ "epoch": 6.977312390924956,
+ "grad_norm": 0.031724050641059875,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 5.000522613525391,
+ "step": 502
+ },
+ {
+ "epoch": 6.991273996509599,
+ "grad_norm": 0.032037995755672455,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.947064399719238,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.01883271522819996,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 3.112166166305542,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6296830177307129,
+ "eval_runtime": 59.5495,
+ "eval_samples_per_second": 41.008,
+ "eval_steps_per_second": 0.655,
+ "step": 504
+ },
+ {
+ "epoch": 7.013961605584642,
+ "grad_norm": 0.031091563403606415,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.919820785522461,
+ "step": 505
+ },
+ {
+ "epoch": 7.027923211169284,
+ "grad_norm": 0.04074421525001526,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.945765972137451,
+ "step": 506
+ },
+ {
+ "epoch": 7.041884816753926,
+ "grad_norm": 0.04225052148103714,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.935208797454834,
+ "step": 507
+ },
+ {
+ "epoch": 7.055846422338569,
+ "grad_norm": 0.0399051196873188,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.951314926147461,
+ "step": 508
+ },
+ {
+ "epoch": 7.069808027923211,
+ "grad_norm": 0.04785890504717827,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.92843770980835,
+ "step": 509
+ },
+ {
+ "epoch": 7.0837696335078535,
+ "grad_norm": 0.04922301694750786,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.901020050048828,
+ "step": 510
+ },
+ {
+ "epoch": 7.097731239092496,
+ "grad_norm": 0.05307295173406601,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.961151599884033,
+ "step": 511
+ },
+ {
+ "epoch": 7.111692844677138,
+ "grad_norm": 0.07003362476825714,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.9427876472473145,
+ "step": 512
+ },
+ {
+ "epoch": 7.12565445026178,
+ "grad_norm": 0.06498776376247406,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.947838306427002,
+ "step": 513
+ },
+ {
+ "epoch": 7.139616055846423,
+ "grad_norm": 0.058723628520965576,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.938859939575195,
+ "step": 514
+ },
+ {
+ "epoch": 7.153577661431065,
+ "grad_norm": 0.06590746343135834,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.962221145629883,
+ "step": 515
+ },
+ {
+ "epoch": 7.167539267015707,
+ "grad_norm": 0.052714601159095764,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.931011199951172,
+ "step": 516
+ },
+ {
+ "epoch": 7.181500872600349,
+ "grad_norm": 0.04296804964542389,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.93586540222168,
+ "step": 517
+ },
+ {
+ "epoch": 7.195462478184991,
+ "grad_norm": 0.03699173778295517,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.910763263702393,
+ "step": 518
+ },
+ {
+ "epoch": 7.209424083769633,
+ "grad_norm": 0.0317913293838501,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.9256439208984375,
+ "step": 519
+ },
+ {
+ "epoch": 7.223385689354275,
+ "grad_norm": 0.030313925817608833,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.938329696655273,
+ "step": 520
+ },
+ {
+ "epoch": 7.237347294938918,
+ "grad_norm": 0.02879229187965393,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.924966812133789,
+ "step": 521
+ },
+ {
+ "epoch": 7.2513089005235605,
+ "grad_norm": 0.02689451351761818,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.9079742431640625,
+ "step": 522
+ },
+ {
+ "epoch": 7.265270506108203,
+ "grad_norm": 0.028403930366039276,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.913238525390625,
+ "step": 523
+ },
+ {
+ "epoch": 7.279232111692845,
+ "grad_norm": 0.025912854820489883,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.896944046020508,
+ "step": 524
+ },
+ {
+ "epoch": 7.293193717277487,
+ "grad_norm": 0.024899760261178017,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.884920120239258,
+ "step": 525
+ },
+ {
+ "epoch": 7.307155322862129,
+ "grad_norm": 0.026912573724985123,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.9181060791015625,
+ "step": 526
+ },
+ {
+ "epoch": 7.321116928446771,
+ "grad_norm": 0.029539721086621284,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.914867401123047,
+ "step": 527
+ },
+ {
+ "epoch": 7.335078534031414,
+ "grad_norm": 0.028527872636914253,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.907217979431152,
+ "step": 528
+ },
+ {
+ "epoch": 7.349040139616056,
+ "grad_norm": 0.025879016146063805,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.905783176422119,
+ "step": 529
+ },
+ {
+ "epoch": 7.363001745200698,
+ "grad_norm": 0.026395199820399284,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.880505561828613,
+ "step": 530
+ },
+ {
+ "epoch": 7.37696335078534,
+ "grad_norm": 0.025072351098060608,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.870449066162109,
+ "step": 531
+ },
+ {
+ "epoch": 7.390924956369982,
+ "grad_norm": 0.02751019410789013,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.888952255249023,
+ "step": 532
+ },
+ {
+ "epoch": 7.4048865619546245,
+ "grad_norm": 0.025810658931732178,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.894552230834961,
+ "step": 533
+ },
+ {
+ "epoch": 7.418848167539267,
+ "grad_norm": 0.02399870567023754,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.8768229484558105,
+ "step": 534
+ },
+ {
+ "epoch": 7.43280977312391,
+ "grad_norm": 0.024639368057250977,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.9030914306640625,
+ "step": 535
+ },
+ {
+ "epoch": 7.446771378708552,
+ "grad_norm": 0.02543291077017784,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.86467170715332,
+ "step": 536
+ },
+ {
+ "epoch": 7.460732984293194,
+ "grad_norm": 0.02702544629573822,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.888283729553223,
+ "step": 537
+ },
+ {
+ "epoch": 7.474694589877836,
+ "grad_norm": 0.028904743492603302,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.866081237792969,
+ "step": 538
+ },
+ {
+ "epoch": 7.488656195462478,
+ "grad_norm": 0.03055848367512226,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.890739440917969,
+ "step": 539
+ },
+ {
+ "epoch": 7.50261780104712,
+ "grad_norm": 0.03410084918141365,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.862918853759766,
+ "step": 540
+ },
+ {
+ "epoch": 7.516579406631763,
+ "grad_norm": 0.03738495334982872,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.864728927612305,
+ "step": 541
+ },
+ {
+ "epoch": 7.530541012216405,
+ "grad_norm": 0.037646882236003876,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.865185737609863,
+ "step": 542
+ },
+ {
+ "epoch": 7.544502617801047,
+ "grad_norm": 0.03418618440628052,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.866114139556885,
+ "step": 543
+ },
+ {
+ "epoch": 7.558464223385689,
+ "grad_norm": 0.036924004554748535,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.883186340332031,
+ "step": 544
+ },
+ {
+ "epoch": 7.5724258289703315,
+ "grad_norm": 0.03590598329901695,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.863725662231445,
+ "step": 545
+ },
+ {
+ "epoch": 7.5863874345549736,
+ "grad_norm": 0.037363454699516296,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.841030120849609,
+ "step": 546
+ },
+ {
+ "epoch": 7.600349040139616,
+ "grad_norm": 0.046537864953279495,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.873837471008301,
+ "step": 547
+ },
+ {
+ "epoch": 7.614310645724259,
+ "grad_norm": 0.057226769626140594,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.865679740905762,
+ "step": 548
+ },
+ {
+ "epoch": 7.628272251308901,
+ "grad_norm": 0.06175961345434189,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.888645648956299,
+ "step": 549
+ },
+ {
+ "epoch": 7.642233856893543,
+ "grad_norm": 0.05063694715499878,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.877828598022461,
+ "step": 550
+ },
+ {
+ "epoch": 7.656195462478185,
+ "grad_norm": 0.03950628638267517,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.88637638092041,
+ "step": 551
+ },
+ {
+ "epoch": 7.670157068062827,
+ "grad_norm": 0.04183337092399597,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.8597917556762695,
+ "step": 552
+ },
+ {
+ "epoch": 7.684118673647469,
+ "grad_norm": 0.04642999917268753,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.880380153656006,
+ "step": 553
+ },
+ {
+ "epoch": 7.698080279232112,
+ "grad_norm": 0.05333106964826584,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.853932857513428,
+ "step": 554
+ },
+ {
+ "epoch": 7.712041884816754,
+ "grad_norm": 0.04764975979924202,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.870235443115234,
+ "step": 555
+ },
+ {
+ "epoch": 7.726003490401396,
+ "grad_norm": 0.043072231113910675,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.822697639465332,
+ "step": 556
+ },
+ {
+ "epoch": 7.739965095986038,
+ "grad_norm": 0.03913857042789459,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.849332809448242,
+ "step": 557
+ },
+ {
+ "epoch": 7.7539267015706805,
+ "grad_norm": 0.034836407750844955,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.842280387878418,
+ "step": 558
+ },
+ {
+ "epoch": 7.767888307155323,
+ "grad_norm": 0.033344294875860214,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.837811470031738,
+ "step": 559
+ },
+ {
+ "epoch": 7.781849912739965,
+ "grad_norm": 0.03342701122164726,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.876040458679199,
+ "step": 560
+ },
+ {
+ "epoch": 7.795811518324607,
+ "grad_norm": 0.034638747572898865,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.830340385437012,
+ "step": 561
+ },
+ {
+ "epoch": 7.80977312390925,
+ "grad_norm": 0.03358135372400284,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.828607559204102,
+ "step": 562
+ },
+ {
+ "epoch": 7.823734729493892,
+ "grad_norm": 0.036001041531562805,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.827905654907227,
+ "step": 563
+ },
+ {
+ "epoch": 7.837696335078534,
+ "grad_norm": 0.038079630583524704,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.866955280303955,
+ "step": 564
+ },
+ {
+ "epoch": 7.851657940663176,
+ "grad_norm": 0.03728470206260681,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.828039169311523,
+ "step": 565
+ },
+ {
+ "epoch": 7.865619546247818,
+ "grad_norm": 0.03920375928282738,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.841631889343262,
+ "step": 566
+ },
+ {
+ "epoch": 7.879581151832461,
+ "grad_norm": 0.032353926450014114,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.861110210418701,
+ "step": 567
+ },
+ {
+ "epoch": 7.893542757417103,
+ "grad_norm": 0.03241422027349472,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.818441390991211,
+ "step": 568
+ },
+ {
+ "epoch": 7.907504363001745,
+ "grad_norm": 0.03879080340266228,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.823749542236328,
+ "step": 569
+ },
+ {
+ "epoch": 7.9214659685863875,
+ "grad_norm": 0.03486498072743416,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.830697059631348,
+ "step": 570
+ },
+ {
+ "epoch": 7.93542757417103,
+ "grad_norm": 0.03279713913798332,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.815736770629883,
+ "step": 571
+ },
+ {
+ "epoch": 7.949389179755672,
+ "grad_norm": 0.03313925117254257,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.8364033699035645,
+ "step": 572
+ },
+ {
+ "epoch": 7.963350785340314,
+ "grad_norm": 0.03405062481760979,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.816010475158691,
+ "step": 573
+ },
+ {
+ "epoch": 7.977312390924956,
+ "grad_norm": 0.03259236365556717,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.813413619995117,
+ "step": 574
+ },
+ {
+ "epoch": 7.991273996509599,
+ "grad_norm": 0.030408985912799835,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.833840370178223,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.022680889815092087,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 2.99565052986145,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6126086115837097,
+ "eval_runtime": 60.2662,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 576
+ },
+ {
+ "epoch": 8.013961605584642,
+ "grad_norm": 0.029732126742601395,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.7816314697265625,
+ "step": 577
+ },
+ {
+ "epoch": 8.027923211169284,
+ "grad_norm": 0.04357453063130379,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.7575225830078125,
+ "step": 578
+ },
+ {
+ "epoch": 8.041884816753926,
+ "grad_norm": 0.051903221756219864,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.802519798278809,
+ "step": 579
+ },
+ {
+ "epoch": 8.055846422338568,
+ "grad_norm": 0.0539076030254364,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.799004554748535,
+ "step": 580
+ },
+ {
+ "epoch": 8.06980802792321,
+ "grad_norm": 0.043642807751894,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.7896528244018555,
+ "step": 581
+ },
+ {
+ "epoch": 8.083769633507853,
+ "grad_norm": 0.0469554178416729,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.765430927276611,
+ "step": 582
+ },
+ {
+ "epoch": 8.097731239092496,
+ "grad_norm": 0.042060062289237976,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.784132957458496,
+ "step": 583
+ },
+ {
+ "epoch": 8.111692844677139,
+ "grad_norm": 0.035101912915706635,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.767234802246094,
+ "step": 584
+ },
+ {
+ "epoch": 8.12565445026178,
+ "grad_norm": 0.042576879262924194,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.762052536010742,
+ "step": 585
+ },
+ {
+ "epoch": 8.139616055846423,
+ "grad_norm": 0.04231419041752815,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.752150535583496,
+ "step": 586
+ },
+ {
+ "epoch": 8.153577661431065,
+ "grad_norm": 0.046912599354982376,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.771881103515625,
+ "step": 587
+ },
+ {
+ "epoch": 8.167539267015707,
+ "grad_norm": 0.051271144300699234,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.773284912109375,
+ "step": 588
+ },
+ {
+ "epoch": 8.181500872600349,
+ "grad_norm": 0.04655958339571953,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.769917964935303,
+ "step": 589
+ },
+ {
+ "epoch": 8.195462478184991,
+ "grad_norm": 0.04463687539100647,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.77938175201416,
+ "step": 590
+ },
+ {
+ "epoch": 8.209424083769633,
+ "grad_norm": 0.047959886491298676,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.785962104797363,
+ "step": 591
+ },
+ {
+ "epoch": 8.223385689354275,
+ "grad_norm": 0.04414986073970795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.786281585693359,
+ "step": 592
+ },
+ {
+ "epoch": 8.237347294938917,
+ "grad_norm": 0.044150348752737045,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.777070045471191,
+ "step": 593
+ },
+ {
+ "epoch": 8.25130890052356,
+ "grad_norm": 0.037002332508563995,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.77547550201416,
+ "step": 594
+ },
+ {
+ "epoch": 8.265270506108202,
+ "grad_norm": 0.03417367860674858,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.782101154327393,
+ "step": 595
+ },
+ {
+ "epoch": 8.279232111692846,
+ "grad_norm": 0.03069210797548294,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.745595455169678,
+ "step": 596
+ },
+ {
+ "epoch": 8.293193717277488,
+ "grad_norm": 0.03228718787431717,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.7663655281066895,
+ "step": 597
+ },
+ {
+ "epoch": 8.30715532286213,
+ "grad_norm": 0.028603695333003998,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.767274856567383,
+ "step": 598
+ },
+ {
+ "epoch": 8.321116928446772,
+ "grad_norm": 0.03098377026617527,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.7853684425354,
+ "step": 599
+ },
+ {
+ "epoch": 8.335078534031414,
+ "grad_norm": 0.029242439195513725,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.738029956817627,
+ "step": 600
+ },
+ {
+ "epoch": 8.349040139616056,
+ "grad_norm": 0.02751081995666027,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.76801872253418,
+ "step": 601
+ },
+ {
+ "epoch": 8.363001745200698,
+ "grad_norm": 0.02838125079870224,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.743884086608887,
+ "step": 602
+ },
+ {
+ "epoch": 8.37696335078534,
+ "grad_norm": 0.029203062877058983,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.746668815612793,
+ "step": 603
+ },
+ {
+ "epoch": 8.390924956369982,
+ "grad_norm": 0.02679944597184658,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.760764122009277,
+ "step": 604
+ },
+ {
+ "epoch": 8.404886561954624,
+ "grad_norm": 0.028955848887562752,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.7353668212890625,
+ "step": 605
+ },
+ {
+ "epoch": 8.418848167539267,
+ "grad_norm": 0.02880851738154888,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.766837120056152,
+ "step": 606
+ },
+ {
+ "epoch": 8.432809773123909,
+ "grad_norm": 0.02495497092604637,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.78959321975708,
+ "step": 607
+ },
+ {
+ "epoch": 8.44677137870855,
+ "grad_norm": 0.026064753532409668,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.736433029174805,
+ "step": 608
+ },
+ {
+ "epoch": 8.460732984293193,
+ "grad_norm": 0.027566388249397278,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.782382965087891,
+ "step": 609
+ },
+ {
+ "epoch": 8.474694589877837,
+ "grad_norm": 0.02967359498143196,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.735286712646484,
+ "step": 610
+ },
+ {
+ "epoch": 8.488656195462479,
+ "grad_norm": 0.03079676441848278,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.725196838378906,
+ "step": 611
+ },
+ {
+ "epoch": 8.502617801047121,
+ "grad_norm": 0.032531846314668655,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.719333648681641,
+ "step": 612
+ },
+ {
+ "epoch": 8.516579406631763,
+ "grad_norm": 0.03120352514088154,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.723845958709717,
+ "step": 613
+ },
+ {
+ "epoch": 8.530541012216405,
+ "grad_norm": 0.03328424692153931,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.7669782638549805,
+ "step": 614
+ },
+ {
+ "epoch": 8.544502617801047,
+ "grad_norm": 0.035601649433374405,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.714925765991211,
+ "step": 615
+ },
+ {
+ "epoch": 8.55846422338569,
+ "grad_norm": 0.03675796464085579,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.715317726135254,
+ "step": 616
+ },
+ {
+ "epoch": 8.572425828970331,
+ "grad_norm": 0.03701988235116005,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.724606990814209,
+ "step": 617
+ },
+ {
+ "epoch": 8.586387434554974,
+ "grad_norm": 0.036719705909490585,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.746044635772705,
+ "step": 618
+ },
+ {
+ "epoch": 8.600349040139616,
+ "grad_norm": 0.034756604582071304,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.74885368347168,
+ "step": 619
+ },
+ {
+ "epoch": 8.614310645724258,
+ "grad_norm": 0.03237922117114067,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.721858501434326,
+ "step": 620
+ },
+ {
+ "epoch": 8.6282722513089,
+ "grad_norm": 0.028420142829418182,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.733814239501953,
+ "step": 621
+ },
+ {
+ "epoch": 8.642233856893542,
+ "grad_norm": 0.026203038170933723,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.726088523864746,
+ "step": 622
+ },
+ {
+ "epoch": 8.656195462478184,
+ "grad_norm": 0.02627231739461422,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.727700233459473,
+ "step": 623
+ },
+ {
+ "epoch": 8.670157068062828,
+ "grad_norm": 0.02920261025428772,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.731454849243164,
+ "step": 624
+ },
+ {
+ "epoch": 8.68411867364747,
+ "grad_norm": 0.02792907878756523,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.734154224395752,
+ "step": 625
+ },
+ {
+ "epoch": 8.698080279232112,
+ "grad_norm": 0.026770325377583504,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.715337753295898,
+ "step": 626
+ },
+ {
+ "epoch": 8.712041884816754,
+ "grad_norm": 0.027346540242433548,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.716512680053711,
+ "step": 627
+ },
+ {
+ "epoch": 8.726003490401396,
+ "grad_norm": 0.031965773552656174,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.7072038650512695,
+ "step": 628
+ },
+ {
+ "epoch": 8.739965095986038,
+ "grad_norm": 0.03708025813102722,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.749785900115967,
+ "step": 629
+ },
+ {
+ "epoch": 8.75392670157068,
+ "grad_norm": 0.04846490919589996,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.735586166381836,
+ "step": 630
+ },
+ {
+ "epoch": 8.767888307155323,
+ "grad_norm": 0.053352758288383484,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.720524787902832,
+ "step": 631
+ },
+ {
+ "epoch": 8.781849912739965,
+ "grad_norm": 0.040559202432632446,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.719901084899902,
+ "step": 632
+ },
+ {
+ "epoch": 8.795811518324607,
+ "grad_norm": 0.03333032503724098,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.708139419555664,
+ "step": 633
+ },
+ {
+ "epoch": 8.809773123909249,
+ "grad_norm": 0.030532941222190857,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.728184223175049,
+ "step": 634
+ },
+ {
+ "epoch": 8.823734729493891,
+ "grad_norm": 0.03337649628520012,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.726020336151123,
+ "step": 635
+ },
+ {
+ "epoch": 8.837696335078533,
+ "grad_norm": 0.03352045267820358,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.7196855545043945,
+ "step": 636
+ },
+ {
+ "epoch": 8.851657940663177,
+ "grad_norm": 0.03662872314453125,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.732823371887207,
+ "step": 637
+ },
+ {
+ "epoch": 8.86561954624782,
+ "grad_norm": 0.037085745483636856,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.728443145751953,
+ "step": 638
+ },
+ {
+ "epoch": 8.879581151832461,
+ "grad_norm": 0.036635689437389374,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.713988780975342,
+ "step": 639
+ },
+ {
+ "epoch": 8.893542757417103,
+ "grad_norm": 0.03719659522175789,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.715950965881348,
+ "step": 640
+ },
+ {
+ "epoch": 8.907504363001745,
+ "grad_norm": 0.03719159588217735,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.7124924659729,
+ "step": 641
+ },
+ {
+ "epoch": 8.921465968586388,
+ "grad_norm": 0.04275721684098244,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.708404064178467,
+ "step": 642
+ },
+ {
+ "epoch": 8.93542757417103,
+ "grad_norm": 0.04204598814249039,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.715372085571289,
+ "step": 643
+ },
+ {
+ "epoch": 8.949389179755672,
+ "grad_norm": 0.03938259929418564,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.729745864868164,
+ "step": 644
+ },
+ {
+ "epoch": 8.963350785340314,
+ "grad_norm": 0.036226850003004074,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.733749866485596,
+ "step": 645
+ },
+ {
+ "epoch": 8.977312390924956,
+ "grad_norm": 0.03542874753475189,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.706825256347656,
+ "step": 646
+ },
+ {
+ "epoch": 8.991273996509598,
+ "grad_norm": 0.029055381193757057,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.718567848205566,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.01990380510687828,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 2.943183183670044,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.601699709892273,
+ "eval_runtime": 60.0221,
+ "eval_samples_per_second": 40.685,
+ "eval_steps_per_second": 0.65,
+ "step": 648
+ },
+ {
+ "epoch": 9.013961605584642,
+ "grad_norm": 0.030904127284884453,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.660327911376953,
+ "step": 649
+ },
+ {
+ "epoch": 9.027923211169284,
+ "grad_norm": 0.0441313236951828,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.701565742492676,
+ "step": 650
+ },
+ {
+ "epoch": 9.041884816753926,
+ "grad_norm": 0.047547727823257446,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.666440486907959,
+ "step": 651
+ },
+ {
+ "epoch": 9.055846422338568,
+ "grad_norm": 0.04549463838338852,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.660548210144043,
+ "step": 652
+ },
+ {
+ "epoch": 9.06980802792321,
+ "grad_norm": 0.050346389412879944,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.640229225158691,
+ "step": 653
+ },
+ {
+ "epoch": 9.083769633507853,
+ "grad_norm": 0.058505453169345856,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.673881530761719,
+ "step": 654
+ },
+ {
+ "epoch": 9.097731239092496,
+ "grad_norm": 0.051989227533340454,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.693943977355957,
+ "step": 655
+ },
+ {
+ "epoch": 9.111692844677139,
+ "grad_norm": 0.052564945071935654,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.644315242767334,
+ "step": 656
+ },
+ {
+ "epoch": 9.12565445026178,
+ "grad_norm": 0.05647841840982437,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.681768417358398,
+ "step": 657
+ },
+ {
+ "epoch": 9.139616055846423,
+ "grad_norm": 0.058640867471694946,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.676457405090332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153577661431065,
+ "grad_norm": 0.05386025458574295,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.659588813781738,
+ "step": 659
+ },
+ {
+ "epoch": 9.167539267015707,
+ "grad_norm": 0.048699021339416504,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.680031776428223,
+ "step": 660
+ },
+ {
+ "epoch": 9.181500872600349,
+ "grad_norm": 0.04374176263809204,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.672671318054199,
+ "step": 661
+ },
+ {
+ "epoch": 9.195462478184991,
+ "grad_norm": 0.042337436228990555,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.673431873321533,
+ "step": 662
+ },
+ {
+ "epoch": 9.209424083769633,
+ "grad_norm": 0.04564661160111427,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.632595062255859,
+ "step": 663
+ },
+ {
+ "epoch": 9.223385689354275,
+ "grad_norm": 0.04404835030436516,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.678891181945801,
+ "step": 664
+ },
+ {
+ "epoch": 9.237347294938917,
+ "grad_norm": 0.040642790496349335,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.633418560028076,
+ "step": 665
+ },
+ {
+ "epoch": 9.25130890052356,
+ "grad_norm": 0.036492060869932175,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.670109748840332,
+ "step": 666
+ },
+ {
+ "epoch": 9.265270506108202,
+ "grad_norm": 0.03735879808664322,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.608570098876953,
+ "step": 667
+ },
+ {
+ "epoch": 9.279232111692846,
+ "grad_norm": 0.034563641995191574,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.632505416870117,
+ "step": 668
+ },
+ {
+ "epoch": 9.293193717277488,
+ "grad_norm": 0.03522484749555588,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.6475629806518555,
+ "step": 669
+ },
+ {
+ "epoch": 9.30715532286213,
+ "grad_norm": 0.03216472268104553,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.627873420715332,
+ "step": 670
+ },
+ {
+ "epoch": 9.321116928446772,
+ "grad_norm": 0.029155759140849113,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.652706146240234,
+ "step": 671
+ },
+ {
+ "epoch": 9.335078534031414,
+ "grad_norm": 0.031060541048645973,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.644545078277588,
+ "step": 672
+ },
+ {
+ "epoch": 9.349040139616056,
+ "grad_norm": 0.034979186952114105,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.642645359039307,
+ "step": 673
+ },
+ {
+ "epoch": 9.363001745200698,
+ "grad_norm": 0.035258810967206955,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.647383689880371,
+ "step": 674
+ },
+ {
+ "epoch": 9.37696335078534,
+ "grad_norm": 0.031552575528621674,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.667779445648193,
+ "step": 675
+ },
+ {
+ "epoch": 9.390924956369982,
+ "grad_norm": 0.03193951025605202,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.63625431060791,
+ "step": 676
+ },
+ {
+ "epoch": 9.404886561954624,
+ "grad_norm": 0.029811711981892586,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.658000946044922,
+ "step": 677
+ },
+ {
+ "epoch": 9.418848167539267,
+ "grad_norm": 0.028464138507843018,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.6545515060424805,
+ "step": 678
+ },
+ {
+ "epoch": 9.432809773123909,
+ "grad_norm": 0.0315735749900341,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.641153335571289,
+ "step": 679
+ },
+ {
+ "epoch": 9.44677137870855,
+ "grad_norm": 0.033412106335163116,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.626931667327881,
+ "step": 680
+ },
+ {
+ "epoch": 9.460732984293193,
+ "grad_norm": 0.03757626190781593,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.671254634857178,
+ "step": 681
+ },
+ {
+ "epoch": 9.474694589877837,
+ "grad_norm": 0.03910793736577034,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.651954174041748,
+ "step": 682
+ },
+ {
+ "epoch": 9.488656195462479,
+ "grad_norm": 0.04015549272298813,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.613324165344238,
+ "step": 683
+ },
+ {
+ "epoch": 9.502617801047121,
+ "grad_norm": 0.040689144283533096,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.639243125915527,
+ "step": 684
+ },
+ {
+ "epoch": 9.516579406631763,
+ "grad_norm": 0.04075361415743828,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.649250030517578,
+ "step": 685
+ },
+ {
+ "epoch": 9.530541012216405,
+ "grad_norm": 0.03914607688784599,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.66444730758667,
+ "step": 686
+ },
+ {
+ "epoch": 9.544502617801047,
+ "grad_norm": 0.033253733068704605,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.647335052490234,
+ "step": 687
+ },
+ {
+ "epoch": 9.55846422338569,
+ "grad_norm": 0.03179008886218071,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.6727800369262695,
+ "step": 688
+ },
+ {
+ "epoch": 9.572425828970331,
+ "grad_norm": 0.03477025404572487,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.655778884887695,
+ "step": 689
+ },
+ {
+ "epoch": 9.586387434554974,
+ "grad_norm": 0.030094800516963005,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.680851936340332,
+ "step": 690
+ },
+ {
+ "epoch": 9.600349040139616,
+ "grad_norm": 0.030153103172779083,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.654638290405273,
+ "step": 691
+ },
+ {
+ "epoch": 9.614310645724258,
+ "grad_norm": 0.029009278863668442,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.625916481018066,
+ "step": 692
+ },
+ {
+ "epoch": 9.6282722513089,
+ "grad_norm": 0.027248440310359,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.656370162963867,
+ "step": 693
+ },
+ {
+ "epoch": 9.642233856893542,
+ "grad_norm": 0.027727462351322174,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.631473541259766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656195462478184,
+ "grad_norm": 0.026710307225584984,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.636524200439453,
+ "step": 695
+ },
+ {
+ "epoch": 9.670157068062828,
+ "grad_norm": 0.02726181596517563,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.629701614379883,
+ "step": 696
+ },
+ {
+ "epoch": 9.68411867364747,
+ "grad_norm": 0.027146652340888977,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.640240669250488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698080279232112,
+ "grad_norm": 0.027165887877345085,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.64462947845459,
+ "step": 698
+ },
+ {
+ "epoch": 9.712041884816754,
+ "grad_norm": 0.03025982715189457,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.6271820068359375,
+ "step": 699
+ },
+ {
+ "epoch": 9.726003490401396,
+ "grad_norm": 0.02837025746703148,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.629697799682617,
+ "step": 700
+ },
+ {
+ "epoch": 9.739965095986038,
+ "grad_norm": 0.026777995750308037,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.646146774291992,
+ "step": 701
+ },
+ {
+ "epoch": 9.75392670157068,
+ "grad_norm": 0.024626802653074265,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.641671657562256,
+ "step": 702
+ },
+ {
+ "epoch": 9.767888307155323,
+ "grad_norm": 0.025329895317554474,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.626345157623291,
+ "step": 703
+ },
+ {
+ "epoch": 9.781849912739965,
+ "grad_norm": 0.027804844081401825,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.621457099914551,
+ "step": 704
+ },
+ {
+ "epoch": 9.795811518324607,
+ "grad_norm": 0.028763437643647194,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.6124114990234375,
+ "step": 705
+ },
+ {
+ "epoch": 9.809773123909249,
+ "grad_norm": 0.030769916251301765,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.604399681091309,
+ "step": 706
+ },
+ {
+ "epoch": 9.823734729493891,
+ "grad_norm": 0.03060949593782425,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.640804290771484,
+ "step": 707
+ },
+ {
+ "epoch": 9.837696335078533,
+ "grad_norm": 0.029832778498530388,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.593376159667969,
+ "step": 708
+ },
+ {
+ "epoch": 9.851657940663177,
+ "grad_norm": 0.02952776849269867,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.619533538818359,
+ "step": 709
+ },
+ {
+ "epoch": 9.86561954624782,
+ "grad_norm": 0.02894212119281292,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.640516757965088,
+ "step": 710
+ },
+ {
+ "epoch": 9.879581151832461,
+ "grad_norm": 0.026435624808073044,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.6342997550964355,
+ "step": 711
+ },
+ {
+ "epoch": 9.893542757417103,
+ "grad_norm": 0.02591666206717491,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.630884170532227,
+ "step": 712
+ },
+ {
+ "epoch": 9.907504363001745,
+ "grad_norm": 0.02702309936285019,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.632085800170898,
+ "step": 713
+ },
+ {
+ "epoch": 9.921465968586388,
+ "grad_norm": 0.026766657829284668,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.6231231689453125,
+ "step": 714
+ },
+ {
+ "epoch": 9.93542757417103,
+ "grad_norm": 0.026666559278964996,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.601619243621826,
+ "step": 715
+ },
+ {
+ "epoch": 9.949389179755672,
+ "grad_norm": 0.026801064610481262,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.641724109649658,
+ "step": 716
+ },
+ {
+ "epoch": 9.963350785340314,
+ "grad_norm": 0.03152798116207123,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.642511367797852,
+ "step": 717
+ },
+ {
+ "epoch": 9.977312390924956,
+ "grad_norm": 0.035406943410634995,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.635779857635498,
+ "step": 718
+ },
+ {
+ "epoch": 9.991273996509598,
+ "grad_norm": 0.037504684180021286,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.629463195800781,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02263174019753933,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 2.876767635345459,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.5942904949188232,
+ "eval_runtime": 60.2086,
+ "eval_samples_per_second": 40.559,
+ "eval_steps_per_second": 0.648,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03707852587103844,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.5765533447265625,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.05454680323600769,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.576363563537598,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.04832114651799202,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.586915969848633,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.050469402223825455,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.5265092849731445,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0622565895318985,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.592833518981934,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0640949010848999,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.5855560302734375,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.06502196192741394,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.591236114501953,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.06167149916291237,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.56747579574585,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.06747374683618546,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.60452938079834,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.07210271060466766,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.561588764190674,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.0733448714017868,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.610766410827637,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.0731528028845787,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.589570045471191,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.0822535902261734,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.631077289581299,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.06921929866075516,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.628293991088867,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.07443586736917496,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.627321243286133,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.07153813540935516,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.641068935394287,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.069196417927742,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.615875720977783,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06483697891235352,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.6165313720703125,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.05482150986790657,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.599890232086182,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.05364019051194191,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.61989164352417,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.053216829895973206,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.564305305480957,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05740506947040558,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.582683563232422,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.05587657541036606,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.61264705657959,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.050252825021743774,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.648050308227539,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.050955042243003845,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.615367889404297,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04903746768832207,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.617378234863281,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04252410680055618,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.609130859375,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.03801053389906883,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.625149250030518,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.03422372043132782,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.606106281280518,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.030507879331707954,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.6125993728637695,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.0340375155210495,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.5690202713012695,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.030507374554872513,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.6042985916137695,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027383938431739807,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.585570812225342,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.028792303055524826,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.566786766052246,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02866457775235176,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.566376686096191,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.028657952323555946,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.581750869750977,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.028227943927049637,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.576788902282715,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.024757765233516693,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.59373664855957,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.027442635968327522,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.570978164672852,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.02590034157037735,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.571286201477051,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.025696570053696632,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.542209625244141,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.025478003546595573,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.5900421142578125,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.02563696913421154,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.544609546661377,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.025326019152998924,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.5808515548706055,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.024502627551555634,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.560579776763916,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.02687518671154976,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.567154884338379,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.025280235335230827,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.560011863708496,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.026266315951943398,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.5758442878723145,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.026058891788125038,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.583309173583984,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.02620471641421318,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.547737121582031,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02626362442970276,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.556955337524414,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.024503158405423164,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.580735206604004,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02580532431602478,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.564000129699707,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.024437161162495613,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.545594215393066,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.026664750650525093,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.56415319442749,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.02499818615615368,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.550689697265625,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.027514580637216568,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.5548248291015625,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.026799403131008148,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.548295021057129,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.02685476280748844,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.5660600662231445,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.026104001328349113,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.556082248687744,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.027479661628603935,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.5580244064331055,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.02638247236609459,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.550073623657227,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.025853842496871948,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.578088760375977,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.02636682614684105,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.58807373046875,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.02684398740530014,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.561017036437988,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.028538964688777924,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.548762321472168,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.028327809646725655,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.548969745635986,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.025131452828645706,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.564047813415527,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.026498543098568916,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.553997039794922,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.030081963166594505,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.5787787437438965,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.03201213851571083,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.566424369812012,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.020518779754638672,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8421616554260254,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.5899484157562256,
+ "eval_runtime": 60.2671,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03692612797021866,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.492715835571289,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.050258126109838486,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.480323791503906,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.04179152473807335,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.484371185302734,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.04160204157233238,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.4905524253845215,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.03974860534071922,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.503803253173828,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.037031810730695724,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.494629859924316,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.038762833923101425,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.495111465454102,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.0412563718855381,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.490113735198975,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.041371624916791916,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.490643501281738,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04151557758450508,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.49881649017334,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04037577286362648,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.492178916931152,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.03676290065050125,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.477460861206055,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.038055796176195145,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.509150981903076,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.041152238845825195,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.47133731842041,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.03728092834353447,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.508072376251221,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.035338856279850006,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.505619049072266,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.037813492119312286,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.509489059448242,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.03789111226797104,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.508646488189697,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.03880828619003296,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.52567720413208,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.03611127287149429,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.498889446258545,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.036903638392686844,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.498601913452148,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03620576485991478,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.490840911865234,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03555903211236,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.506166458129883,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.03441924974322319,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.498001575469971,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.0334339402616024,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.491082191467285,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.031418606638908386,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.527408599853516,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03073921985924244,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.514107704162598,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.029589885845780373,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.48323917388916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.02974501997232437,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.51338529586792,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.029247552156448364,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.504205226898193,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.03132032975554466,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.505674362182617,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03088037669658661,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.477198600769043,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.031054604798555374,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.498721122741699,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03293661028146744,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.527826309204102,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.03181406110525131,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.495999336242676,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.029380083084106445,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.530351638793945,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.029296347871422768,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.492367744445801,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.02958036959171295,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.493037700653076,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03078395500779152,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.47504186630249,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03280661255121231,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.518409729003906,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.03323954716324806,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.460753917694092,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.032254140824079514,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.474285125732422,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.03018898330628872,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.5162353515625,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.02802245132625103,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.538684844970703,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.02821732498705387,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.4850287437438965,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.029189350083470345,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.50770378112793,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.02851196378469467,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.499988555908203,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.02741030789911747,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.499852180480957,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.028708625584840775,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.465211868286133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.028519680723547935,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.502464294433594,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.027846841141581535,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.503248691558838,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030456233769655228,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.490391731262207,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029952052980661392,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.50056266784668,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.02925778552889824,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.49215030670166,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.029679695144295692,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.523841857910156,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.031178541481494904,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.499871730804443,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.031264498829841614,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.50546932220459,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03482109308242798,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.494480133056641,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.03341105580329895,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.532899856567383,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028623664751648903,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.5016279220581055,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028600146993994713,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.511589050292969,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.029127901419997215,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.502058982849121,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.027898237109184265,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.493182182312012,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.028683282434940338,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.519793510437012,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.02887023612856865,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.457289695739746,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.02811296097934246,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.49398946762085,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.028485111892223358,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.478344440460205,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02914245054125786,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.487666130065918,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.028030497953295708,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.518488883972168,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.027973609045147896,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.515271186828613,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.027458522468805313,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.490856647491455,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.01909293793141842,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.822120189666748,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5866912603378296,
+ "eval_runtime": 61.3185,
+ "eval_samples_per_second": 39.825,
+ "eval_steps_per_second": 0.636,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03683992475271225,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.438363075256348,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.05042923241853714,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.443560600280762,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.055914171040058136,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.422282695770264,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.057007886469364166,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.442644119262695,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.05999710038304329,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.44780969619751,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06472347676753998,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.485074996948242,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.06923902779817581,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.457515716552734,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.0710320770740509,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.465476989746094,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.06942574679851532,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.426226615905762,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.06867868453264236,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.438092231750488,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.0688796266913414,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.446664810180664,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.0690719485282898,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.456482887268066,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.06068912148475647,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.4730730056762695,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.05411486327648163,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.4596452713012695,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.056623928248882294,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.418197154998779,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.05634097009897232,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.482892036437988,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.05056057870388031,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.457589149475098,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.04460986703634262,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.451439380645752,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.045022401958703995,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.452855110168457,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.04054102674126625,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.417323112487793,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.03682882338762283,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.465713977813721,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.03528819605708122,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.4502339363098145,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.03568802773952484,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.422399520874023,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.0363088883459568,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.447678089141846,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.03460463136434555,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.437664985656738,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.034751344472169876,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.489151954650879,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.033658601343631744,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.4471635818481445,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.03433239087462425,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.455332279205322,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.032074104994535446,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.435157775878906,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.032168325036764145,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.435118675231934,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.03007678873836994,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.42781925201416,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03094375506043434,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.433105945587158,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.032186444848775864,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.383000373840332,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.030610304325819016,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.473257064819336,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.030257435515522957,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.453041076660156,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.0301323514431715,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.396578788757324,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.030934512615203857,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.445223808288574,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.029907869175076485,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.442287445068359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03163841366767883,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.434650421142578,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.029627414420247078,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.454374313354492,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.031026875600218773,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.413784027099609,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03099491260945797,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.458438873291016,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.02981731668114662,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.438560962677002,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.02927570603787899,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.44493293762207,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.029779717326164246,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.445121765136719,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.02909068390727043,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.4736127853393555,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.029217200353741646,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.434604644775391,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.029872911050915718,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.454591751098633,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.028742222115397453,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.468222141265869,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.030313383787870407,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.45759391784668,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.03014662116765976,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.446331977844238,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029481127858161926,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.478209495544434,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.028485266491770744,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.444445610046387,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.02857700176537037,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.430604934692383,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.029038602486252785,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.43759298324585,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.029576119035482407,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.474579811096191,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.02897879108786583,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.440483093261719,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.029799485579133034,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.4839277267456055,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.030710609629750252,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.466696739196777,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.030270954594016075,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.455197334289551,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.03170182928442955,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.454664707183838,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030684730038046837,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.437619209289551,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.030972573906183243,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.431796073913574,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028914963826537132,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.460618019104004,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.029810264706611633,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.407312393188477,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.029806939885020256,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.466520309448242,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.03020886518061161,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.427349090576172,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030598826706409454,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.427891731262207,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03192012757062912,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.448113918304443,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.030492989346385002,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.450060844421387,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03188254311680794,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.43007230758667,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02081947773694992,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.782536506652832,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5860040187835693,
+ "eval_runtime": 60.6725,
+ "eval_samples_per_second": 40.249,
+ "eval_steps_per_second": 0.643,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.036721356213092804,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.358441352844238,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.056633755564689636,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.361527919769287,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.06175022944808006,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.3827924728393555,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05159056559205055,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.354856491088867,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.04885871335864067,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.359966278076172,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.04407693073153496,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.364612579345703,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.04891050234436989,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.344021320343018,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.04643657058477402,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.366066932678223,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.05241481214761734,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.3390607833862305,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.05837438255548477,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.390261650085449,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.058482322841882706,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.3856201171875,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.06507553160190582,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.414223670959473,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.0575365275144577,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.402576446533203,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05055592209100723,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.389563083648682,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.047167956829071045,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.402750015258789,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.04473405331373215,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.358315467834473,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.042440690100193024,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.392551898956299,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04465719312429428,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.373558044433594,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.04337528720498085,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.37944221496582,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.04322915896773338,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.38493013381958,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.039873406291007996,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.389793395996094,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.0413178876042366,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.427964210510254,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.04567280039191246,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.404172897338867,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04154926910996437,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.391359329223633,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.03774876892566681,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.381457328796387,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03588850423693657,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.392526626586914,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.03757261857390404,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.39273738861084,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03849692642688751,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.409083366394043,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03787294402718544,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.3776044845581055,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.034496866166591644,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.413462162017822,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.03973205015063286,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.359318733215332,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.036192040890455246,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.392844200134277,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.03436632081866264,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.421384811401367,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.035658642649650574,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.386757850646973,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.03434896469116211,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.364978313446045,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03642290085554123,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.385326385498047,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.03427736461162567,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.406352996826172,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.034845270216464996,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.370953559875488,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03390975296497345,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.3799943923950195,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.03134961053729057,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.419156551361084,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.03512260690331459,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.404660701751709,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03319067507982254,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.388535022735596,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03113507106900215,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.403522968292236,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03559132665395737,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.382323265075684,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.03579750657081604,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.37620210647583,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.033675212413072586,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.411443710327148,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.03220067173242569,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.443470001220703,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031610481441020966,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.406038284301758,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.02958764135837555,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.41604471206665,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.0318237729370594,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4253387451171875,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03260074555873871,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.370792388916016,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.03359333053231239,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.397639274597168,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032516930252313614,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.4140777587890625,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.03203396871685982,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.408872604370117,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.03201935440301895,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.402975082397461,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.03330211713910103,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.394018173217773,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.0348181314766407,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.380587100982666,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.03174913674592972,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.372378349304199,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.029864462092518806,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.42018985748291,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.033312153071165085,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.385973930358887,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03203144669532776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.395167827606201,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.031578030437231064,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.3982062339782715,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.030885562300682068,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.400731086730957,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03235667943954468,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.389830589294434,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.032664328813552856,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.401277542114258,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.030354196205735207,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.422520637512207,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.031162403523921967,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.393032073974609,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.032187070697546005,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.464164733886719,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.03445067256689072,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.368991851806641,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03398612141609192,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.418752670288086,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.029730670154094696,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.41429328918457,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022440090775489807,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7315595149993896,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5853753089904785,
+ "eval_runtime": 60.7174,
+ "eval_samples_per_second": 40.219,
+ "eval_steps_per_second": 0.642,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.03754177689552307,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.304028511047363,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.051616180688142776,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.34659481048584,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055258627980947495,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.310445785522461,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.049489911645650864,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3416619300842285,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05377533286809921,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.3143310546875,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.05443088337779045,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.320600986480713,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.050222691148519516,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.334789752960205,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.051904670894145966,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.316756248474121,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.05376426503062248,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.274116516113281,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.053313180804252625,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.321177005767822,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.058234598487615585,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.3169264793396,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05730101093649864,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.307013511657715,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.058140840381383896,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.332961082458496,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05654815956950188,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.315462112426758,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.05129384249448776,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.338476181030273,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.047652702778577805,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.321232795715332,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.050912003964185715,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.318971633911133,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.0514022521674633,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.3334150314331055,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.05457981303334236,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.325802803039551,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04678885266184807,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.307281494140625,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.05048161372542381,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.299831390380859,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.05334261804819107,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.313231468200684,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.05055771768093109,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.36859130859375,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04955561086535454,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.301530361175537,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.04548008739948273,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.339517593383789,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.046585116535425186,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.333036422729492,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.0503716766834259,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.364710807800293,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.040569406002759933,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.369089126586914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.03754054754972458,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.3349432945251465,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03775954991579056,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.333134174346924,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.04066266492009163,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.354358673095703,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.040490977466106415,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.360981464385986,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.0426657609641552,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.357726097106934,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.0415605828166008,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.363457679748535,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.04020509123802185,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.372447967529297,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.04101516678929329,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.292452812194824,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.03546427935361862,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.33892297744751,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.03661105036735535,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.329778671264648,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03565476834774017,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.351012229919434,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.03591311350464821,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.376980781555176,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.03665630891919136,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.354767799377441,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.0371236577630043,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.374916076660156,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.03464849665760994,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.3713059425354,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03345152735710144,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.350670337677002,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.03357568010687828,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.3573713302612305,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03503486514091492,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.362038612365723,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.03455005958676338,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.3263773918151855,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.03173859417438507,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.351345062255859,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03301836550235748,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.398364067077637,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.033733855932950974,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.335268020629883,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034080155193805695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.371492385864258,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.034396376460790634,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.382076740264893,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.03303401917219162,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.36623477935791,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03297369182109833,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3148393630981445,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.034819211810827255,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.358110427856445,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.033389780670404434,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.367427825927734,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.032363444566726685,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.3813676834106445,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03509577363729477,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.386520862579346,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.034685730934143066,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.33738374710083,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.034755002707242966,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.360980987548828,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03461688384413719,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.396650314331055,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03392560034990311,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.356130123138428,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.03342147544026375,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.359145164489746,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.032852694392204285,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.3796162605285645,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03472859039902687,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.361110687255859,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03277098760008812,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.361811637878418,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.03341882303357124,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.369524955749512,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.032347049564123154,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.369096755981445,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03161193057894707,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.339844703674316,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.03322301432490349,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.376675605773926,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03419092297554016,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.346864223480225,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02461826801300049,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7236275672912598,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.5859372019767761,
+ "eval_runtime": 59.0139,
+ "eval_samples_per_second": 41.38,
+ "eval_steps_per_second": 0.661,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013961605584642,
+ "grad_norm": 0.039559684693813324,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 4.289841651916504,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027923211169284,
+ "grad_norm": 0.05554218962788582,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 4.255925178527832,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041884816753926,
+ "grad_norm": 0.05330996215343475,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 4.272125244140625,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055846422338568,
+ "grad_norm": 0.053247954696416855,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 4.290765285491943,
+ "step": 1084
+ },
+ {
+ "epoch": 15.06980802792321,
+ "grad_norm": 0.05471150949597359,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 4.271744728088379,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083769633507853,
+ "grad_norm": 0.05976719409227371,
+ "learning_rate": 0.000491285979372924,
+ "loss": 4.28203821182251,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097731239092496,
+ "grad_norm": 0.05369880795478821,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 4.275313377380371,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111692844677139,
+ "grad_norm": 0.052529267966747284,
+ "learning_rate": 0.000490873338088198,
+ "loss": 4.25678825378418,
+ "step": 1088
+ },
+ {
+ "epoch": 15.12565445026178,
+ "grad_norm": 0.05184522643685341,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 4.260510444641113,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139616055846423,
+ "grad_norm": 0.051606856286525726,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 4.283233642578125,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153577661431065,
+ "grad_norm": 0.05318903923034668,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 4.286595344543457,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167539267015707,
+ "grad_norm": 0.058525051921606064,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 4.282374382019043,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181500872600349,
+ "grad_norm": 0.054776228964328766,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 4.2913126945495605,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195462478184991,
+ "grad_norm": 0.04907633364200592,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 4.297187805175781,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209424083769633,
+ "grad_norm": 0.05163136497139931,
+ "learning_rate": 0.000489424334303338,
+ "loss": 4.2968645095825195,
+ "step": 1095
+ },
+ {
+ "epoch": 15.223385689354275,
+ "grad_norm": 0.053157489746809006,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 4.285839080810547,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237347294938917,
+ "grad_norm": 0.05251535773277283,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 4.311246871948242,
+ "step": 1097
+ },
+ {
+ "epoch": 15.25130890052356,
+ "grad_norm": 0.048785097897052765,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 4.300364971160889,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265270506108202,
+ "grad_norm": 0.050393108278512955,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 4.299890995025635,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279232111692846,
+ "grad_norm": 0.04894037917256355,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 4.312449932098389,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293193717277488,
+ "grad_norm": 0.05274952948093414,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 4.2889814376831055,
+ "step": 1101
+ },
+ {
+ "epoch": 15.30715532286213,
+ "grad_norm": 0.05946570262312889,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 4.2731170654296875,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321116928446772,
+ "grad_norm": 0.054245397448539734,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 4.272463798522949,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335078534031414,
+ "grad_norm": 0.042092181742191315,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 4.299953937530518,
+ "step": 1104
+ },
+ {
+ "epoch": 15.349040139616056,
+ "grad_norm": 0.040222618728876114,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 4.292524814605713,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363001745200698,
+ "grad_norm": 0.04298696666955948,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 4.31342887878418,
+ "step": 1106
+ },
+ {
+ "epoch": 15.37696335078534,
+ "grad_norm": 0.04271050915122032,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 4.277663707733154,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390924956369982,
+ "grad_norm": 0.042364370077848434,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 4.311102390289307,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404886561954624,
+ "grad_norm": 0.04343319311738014,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 4.311610221862793,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418848167539267,
+ "grad_norm": 0.041321512311697006,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 4.29677677154541,
+ "step": 1110
+ },
+ {
+ "epoch": 15.432809773123909,
+ "grad_norm": 0.04107942432165146,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 4.3163018226623535,
+ "step": 1111
+ },
+ {
+ "epoch": 15.44677137870855,
+ "grad_norm": 0.03844515234231949,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 4.311301231384277,
+ "step": 1112
+ },
+ {
+ "epoch": 15.460732984293193,
+ "grad_norm": 0.03731007128953934,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 4.31325626373291,
+ "step": 1113
+ },
+ {
+ "epoch": 15.474694589877837,
+ "grad_norm": 0.03807826340198517,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 4.296201705932617,
+ "step": 1114
+ },
+ {
+ "epoch": 15.488656195462479,
+ "grad_norm": 0.040137600153684616,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 4.320395469665527,
+ "step": 1115
+ },
+ {
+ "epoch": 15.502617801047121,
+ "grad_norm": 0.038484394550323486,
+ "learning_rate": 0.000485033394934934,
+ "loss": 4.313858985900879,
+ "step": 1116
+ },
+ {
+ "epoch": 15.516579406631763,
+ "grad_norm": 0.03603338822722435,
+ "learning_rate": 0.000484822676912974,
+ "loss": 4.297666549682617,
+ "step": 1117
+ },
+ {
+ "epoch": 15.530541012216405,
+ "grad_norm": 0.038647182285785675,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 4.305002689361572,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544502617801047,
+ "grad_norm": 0.03907139599323273,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 4.329583644866943,
+ "step": 1119
+ },
+ {
+ "epoch": 15.55846422338569,
+ "grad_norm": 0.03621622920036316,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 4.290896415710449,
+ "step": 1120
+ },
+ {
+ "epoch": 15.572425828970331,
+ "grad_norm": 0.037496887147426605,
+ "learning_rate": 0.000483978341612154,
+ "loss": 4.33336877822876,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586387434554974,
+ "grad_norm": 0.03778311610221863,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 4.285354137420654,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600349040139616,
+ "grad_norm": 0.03782997280359268,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 4.29697847366333,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614310645724258,
+ "grad_norm": 0.03665657714009285,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 4.320103168487549,
+ "step": 1124
+ },
+ {
+ "epoch": 15.6282722513089,
+ "grad_norm": 0.03796127438545227,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 4.317153453826904,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642233856893542,
+ "grad_norm": 0.036425214260816574,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 4.293642520904541,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656195462478184,
+ "grad_norm": 0.03629784658551216,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 4.291100025177002,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670157068062828,
+ "grad_norm": 0.03674762696027756,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 4.295144557952881,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68411867364747,
+ "grad_norm": 0.036291178315877914,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 4.313133239746094,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698080279232112,
+ "grad_norm": 0.03817860782146454,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 4.294533729553223,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712041884816754,
+ "grad_norm": 0.036819685250520706,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 4.299263954162598,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726003490401396,
+ "grad_norm": 0.035103823989629745,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 4.287391185760498,
+ "step": 1132
+ },
+ {
+ "epoch": 15.739965095986038,
+ "grad_norm": 0.035701096057891846,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 4.318216323852539,
+ "step": 1133
+ },
+ {
+ "epoch": 15.75392670157068,
+ "grad_norm": 0.03639856353402138,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 4.301465034484863,
+ "step": 1134
+ },
+ {
+ "epoch": 15.767888307155323,
+ "grad_norm": 0.03593013063073158,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 4.311532974243164,
+ "step": 1135
+ },
+ {
+ "epoch": 15.781849912739965,
+ "grad_norm": 0.03564770147204399,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 4.3400726318359375,
+ "step": 1136
+ },
+ {
+ "epoch": 15.795811518324607,
+ "grad_norm": 0.03598494827747345,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 4.30976676940918,
+ "step": 1137
+ },
+ {
+ "epoch": 15.809773123909249,
+ "grad_norm": 0.034391652792692184,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 4.328464508056641,
+ "step": 1138
+ },
+ {
+ "epoch": 15.823734729493891,
+ "grad_norm": 0.03445672616362572,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 4.310579299926758,
+ "step": 1139
+ },
+ {
+ "epoch": 15.837696335078533,
+ "grad_norm": 0.03643282502889633,
+ "learning_rate": 0.000479936075927473,
+ "loss": 4.320905685424805,
+ "step": 1140
+ },
+ {
+ "epoch": 15.851657940663177,
+ "grad_norm": 0.03489850088953972,
+ "learning_rate": 0.000479721889242305,
+ "loss": 4.317229270935059,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86561954624782,
+ "grad_norm": 0.03609171137213707,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 4.309203147888184,
+ "step": 1142
+ },
+ {
+ "epoch": 15.879581151832461,
+ "grad_norm": 0.03705338388681412,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 4.287387371063232,
+ "step": 1143
+ },
+ {
+ "epoch": 15.893542757417103,
+ "grad_norm": 0.03697717934846878,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 4.302860260009766,
+ "step": 1144
+ },
+ {
+ "epoch": 15.907504363001745,
+ "grad_norm": 0.035338208079338074,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 4.3152265548706055,
+ "step": 1145
+ },
+ {
+ "epoch": 15.921465968586388,
+ "grad_norm": 0.03548634052276611,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 4.327335357666016,
+ "step": 1146
+ },
+ {
+ "epoch": 15.93542757417103,
+ "grad_norm": 0.035512328147888184,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 4.3067803382873535,
+ "step": 1147
+ },
+ {
+ "epoch": 15.949389179755672,
+ "grad_norm": 0.03799289092421532,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 4.320701599121094,
+ "step": 1148
+ },
+ {
+ "epoch": 15.963350785340314,
+ "grad_norm": 0.03866339474916458,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 4.321087837219238,
+ "step": 1149
+ },
+ {
+ "epoch": 15.977312390924956,
+ "grad_norm": 0.036002084612846375,
+ "learning_rate": 0.000477787822648007,
+ "loss": 4.287045478820801,
+ "step": 1150
+ },
+ {
+ "epoch": 15.991273996509598,
+ "grad_norm": 0.03433441370725632,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 4.2562665939331055,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.02580358274281025,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 2.67783260345459,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.5870338082313538,
+ "eval_runtime": 60.0469,
+ "eval_samples_per_second": 40.668,
+ "eval_steps_per_second": 0.649,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013961605584644,
+ "grad_norm": 0.04105810821056366,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 4.224295616149902,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027923211169284,
+ "grad_norm": 0.049538832157850266,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 4.222184181213379,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041884816753928,
+ "grad_norm": 0.05089569091796875,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 4.2067060470581055,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05584642233857,
+ "grad_norm": 0.04600589722394943,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 4.238527774810791,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069808027923212,
+ "grad_norm": 0.043311409652233124,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 4.242960453033447,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083769633507853,
+ "grad_norm": 0.047897014766931534,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 4.199069023132324,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097731239092496,
+ "grad_norm": 0.05633604899048805,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 4.211665153503418,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111692844677137,
+ "grad_norm": 0.056110311299562454,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 4.20163631439209,
+ "step": 1160
+ },
+ {
+ "epoch": 16.12565445026178,
+ "grad_norm": 0.05426819995045662,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 4.21889591217041,
+ "step": 1161
+ },
+ {
+ "epoch": 16.13961605584642,
+ "grad_norm": 0.05489835515618324,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 4.220406532287598,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153577661431065,
+ "grad_norm": 0.053941354155540466,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 4.231909275054932,
+ "step": 1163
+ },
+ {
+ "epoch": 16.167539267015705,
+ "grad_norm": 0.058548808097839355,
+ "learning_rate": 0.000474756648870413,
+ "loss": 4.237003326416016,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18150087260035,
+ "grad_norm": 0.05370597913861275,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 4.240085124969482,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195462478184993,
+ "grad_norm": 0.05684991553425789,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 4.203842639923096,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209424083769633,
+ "grad_norm": 0.060745254158973694,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 4.247998237609863,
+ "step": 1167
+ },
+ {
+ "epoch": 16.223385689354277,
+ "grad_norm": 0.058296725153923035,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 4.226161003112793,
+ "step": 1168
+ },
+ {
+ "epoch": 16.237347294938917,
+ "grad_norm": 0.05310584232211113,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 4.229870796203613,
+ "step": 1169
+ },
+ {
+ "epoch": 16.25130890052356,
+ "grad_norm": 0.04710390046238899,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 4.240084648132324,
+ "step": 1170
+ },
+ {
+ "epoch": 16.2652705061082,
+ "grad_norm": 0.05003240332007408,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 4.240972518920898,
+ "step": 1171
+ },
+ {
+ "epoch": 16.279232111692846,
+ "grad_norm": 0.05552851781249046,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 4.248024940490723,
+ "step": 1172
+ },
+ {
+ "epoch": 16.293193717277486,
+ "grad_norm": 0.049967192113399506,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 4.279229164123535,
+ "step": 1173
+ },
+ {
+ "epoch": 16.30715532286213,
+ "grad_norm": 0.04883681237697601,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 4.269807815551758,
+ "step": 1174
+ },
+ {
+ "epoch": 16.32111692844677,
+ "grad_norm": 0.04931480810046196,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 4.232419490814209,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335078534031414,
+ "grad_norm": 0.048746511340141296,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 4.267911911010742,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349040139616054,
+ "grad_norm": 0.04678189381957054,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 4.262979507446289,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363001745200698,
+ "grad_norm": 0.048893529921770096,
+ "learning_rate": 0.000471698352726896,
+ "loss": 4.254408359527588,
+ "step": 1178
+ },
+ {
+ "epoch": 16.376963350785342,
+ "grad_norm": 0.045027732849121094,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 4.2445902824401855,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390924956369982,
+ "grad_norm": 0.04324870929121971,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 4.2685136795043945,
+ "step": 1180
+ },
+ {
+ "epoch": 16.404886561954626,
+ "grad_norm": 0.0470522940158844,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 4.237505912780762,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418848167539267,
+ "grad_norm": 0.04896962642669678,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 4.2523345947265625,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43280977312391,
+ "grad_norm": 0.04562787711620331,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 4.2192792892456055,
+ "step": 1183
+ },
+ {
+ "epoch": 16.44677137870855,
+ "grad_norm": 0.04414575919508934,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 4.250238418579102,
+ "step": 1184
+ },
+ {
+ "epoch": 16.460732984293195,
+ "grad_norm": 0.04363945871591568,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 4.256250381469727,
+ "step": 1185
+ },
+ {
+ "epoch": 16.474694589877835,
+ "grad_norm": 0.04159797728061676,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 4.274903297424316,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48865619546248,
+ "grad_norm": 0.04248544201254845,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 4.278222560882568,
+ "step": 1187
+ },
+ {
+ "epoch": 16.50261780104712,
+ "grad_norm": 0.04379890486598015,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 4.267505645751953,
+ "step": 1188
+ },
+ {
+ "epoch": 16.516579406631763,
+ "grad_norm": 0.04319411888718605,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 4.2607269287109375,
+ "step": 1189
+ },
+ {
+ "epoch": 16.530541012216403,
+ "grad_norm": 0.04143810644745827,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 4.267977714538574,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544502617801047,
+ "grad_norm": 0.03914840891957283,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 4.260498523712158,
+ "step": 1191
+ },
+ {
+ "epoch": 16.55846422338569,
+ "grad_norm": 0.04107543081045151,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 4.2550482749938965,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57242582897033,
+ "grad_norm": 0.041316065937280655,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 4.3075456619262695,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586387434554975,
+ "grad_norm": 0.04069342464208603,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 4.234392166137695,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600349040139616,
+ "grad_norm": 0.03816857561469078,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 4.251038551330566,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61431064572426,
+ "grad_norm": 0.04095374792814255,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 4.284790992736816,
+ "step": 1196
+ },
+ {
+ "epoch": 16.6282722513089,
+ "grad_norm": 0.03788160905241966,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 4.255716323852539,
+ "step": 1197
+ },
+ {
+ "epoch": 16.642233856893544,
+ "grad_norm": 0.03784061595797539,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 4.269606590270996,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656195462478184,
+ "grad_norm": 0.04087168723344803,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 4.231165885925293,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670157068062828,
+ "grad_norm": 0.040546853095293045,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 4.288422584533691,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68411867364747,
+ "grad_norm": 0.03777262568473816,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 4.277492523193359,
+ "step": 1201
+ },
+ {
+ "epoch": 16.698080279232112,
+ "grad_norm": 0.039857733994722366,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 4.239383697509766,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712041884816752,
+ "grad_norm": 0.04387792944908142,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 4.300265312194824,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726003490401396,
+ "grad_norm": 0.04214833676815033,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 4.255545616149902,
+ "step": 1204
+ },
+ {
+ "epoch": 16.739965095986037,
+ "grad_norm": 0.04290056228637695,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 4.275530815124512,
+ "step": 1205
+ },
+ {
+ "epoch": 16.75392670157068,
+ "grad_norm": 0.03975554555654526,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 4.282685279846191,
+ "step": 1206
+ },
+ {
+ "epoch": 16.767888307155324,
+ "grad_norm": 0.039464011788368225,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 4.285504341125488,
+ "step": 1207
+ },
+ {
+ "epoch": 16.781849912739965,
+ "grad_norm": 0.039714012295007706,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 4.273477554321289,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79581151832461,
+ "grad_norm": 0.03936111181974411,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 4.304205894470215,
+ "step": 1209
+ },
+ {
+ "epoch": 16.80977312390925,
+ "grad_norm": 0.03782064467668533,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 4.307382583618164,
+ "step": 1210
+ },
+ {
+ "epoch": 16.823734729493893,
+ "grad_norm": 0.040340717881917953,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 4.308620452880859,
+ "step": 1211
+ },
+ {
+ "epoch": 16.837696335078533,
+ "grad_norm": 0.0403442308306694,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 4.248626708984375,
+ "step": 1212
+ },
+ {
+ "epoch": 16.851657940663177,
+ "grad_norm": 0.038351960480213165,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 4.261248588562012,
+ "step": 1213
+ },
+ {
+ "epoch": 16.865619546247817,
+ "grad_norm": 0.03965333476662636,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 4.278291702270508,
+ "step": 1214
+ },
+ {
+ "epoch": 16.87958115183246,
+ "grad_norm": 0.040909092873334885,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 4.241847991943359,
+ "step": 1215
+ },
+ {
+ "epoch": 16.8935427574171,
+ "grad_norm": 0.038684047758579254,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 4.281346321105957,
+ "step": 1216
+ },
+ {
+ "epoch": 16.907504363001745,
+ "grad_norm": 0.03698103874921799,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 4.3085432052612305,
+ "step": 1217
+ },
+ {
+ "epoch": 16.921465968586386,
+ "grad_norm": 0.03908613696694374,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 4.24220085144043,
+ "step": 1218
+ },
+ {
+ "epoch": 16.93542757417103,
+ "grad_norm": 0.03873388469219208,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 4.282486915588379,
+ "step": 1219
+ },
+ {
+ "epoch": 16.949389179755673,
+ "grad_norm": 0.03709883987903595,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 4.279009819030762,
+ "step": 1220
+ },
+ {
+ "epoch": 16.963350785340314,
+ "grad_norm": 0.03937513753771782,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 4.273211479187012,
+ "step": 1221
+ },
+ {
+ "epoch": 16.977312390924958,
+ "grad_norm": 0.038478728383779526,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 4.281154155731201,
+ "step": 1222
+ },
+ {
+ "epoch": 16.991273996509598,
+ "grad_norm": 0.03913680836558342,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 4.269218444824219,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.02641749382019043,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 2.6905035972595215,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.5879416465759277,
+ "eval_runtime": 60.2807,
+ "eval_samples_per_second": 40.511,
+ "eval_steps_per_second": 0.647,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013961605584644,
+ "grad_norm": 0.048571839928627014,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 4.2008538246154785,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027923211169284,
+ "grad_norm": 0.0640900656580925,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 4.180826663970947,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041884816753928,
+ "grad_norm": 0.06458134204149246,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 4.199562072753906,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05584642233857,
+ "grad_norm": 0.0563119500875473,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 4.171557426452637,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069808027923212,
+ "grad_norm": 0.050914082676172256,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 4.183539390563965,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083769633507853,
+ "grad_norm": 0.053884491324424744,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 4.210450172424316,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097731239092496,
+ "grad_norm": 0.05301345884799957,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 4.1553192138671875,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111692844677137,
+ "grad_norm": 0.051138825714588165,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 4.158510208129883,
+ "step": 1232
+ },
+ {
+ "epoch": 17.12565445026178,
+ "grad_norm": 0.05355675891041756,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 4.167689323425293,
+ "step": 1233
+ },
+ {
+ "epoch": 17.13961605584642,
+ "grad_norm": 0.06062275916337967,
+ "learning_rate": 0.000459203665939781,
+ "loss": 4.170838356018066,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153577661431065,
+ "grad_norm": 0.058666616678237915,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 4.187441825866699,
+ "step": 1235
+ },
+ {
+ "epoch": 17.167539267015705,
+ "grad_norm": 0.057622168213129044,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 4.190746784210205,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18150087260035,
+ "grad_norm": 0.069090336561203,
+ "learning_rate": 0.000458522886595554,
+ "loss": 4.224946022033691,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195462478184993,
+ "grad_norm": 0.06552847474813461,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 4.204067230224609,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209424083769633,
+ "grad_norm": 0.06602177768945694,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 4.189286708831787,
+ "step": 1239
+ },
+ {
+ "epoch": 17.223385689354277,
+ "grad_norm": 0.06751416623592377,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 4.208927154541016,
+ "step": 1240
+ },
+ {
+ "epoch": 17.237347294938917,
+ "grad_norm": 0.0706794261932373,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 4.212923049926758,
+ "step": 1241
+ },
+ {
+ "epoch": 17.25130890052356,
+ "grad_norm": 0.05346836522221565,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 4.2221784591674805,
+ "step": 1242
+ },
+ {
+ "epoch": 17.2652705061082,
+ "grad_norm": 0.05395734682679176,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 4.196428298950195,
+ "step": 1243
+ },
+ {
+ "epoch": 17.279232111692846,
+ "grad_norm": 0.05496550351381302,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 4.208031177520752,
+ "step": 1244
+ },
+ {
+ "epoch": 17.293193717277486,
+ "grad_norm": 0.05320895463228226,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 4.232810974121094,
+ "step": 1245
+ },
+ {
+ "epoch": 17.30715532286213,
+ "grad_norm": 0.05064759403467178,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 4.207826614379883,
+ "step": 1246
+ },
+ {
+ "epoch": 17.32111692844677,
+ "grad_norm": 0.04287239536643028,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 4.195648670196533,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335078534031414,
+ "grad_norm": 0.04527301341295242,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 4.19954776763916,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349040139616054,
+ "grad_norm": 0.04421805590391159,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 4.248671054840088,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363001745200698,
+ "grad_norm": 0.04487175494432449,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 4.2072954177856445,
+ "step": 1250
+ },
+ {
+ "epoch": 17.376963350785342,
+ "grad_norm": 0.04415794834494591,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 4.222837448120117,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390924956369982,
+ "grad_norm": 0.04474013298749924,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 4.213104248046875,
+ "step": 1252
+ },
+ {
+ "epoch": 17.404886561954626,
+ "grad_norm": 0.042631037533283234,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 4.209726333618164,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418848167539267,
+ "grad_norm": 0.04424271360039711,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 4.202699661254883,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43280977312391,
+ "grad_norm": 0.04151812940835953,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 4.190732955932617,
+ "step": 1255
+ },
+ {
+ "epoch": 17.44677137870855,
+ "grad_norm": 0.041588131338357925,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 4.222360610961914,
+ "step": 1256
+ },
+ {
+ "epoch": 17.460732984293195,
+ "grad_norm": 0.042378414422273636,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 4.198424339294434,
+ "step": 1257
+ },
+ {
+ "epoch": 17.474694589877835,
+ "grad_norm": 0.041907574981451035,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 4.224804401397705,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48865619546248,
+ "grad_norm": 0.043774355202913284,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 4.2276611328125,
+ "step": 1259
+ },
+ {
+ "epoch": 17.50261780104712,
+ "grad_norm": 0.04238573834300041,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 4.228109359741211,
+ "step": 1260
+ },
+ {
+ "epoch": 17.516579406631763,
+ "grad_norm": 0.045351386070251465,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 4.208807468414307,
+ "step": 1261
+ },
+ {
+ "epoch": 17.530541012216403,
+ "grad_norm": 0.04300729185342789,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 4.257306098937988,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544502617801047,
+ "grad_norm": 0.041442908346652985,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 4.208004951477051,
+ "step": 1263
+ },
+ {
+ "epoch": 17.55846422338569,
+ "grad_norm": 0.04337337985634804,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 4.23393440246582,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57242582897033,
+ "grad_norm": 0.04335462674498558,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 4.228087425231934,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586387434554975,
+ "grad_norm": 0.03987223282456398,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 4.213971138000488,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600349040139616,
+ "grad_norm": 0.041760094463825226,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 4.248366355895996,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61431064572426,
+ "grad_norm": 0.04420867934823036,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 4.278865814208984,
+ "step": 1268
+ },
+ {
+ "epoch": 17.6282722513089,
+ "grad_norm": 0.04244982823729515,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 4.223642349243164,
+ "step": 1269
+ },
+ {
+ "epoch": 17.642233856893544,
+ "grad_norm": 0.04280664771795273,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 4.231023788452148,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656195462478184,
+ "grad_norm": 0.04294133558869362,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 4.222380638122559,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670157068062828,
+ "grad_norm": 0.04179249703884125,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 4.215641975402832,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68411867364747,
+ "grad_norm": 0.04038798063993454,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 4.228580474853516,
+ "step": 1273
+ },
+ {
+ "epoch": 17.698080279232112,
+ "grad_norm": 0.04211528226733208,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 4.245234489440918,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712041884816752,
+ "grad_norm": 0.04102989658713341,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 4.218495845794678,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726003490401396,
+ "grad_norm": 0.04038692265748978,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 4.2466511726379395,
+ "step": 1276
+ },
+ {
+ "epoch": 17.739965095986037,
+ "grad_norm": 0.03958791121840477,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 4.237762451171875,
+ "step": 1277
+ },
+ {
+ "epoch": 17.75392670157068,
+ "grad_norm": 0.04129907861351967,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 4.238739013671875,
+ "step": 1278
+ },
+ {
+ "epoch": 17.767888307155324,
+ "grad_norm": 0.042581796646118164,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 4.235558032989502,
+ "step": 1279
+ },
+ {
+ "epoch": 17.781849912739965,
+ "grad_norm": 0.04339978098869324,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 4.226593971252441,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79581151832461,
+ "grad_norm": 0.03808023780584335,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 4.238213062286377,
+ "step": 1281
+ },
+ {
+ "epoch": 17.80977312390925,
+ "grad_norm": 0.04022229090332985,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 4.214224815368652,
+ "step": 1282
+ },
+ {
+ "epoch": 17.823734729493893,
+ "grad_norm": 0.04379156231880188,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 4.253068447113037,
+ "step": 1283
+ },
+ {
+ "epoch": 17.837696335078533,
+ "grad_norm": 0.04232964664697647,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 4.249903202056885,
+ "step": 1284
+ },
+ {
+ "epoch": 17.851657940663177,
+ "grad_norm": 0.040980905294418335,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 4.252640724182129,
+ "step": 1285
+ },
+ {
+ "epoch": 17.865619546247817,
+ "grad_norm": 0.04015526548027992,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 4.234862327575684,
+ "step": 1286
+ },
+ {
+ "epoch": 17.87958115183246,
+ "grad_norm": 0.04273828864097595,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 4.216028213500977,
+ "step": 1287
+ },
+ {
+ "epoch": 17.8935427574171,
+ "grad_norm": 0.04210563376545906,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 4.23359489440918,
+ "step": 1288
+ },
+ {
+ "epoch": 17.907504363001745,
+ "grad_norm": 0.038766685873270035,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 4.220360279083252,
+ "step": 1289
+ },
+ {
+ "epoch": 17.921465968586386,
+ "grad_norm": 0.03921609744429588,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 4.253805160522461,
+ "step": 1290
+ },
+ {
+ "epoch": 17.93542757417103,
+ "grad_norm": 0.04071544483304024,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 4.211221694946289,
+ "step": 1291
+ },
+ {
+ "epoch": 17.949389179755673,
+ "grad_norm": 0.03996877372264862,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 4.241739273071289,
+ "step": 1292
+ },
+ {
+ "epoch": 17.963350785340314,
+ "grad_norm": 0.03878853842616081,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 4.237001419067383,
+ "step": 1293
+ },
+ {
+ "epoch": 17.977312390924958,
+ "grad_norm": 0.03889666125178337,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 4.244929790496826,
+ "step": 1294
+ },
+ {
+ "epoch": 17.991273996509598,
+ "grad_norm": 0.03876441717147827,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 4.236083030700684,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.027286000549793243,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 2.636739730834961,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.5896694660186768,
+ "eval_runtime": 60.4742,
+ "eval_samples_per_second": 40.381,
+ "eval_steps_per_second": 0.645,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013961605584644,
+ "grad_norm": 0.04279542714357376,
+ "learning_rate": 0.000444677103403854,
+ "loss": 4.136401653289795,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027923211169284,
+ "grad_norm": 0.051685452461242676,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 4.159584045410156,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041884816753928,
+ "grad_norm": 0.05411457270383835,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 4.160367965698242,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05584642233857,
+ "grad_norm": 0.048574138432741165,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 4.134189128875732,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069808027923212,
+ "grad_norm": 0.04528478533029556,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 4.1022725105285645,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083769633507853,
+ "grad_norm": 0.04474090039730072,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 4.16152286529541,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097731239092496,
+ "grad_norm": 0.04882791265845299,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 4.1553144454956055,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111692844677137,
+ "grad_norm": 0.04831654578447342,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 4.147022724151611,
+ "step": 1304
+ },
+ {
+ "epoch": 18.12565445026178,
+ "grad_norm": 0.047365427017211914,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 4.139405250549316,
+ "step": 1305
+ },
+ {
+ "epoch": 18.13961605584642,
+ "grad_norm": 0.0514012910425663,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 4.115568161010742,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153577661431065,
+ "grad_norm": 0.05225161090493202,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 4.166429042816162,
+ "step": 1307
+ },
+ {
+ "epoch": 18.167539267015705,
+ "grad_norm": 0.05575249716639519,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 4.1414289474487305,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18150087260035,
+ "grad_norm": 0.05405383184552193,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 4.138680934906006,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195462478184993,
+ "grad_norm": 0.05185031518340111,
+ "learning_rate": 0.000441621628895411,
+ "loss": 4.159243583679199,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209424083769633,
+ "grad_norm": 0.05070916935801506,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 4.1844892501831055,
+ "step": 1311
+ },
+ {
+ "epoch": 18.223385689354277,
+ "grad_norm": 0.05037331581115723,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 4.132781028747559,
+ "step": 1312
+ },
+ {
+ "epoch": 18.237347294938917,
+ "grad_norm": 0.05082111805677414,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 4.143031120300293,
+ "step": 1313
+ },
+ {
+ "epoch": 18.25130890052356,
+ "grad_norm": 0.04742169380187988,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 4.137580871582031,
+ "step": 1314
+ },
+ {
+ "epoch": 18.2652705061082,
+ "grad_norm": 0.049042731523513794,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 4.159721374511719,
+ "step": 1315
+ },
+ {
+ "epoch": 18.279232111692846,
+ "grad_norm": 0.045649368315935135,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 4.144008636474609,
+ "step": 1316
+ },
+ {
+ "epoch": 18.293193717277486,
+ "grad_norm": 0.0473787859082222,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 4.164243698120117,
+ "step": 1317
+ },
+ {
+ "epoch": 18.30715532286213,
+ "grad_norm": 0.047626055777072906,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 4.164441108703613,
+ "step": 1318
+ },
+ {
+ "epoch": 18.32111692844677,
+ "grad_norm": 0.04738573357462883,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 4.180338382720947,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335078534031414,
+ "grad_norm": 0.04712498560547829,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 4.146285533905029,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349040139616054,
+ "grad_norm": 0.045505691319704056,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 4.162572860717773,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363001745200698,
+ "grad_norm": 0.0489528588950634,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 4.160312175750732,
+ "step": 1322
+ },
+ {
+ "epoch": 18.376963350785342,
+ "grad_norm": 0.047517575323581696,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 4.164361953735352,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390924956369982,
+ "grad_norm": 0.044214535504579544,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 4.182154655456543,
+ "step": 1324
+ },
+ {
+ "epoch": 18.404886561954626,
+ "grad_norm": 0.043775733560323715,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 4.130522727966309,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418848167539267,
+ "grad_norm": 0.044364530593156815,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 4.151247024536133,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43280977312391,
+ "grad_norm": 0.046320103108882904,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 4.204159736633301,
+ "step": 1327
+ },
+ {
+ "epoch": 18.44677137870855,
+ "grad_norm": 0.04802565276622772,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 4.182550430297852,
+ "step": 1328
+ },
+ {
+ "epoch": 18.460732984293195,
+ "grad_norm": 0.049340926110744476,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 4.163658142089844,
+ "step": 1329
+ },
+ {
+ "epoch": 18.474694589877835,
+ "grad_norm": 0.046701837331056595,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 4.170445442199707,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48865619546248,
+ "grad_norm": 0.04746273159980774,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 4.220232963562012,
+ "step": 1331
+ },
+ {
+ "epoch": 18.50261780104712,
+ "grad_norm": 0.04644489288330078,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 4.190027236938477,
+ "step": 1332
+ },
+ {
+ "epoch": 18.516579406631763,
+ "grad_norm": 0.04394262284040451,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 4.159570693969727,
+ "step": 1333
+ },
+ {
+ "epoch": 18.530541012216403,
+ "grad_norm": 0.04609063267707825,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 4.1657586097717285,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544502617801047,
+ "grad_norm": 0.0450211800634861,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 4.184865474700928,
+ "step": 1335
+ },
+ {
+ "epoch": 18.55846422338569,
+ "grad_norm": 0.0475253090262413,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 4.185857772827148,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57242582897033,
+ "grad_norm": 0.04332203418016434,
+ "learning_rate": 0.000435215814386134,
+ "loss": 4.160165309906006,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586387434554975,
+ "grad_norm": 0.0427570678293705,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 4.1727495193481445,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600349040139616,
+ "grad_norm": 0.04651174694299698,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 4.204797744750977,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61431064572426,
+ "grad_norm": 0.04324701428413391,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 4.186931610107422,
+ "step": 1340
+ },
+ {
+ "epoch": 18.6282722513089,
+ "grad_norm": 0.044644393026828766,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 4.209554195404053,
+ "step": 1341
+ },
+ {
+ "epoch": 18.642233856893544,
+ "grad_norm": 0.042982228100299835,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 4.205570220947266,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656195462478184,
+ "grad_norm": 0.04385090991854668,
+ "learning_rate": 0.000433781621332752,
+ "loss": 4.177936553955078,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670157068062828,
+ "grad_norm": 0.04771510139107704,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 4.1795806884765625,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68411867364747,
+ "grad_norm": 0.04551433399319649,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 4.192194938659668,
+ "step": 1345
+ },
+ {
+ "epoch": 18.698080279232112,
+ "grad_norm": 0.043501488864421844,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 4.177781105041504,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712041884816752,
+ "grad_norm": 0.04394034296274185,
+ "learning_rate": 0.000432823382695327,
+ "loss": 4.206354141235352,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726003490401396,
+ "grad_norm": 0.04911153391003609,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 4.205874919891357,
+ "step": 1348
+ },
+ {
+ "epoch": 18.739965095986037,
+ "grad_norm": 0.04602045193314552,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 4.1947021484375,
+ "step": 1349
+ },
+ {
+ "epoch": 18.75392670157068,
+ "grad_norm": 0.04393904283642769,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 4.188337802886963,
+ "step": 1350
+ },
+ {
+ "epoch": 18.767888307155324,
+ "grad_norm": 0.04526393488049507,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 4.175235748291016,
+ "step": 1351
+ },
+ {
+ "epoch": 18.781849912739965,
+ "grad_norm": 0.041930247098207474,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 4.210986137390137,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79581151832461,
+ "grad_norm": 0.044262565672397614,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 4.176301002502441,
+ "step": 1353
+ },
+ {
+ "epoch": 18.80977312390925,
+ "grad_norm": 0.044190581887960434,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 4.1888837814331055,
+ "step": 1354
+ },
+ {
+ "epoch": 18.823734729493893,
+ "grad_norm": 0.045371655374765396,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 4.182244777679443,
+ "step": 1355
+ },
+ {
+ "epoch": 18.837696335078533,
+ "grad_norm": 0.04302282631397247,
+ "learning_rate": 0.000430661245733797,
+ "loss": 4.189350128173828,
+ "step": 1356
+ },
+ {
+ "epoch": 18.851657940663177,
+ "grad_norm": 0.04402372986078262,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 4.233372211456299,
+ "step": 1357
+ },
+ {
+ "epoch": 18.865619546247817,
+ "grad_norm": 0.04397045075893402,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 4.21682071685791,
+ "step": 1358
+ },
+ {
+ "epoch": 18.87958115183246,
+ "grad_norm": 0.04319734871387482,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 4.204183578491211,
+ "step": 1359
+ },
+ {
+ "epoch": 18.8935427574171,
+ "grad_norm": 0.04600509628653526,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 4.206524848937988,
+ "step": 1360
+ },
+ {
+ "epoch": 18.907504363001745,
+ "grad_norm": 0.04364705830812454,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 4.213656425476074,
+ "step": 1361
+ },
+ {
+ "epoch": 18.921465968586386,
+ "grad_norm": 0.044004205614328384,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 4.1841583251953125,
+ "step": 1362
+ },
+ {
+ "epoch": 18.93542757417103,
+ "grad_norm": 0.04480595886707306,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 4.166214942932129,
+ "step": 1363
+ },
+ {
+ "epoch": 18.949389179755673,
+ "grad_norm": 0.04260887950658798,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 4.198726654052734,
+ "step": 1364
+ },
+ {
+ "epoch": 18.963350785340314,
+ "grad_norm": 0.04513927176594734,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 4.2122392654418945,
+ "step": 1365
+ },
+ {
+ "epoch": 18.977312390924958,
+ "grad_norm": 0.04286042973399162,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 4.192383766174316,
+ "step": 1366
+ },
+ {
+ "epoch": 18.991273996509598,
+ "grad_norm": 0.04287346079945564,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 4.221860408782959,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.030670644715428352,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 2.624823570251465,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.5910831689834595,
+ "eval_runtime": 60.5517,
+ "eval_samples_per_second": 40.329,
+ "eval_steps_per_second": 0.644,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013961605584644,
+ "grad_norm": 0.046196095645427704,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 4.104501724243164,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027923211169284,
+ "grad_norm": 0.05348595231771469,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 4.109098434448242,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041884816753928,
+ "grad_norm": 0.05256905406713486,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 4.115832328796387,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05584642233857,
+ "grad_norm": 0.05241430550813675,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 4.079610824584961,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069808027923212,
+ "grad_norm": 0.04983370006084442,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 4.095335006713867,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083769633507853,
+ "grad_norm": 0.04900007322430611,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 4.093306541442871,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097731239092496,
+ "grad_norm": 0.04995666816830635,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 4.111114501953125,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111692844677137,
+ "grad_norm": 0.04875681549310684,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 4.091944694519043,
+ "step": 1376
+ },
+ {
+ "epoch": 19.12565445026178,
+ "grad_norm": 0.05283811315894127,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 4.112549304962158,
+ "step": 1377
+ },
+ {
+ "epoch": 19.13961605584642,
+ "grad_norm": 0.04801992326974869,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 4.108818054199219,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153577661431065,
+ "grad_norm": 0.04751930385828018,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 4.089441299438477,
+ "step": 1379
+ },
+ {
+ "epoch": 19.167539267015705,
+ "grad_norm": 0.045653726905584335,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 4.07722282409668,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18150087260035,
+ "grad_norm": 0.045616209506988525,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 4.097578048706055,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195462478184993,
+ "grad_norm": 0.04929365590214729,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 4.082221984863281,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209424083769633,
+ "grad_norm": 0.04915834963321686,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 4.123330116271973,
+ "step": 1383
+ },
+ {
+ "epoch": 19.223385689354277,
+ "grad_norm": 0.0470132976770401,
+ "learning_rate": 0.000423881957237287,
+ "loss": 4.113927841186523,
+ "step": 1384
+ },
+ {
+ "epoch": 19.237347294938917,
+ "grad_norm": 0.04695259779691696,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 4.096814155578613,
+ "step": 1385
+ },
+ {
+ "epoch": 19.25130890052356,
+ "grad_norm": 0.04730398207902908,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 4.122525215148926,
+ "step": 1386
+ },
+ {
+ "epoch": 19.2652705061082,
+ "grad_norm": 0.050295207649469376,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 4.117993354797363,
+ "step": 1387
+ },
+ {
+ "epoch": 19.279232111692846,
+ "grad_norm": 0.05082492157816887,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 4.14082145690918,
+ "step": 1388
+ },
+ {
+ "epoch": 19.293193717277486,
+ "grad_norm": 0.04960232973098755,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 4.139887809753418,
+ "step": 1389
+ },
+ {
+ "epoch": 19.30715532286213,
+ "grad_norm": 0.04851173236966133,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 4.138492584228516,
+ "step": 1390
+ },
+ {
+ "epoch": 19.32111692844677,
+ "grad_norm": 0.047838468104600906,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 4.118064880371094,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335078534031414,
+ "grad_norm": 0.05148220434784889,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 4.136533737182617,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349040139616054,
+ "grad_norm": 0.053177181631326675,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 4.1125168800354,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363001745200698,
+ "grad_norm": 0.04861171543598175,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 4.1437153816223145,
+ "step": 1394
+ },
+ {
+ "epoch": 19.376963350785342,
+ "grad_norm": 0.05194924771785736,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 4.097813606262207,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390924956369982,
+ "grad_norm": 0.049609847366809845,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 4.068940162658691,
+ "step": 1396
+ },
+ {
+ "epoch": 19.404886561954626,
+ "grad_norm": 0.047042325139045715,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 4.121333122253418,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418848167539267,
+ "grad_norm": 0.047646693885326385,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 4.114943981170654,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43280977312391,
+ "grad_norm": 0.04676460474729538,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 4.139252185821533,
+ "step": 1399
+ },
+ {
+ "epoch": 19.44677137870855,
+ "grad_norm": 0.04905956983566284,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 4.135970115661621,
+ "step": 1400
+ },
+ {
+ "epoch": 19.460732984293195,
+ "grad_norm": 0.050025034695863724,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 4.15662956237793,
+ "step": 1401
+ },
+ {
+ "epoch": 19.474694589877835,
+ "grad_norm": 0.04740425944328308,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 4.140295505523682,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48865619546248,
+ "grad_norm": 0.046904709190130234,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 4.14083194732666,
+ "step": 1403
+ },
+ {
+ "epoch": 19.50261780104712,
+ "grad_norm": 0.048393651843070984,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 4.161823272705078,
+ "step": 1404
+ },
+ {
+ "epoch": 19.516579406631763,
+ "grad_norm": 0.04766222462058067,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 4.151817798614502,
+ "step": 1405
+ },
+ {
+ "epoch": 19.530541012216403,
+ "grad_norm": 0.04797709733247757,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 4.173160552978516,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544502617801047,
+ "grad_norm": 0.05012711137533188,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 4.161989688873291,
+ "step": 1407
+ },
+ {
+ "epoch": 19.55846422338569,
+ "grad_norm": 0.04726000875234604,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 4.136838436126709,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57242582897033,
+ "grad_norm": 0.046251531690359116,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 4.1526994705200195,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586387434554975,
+ "grad_norm": 0.047958049923181534,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 4.13817024230957,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600349040139616,
+ "grad_norm": 0.046333350241184235,
+ "learning_rate": 0.000417272427439646,
+ "loss": 4.160187721252441,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61431064572426,
+ "grad_norm": 0.0488678440451622,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 4.166190147399902,
+ "step": 1412
+ },
+ {
+ "epoch": 19.6282722513089,
+ "grad_norm": 0.04784323647618294,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 4.164414405822754,
+ "step": 1413
+ },
+ {
+ "epoch": 19.642233856893544,
+ "grad_norm": 0.05064673349261284,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 4.149814128875732,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656195462478184,
+ "grad_norm": 0.05072085186839104,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 4.167759895324707,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670157068062828,
+ "grad_norm": 0.0499395951628685,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 4.162792205810547,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68411867364747,
+ "grad_norm": 0.046766918152570724,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 4.1416521072387695,
+ "step": 1417
+ },
+ {
+ "epoch": 19.698080279232112,
+ "grad_norm": 0.04523273929953575,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 4.161751747131348,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712041884816752,
+ "grad_norm": 0.046106480062007904,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 4.166394233703613,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726003490401396,
+ "grad_norm": 0.04666793718934059,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 4.124961853027344,
+ "step": 1420
+ },
+ {
+ "epoch": 19.739965095986037,
+ "grad_norm": 0.047407012432813644,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 4.1794328689575195,
+ "step": 1421
+ },
+ {
+ "epoch": 19.75392670157068,
+ "grad_norm": 0.04690873995423317,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 4.160114765167236,
+ "step": 1422
+ },
+ {
+ "epoch": 19.767888307155324,
+ "grad_norm": 0.04589436575770378,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 4.137969970703125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.781849912739965,
+ "grad_norm": 0.04641961678862572,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 4.161533355712891,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79581151832461,
+ "grad_norm": 0.045799050480127335,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 4.187039375305176,
+ "step": 1425
+ },
+ {
+ "epoch": 19.80977312390925,
+ "grad_norm": 0.04881894588470459,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 4.158107757568359,
+ "step": 1426
+ },
+ {
+ "epoch": 19.823734729493893,
+ "grad_norm": 0.04506978392601013,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 4.143360137939453,
+ "step": 1427
+ },
+ {
+ "epoch": 19.837696335078533,
+ "grad_norm": 0.04504861310124397,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 4.174564361572266,
+ "step": 1428
+ },
+ {
+ "epoch": 19.851657940663177,
+ "grad_norm": 0.04639774188399315,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 4.162899971008301,
+ "step": 1429
+ },
+ {
+ "epoch": 19.865619546247817,
+ "grad_norm": 0.047444432973861694,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 4.143390655517578,
+ "step": 1430
+ },
+ {
+ "epoch": 19.87958115183246,
+ "grad_norm": 0.04461613669991493,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 4.171284198760986,
+ "step": 1431
+ },
+ {
+ "epoch": 19.8935427574171,
+ "grad_norm": 0.046383798122406006,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 4.14809513092041,
+ "step": 1432
+ },
+ {
+ "epoch": 19.907504363001745,
+ "grad_norm": 0.04642366245388985,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 4.178903579711914,
+ "step": 1433
+ },
+ {
+ "epoch": 19.921465968586386,
+ "grad_norm": 0.0451771542429924,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 4.171825408935547,
+ "step": 1434
+ },
+ {
+ "epoch": 19.93542757417103,
+ "grad_norm": 0.04767893627285957,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 4.169270038604736,
+ "step": 1435
+ },
+ {
+ "epoch": 19.949389179755673,
+ "grad_norm": 0.04673412814736366,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 4.175506591796875,
+ "step": 1436
+ },
+ {
+ "epoch": 19.963350785340314,
+ "grad_norm": 0.04465145245194435,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 4.149707794189453,
+ "step": 1437
+ },
+ {
+ "epoch": 19.977312390924958,
+ "grad_norm": 0.04734131321310997,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 4.166044235229492,
+ "step": 1438
+ },
+ {
+ "epoch": 19.991273996509598,
+ "grad_norm": 0.04706178605556488,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 4.152081489562988,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.031556785106658936,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 2.60109543800354,
+ "step": 1440
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 6.14458849272791e+17,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1440/training_args.bin b/runs/l2r90-wd0033-ssl/checkpoint-1440/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..313c41733ff55284c06034a59bb885b30e40ffc8
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1440/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7ddd0bdddba68d3e19ab33477000f3fc684fe4ee8e3580eeb8c02646e5fc27ac
+size 4856
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1800/chat_template.jinja b/runs/l2r90-wd0033-ssl/checkpoint-1800/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1800/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1800/config.json b/runs/l2r90-wd0033-ssl/checkpoint-1800/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1800/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1800/generation_config.json b/runs/l2r90-wd0033-ssl/checkpoint-1800/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1800/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1800/model.safetensors b/runs/l2r90-wd0033-ssl/checkpoint-1800/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..0629aff5ce2d0e986759c294366fd99b2fba65bd
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1800/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:db20d808fba6fab9a9ce56789f991dc15b9bdde609eb5f24576bb94d1ece66b5
+size 583356232
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1800/optimizer.pt b/runs/l2r90-wd0033-ssl/checkpoint-1800/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..4e223201bee80febae08a7092cb3137e43f37894
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1800/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c0f0da630f8f8caaf464a1f393185a36537f63cca381b90538544755f99609dc
+size 1166827898
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1800/rng_state_0.pth b/runs/l2r90-wd0033-ssl/checkpoint-1800/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..7d6b18c29f439e3010a8b54373d317696d66573f
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1800/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bedcb09a1c8950a65096d4085b70ba38f132583fa00a66de92fcc356e19a09d9
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1800/rng_state_1.pth b/runs/l2r90-wd0033-ssl/checkpoint-1800/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..3989a9d6a9b159dc1be67742f3434f6c50d1fb5f
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1800/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:68d1b7677dc8ab3e76010e23c7c2711560a2616f56d2a2907e1a8544b829f4e7
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1800/scheduler.pt b/runs/l2r90-wd0033-ssl/checkpoint-1800/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..6e0ba2afe167e5831737cd4b431b166033f17fab
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1800/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ee9cab552a4cabfd465ec6f079c65fbe7304d4a3df75e2e2b3e2eb614e447c70
+size 1064
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1800/tokenizer.json b/runs/l2r90-wd0033-ssl/checkpoint-1800/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1800/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1800/tokenizer_config.json b/runs/l2r90-wd0033-ssl/checkpoint-1800/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1800/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1800/trainer_state.json b/runs/l2r90-wd0033-ssl/checkpoint-1800/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..bb0df24be2ef7b4d288dd344c585b8210ab0dd99
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1800/trainer_state.json
@@ -0,0 +1,12818 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 25.0,
+ "eval_steps": 500,
+ "global_step": 1800,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013961605584642234,
+ "grad_norm": 0.26733073592185974,
+ "learning_rate": 0.0,
+ "loss": 7.990724086761475,
+ "step": 1
+ },
+ {
+ "epoch": 0.027923211169284468,
+ "grad_norm": 0.26712340116500854,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 7.990110874176025,
+ "step": 2
+ },
+ {
+ "epoch": 0.041884816753926704,
+ "grad_norm": 0.2646932601928711,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 7.956935405731201,
+ "step": 3
+ },
+ {
+ "epoch": 0.055846422338568937,
+ "grad_norm": 0.24684545397758484,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 7.890539646148682,
+ "step": 4
+ },
+ {
+ "epoch": 0.06980802792321117,
+ "grad_norm": 0.21016643941402435,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 7.8146138191223145,
+ "step": 5
+ },
+ {
+ "epoch": 0.08376963350785341,
+ "grad_norm": 0.17992742359638214,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 7.724307060241699,
+ "step": 6
+ },
+ {
+ "epoch": 0.09773123909249563,
+ "grad_norm": 0.14356893301010132,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 7.645662307739258,
+ "step": 7
+ },
+ {
+ "epoch": 0.11169284467713787,
+ "grad_norm": 0.10830964893102646,
+ "learning_rate": 4.2e-05,
+ "loss": 7.582019329071045,
+ "step": 8
+ },
+ {
+ "epoch": 0.1256544502617801,
+ "grad_norm": 0.08265161514282227,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 7.5230793952941895,
+ "step": 9
+ },
+ {
+ "epoch": 0.13961605584642234,
+ "grad_norm": 0.06423710286617279,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 7.473812103271484,
+ "step": 10
+ },
+ {
+ "epoch": 0.15357766143106458,
+ "grad_norm": 0.05132246017456055,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 7.440786838531494,
+ "step": 11
+ },
+ {
+ "epoch": 0.16753926701570682,
+ "grad_norm": 0.040572404861450195,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 7.41290283203125,
+ "step": 12
+ },
+ {
+ "epoch": 0.18150087260034903,
+ "grad_norm": 0.03138705715537071,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 7.39749002456665,
+ "step": 13
+ },
+ {
+ "epoch": 0.19546247818499127,
+ "grad_norm": 0.02553008496761322,
+ "learning_rate": 7.8e-05,
+ "loss": 7.387092590332031,
+ "step": 14
+ },
+ {
+ "epoch": 0.2094240837696335,
+ "grad_norm": 0.022178977727890015,
+ "learning_rate": 8.4e-05,
+ "loss": 7.378993511199951,
+ "step": 15
+ },
+ {
+ "epoch": 0.22338568935427575,
+ "grad_norm": 0.019915293902158737,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 7.37220573425293,
+ "step": 16
+ },
+ {
+ "epoch": 0.23734729493891799,
+ "grad_norm": 0.017390629276633263,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 7.366759300231934,
+ "step": 17
+ },
+ {
+ "epoch": 0.2513089005235602,
+ "grad_norm": 0.015472771599888802,
+ "learning_rate": 0.000102,
+ "loss": 7.3652663230896,
+ "step": 18
+ },
+ {
+ "epoch": 0.26527050610820246,
+ "grad_norm": 0.013599214144051075,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 7.359714984893799,
+ "step": 19
+ },
+ {
+ "epoch": 0.2792321116928447,
+ "grad_norm": 0.012442861683666706,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 7.353302001953125,
+ "step": 20
+ },
+ {
+ "epoch": 0.2931937172774869,
+ "grad_norm": 0.010471965186297894,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 7.357417583465576,
+ "step": 21
+ },
+ {
+ "epoch": 0.30715532286212915,
+ "grad_norm": 0.009647224098443985,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 7.351763725280762,
+ "step": 22
+ },
+ {
+ "epoch": 0.32111692844677137,
+ "grad_norm": 0.008228210732340813,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 7.345434188842773,
+ "step": 23
+ },
+ {
+ "epoch": 0.33507853403141363,
+ "grad_norm": 0.007454514969140291,
+ "learning_rate": 0.000138,
+ "loss": 7.342520236968994,
+ "step": 24
+ },
+ {
+ "epoch": 0.34904013961605584,
+ "grad_norm": 0.0069856527261435986,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 7.338870048522949,
+ "step": 25
+ },
+ {
+ "epoch": 0.36300174520069806,
+ "grad_norm": 0.006091665010899305,
+ "learning_rate": 0.00015,
+ "loss": 7.328121662139893,
+ "step": 26
+ },
+ {
+ "epoch": 0.3769633507853403,
+ "grad_norm": 0.005525320768356323,
+ "learning_rate": 0.000156,
+ "loss": 7.335029602050781,
+ "step": 27
+ },
+ {
+ "epoch": 0.39092495636998253,
+ "grad_norm": 0.0061133443377912045,
+ "learning_rate": 0.000162,
+ "loss": 7.319417953491211,
+ "step": 28
+ },
+ {
+ "epoch": 0.4048865619546248,
+ "grad_norm": 0.006040909793227911,
+ "learning_rate": 0.000168,
+ "loss": 7.326376438140869,
+ "step": 29
+ },
+ {
+ "epoch": 0.418848167539267,
+ "grad_norm": 0.006003216840326786,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 7.319552421569824,
+ "step": 30
+ },
+ {
+ "epoch": 0.4328097731239092,
+ "grad_norm": 0.007096003741025925,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 7.323486328125,
+ "step": 31
+ },
+ {
+ "epoch": 0.4467713787085515,
+ "grad_norm": 0.007725459989160299,
+ "learning_rate": 0.000186,
+ "loss": 7.3190717697143555,
+ "step": 32
+ },
+ {
+ "epoch": 0.4607329842931937,
+ "grad_norm": 0.007338670082390308,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 7.317154884338379,
+ "step": 33
+ },
+ {
+ "epoch": 0.47469458987783597,
+ "grad_norm": 0.007709544617682695,
+ "learning_rate": 0.000198,
+ "loss": 7.311826705932617,
+ "step": 34
+ },
+ {
+ "epoch": 0.4886561954624782,
+ "grad_norm": 0.009249784983694553,
+ "learning_rate": 0.000204,
+ "loss": 7.322498798370361,
+ "step": 35
+ },
+ {
+ "epoch": 0.5026178010471204,
+ "grad_norm": 0.006950767710804939,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 7.314571380615234,
+ "step": 36
+ },
+ {
+ "epoch": 0.5165794066317626,
+ "grad_norm": 0.007583985570818186,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 7.309540748596191,
+ "step": 37
+ },
+ {
+ "epoch": 0.5305410122164049,
+ "grad_norm": 0.014731820672750473,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 7.307004928588867,
+ "step": 38
+ },
+ {
+ "epoch": 0.5445026178010471,
+ "grad_norm": 0.016276473179459572,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 7.299642562866211,
+ "step": 39
+ },
+ {
+ "epoch": 0.5584642233856894,
+ "grad_norm": 0.020018689334392548,
+ "learning_rate": 0.000234,
+ "loss": 7.298466682434082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5724258289703316,
+ "grad_norm": 0.014941536821424961,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 7.289968967437744,
+ "step": 41
+ },
+ {
+ "epoch": 0.5863874345549738,
+ "grad_norm": 0.018299812451004982,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 7.276564598083496,
+ "step": 42
+ },
+ {
+ "epoch": 0.6003490401396161,
+ "grad_norm": 0.01615075021982193,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.2575554847717285,
+ "step": 43
+ },
+ {
+ "epoch": 0.6143106457242583,
+ "grad_norm": 0.013063831254839897,
+ "learning_rate": 0.000258,
+ "loss": 7.253885269165039,
+ "step": 44
+ },
+ {
+ "epoch": 0.6282722513089005,
+ "grad_norm": 0.015211676247417927,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.245779037475586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6422338568935427,
+ "grad_norm": 0.02093500830233097,
+ "learning_rate": 0.00027,
+ "loss": 7.225130081176758,
+ "step": 46
+ },
+ {
+ "epoch": 0.6561954624781849,
+ "grad_norm": 0.01808053068816662,
+ "learning_rate": 0.000276,
+ "loss": 7.209654331207275,
+ "step": 47
+ },
+ {
+ "epoch": 0.6701570680628273,
+ "grad_norm": 0.02209644205868244,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.197124481201172,
+ "step": 48
+ },
+ {
+ "epoch": 0.6841186736474695,
+ "grad_norm": 0.023524878546595573,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.1841936111450195,
+ "step": 49
+ },
+ {
+ "epoch": 0.6980802792321117,
+ "grad_norm": 0.031148087233304977,
+ "learning_rate": 0.000294,
+ "loss": 7.188040733337402,
+ "step": 50
+ },
+ {
+ "epoch": 0.7120418848167539,
+ "grad_norm": 0.02931929938495159,
+ "learning_rate": 0.0003,
+ "loss": 7.17371940612793,
+ "step": 51
+ },
+ {
+ "epoch": 0.7260034904013961,
+ "grad_norm": 0.01611483097076416,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.1714324951171875,
+ "step": 52
+ },
+ {
+ "epoch": 0.7399650959860384,
+ "grad_norm": 0.02619907818734646,
+ "learning_rate": 0.000312,
+ "loss": 7.158261299133301,
+ "step": 53
+ },
+ {
+ "epoch": 0.7539267015706806,
+ "grad_norm": 0.049462560564279556,
+ "learning_rate": 0.000318,
+ "loss": 7.167076110839844,
+ "step": 54
+ },
+ {
+ "epoch": 0.7678883071553229,
+ "grad_norm": 0.09835070371627808,
+ "learning_rate": 0.000324,
+ "loss": 7.154292106628418,
+ "step": 55
+ },
+ {
+ "epoch": 0.7818499127399651,
+ "grad_norm": 0.11406510323286057,
+ "learning_rate": 0.00033,
+ "loss": 7.162016868591309,
+ "step": 56
+ },
+ {
+ "epoch": 0.7958115183246073,
+ "grad_norm": 0.027026303112506866,
+ "learning_rate": 0.000336,
+ "loss": 7.141416549682617,
+ "step": 57
+ },
+ {
+ "epoch": 0.8097731239092496,
+ "grad_norm": 0.06690579652786255,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.13959264755249,
+ "step": 58
+ },
+ {
+ "epoch": 0.8237347294938918,
+ "grad_norm": 0.020919431000947952,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.126513481140137,
+ "step": 59
+ },
+ {
+ "epoch": 0.837696335078534,
+ "grad_norm": 0.05618130415678024,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.124345779418945,
+ "step": 60
+ },
+ {
+ "epoch": 0.8516579406631762,
+ "grad_norm": 0.03402302414178848,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.100615501403809,
+ "step": 61
+ },
+ {
+ "epoch": 0.8656195462478184,
+ "grad_norm": 0.03646906837821007,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.116507530212402,
+ "step": 62
+ },
+ {
+ "epoch": 0.8795811518324608,
+ "grad_norm": 0.03076397068798542,
+ "learning_rate": 0.000372,
+ "loss": 7.097659111022949,
+ "step": 63
+ },
+ {
+ "epoch": 0.893542757417103,
+ "grad_norm": 0.03329646959900856,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.097290992736816,
+ "step": 64
+ },
+ {
+ "epoch": 0.9075043630017452,
+ "grad_norm": 0.01998630352318287,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.092268943786621,
+ "step": 65
+ },
+ {
+ "epoch": 0.9214659685863874,
+ "grad_norm": 0.03278940171003342,
+ "learning_rate": 0.00039,
+ "loss": 7.088292121887207,
+ "step": 66
+ },
+ {
+ "epoch": 0.9354275741710296,
+ "grad_norm": 0.024620421230793,
+ "learning_rate": 0.000396,
+ "loss": 7.078650951385498,
+ "step": 67
+ },
+ {
+ "epoch": 0.9493891797556719,
+ "grad_norm": 0.03750383481383324,
+ "learning_rate": 0.000402,
+ "loss": 7.052778244018555,
+ "step": 68
+ },
+ {
+ "epoch": 0.9633507853403142,
+ "grad_norm": 0.07558075338602066,
+ "learning_rate": 0.000408,
+ "loss": 7.045991897583008,
+ "step": 69
+ },
+ {
+ "epoch": 0.9773123909249564,
+ "grad_norm": 0.16021482646465302,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.023152828216553,
+ "step": 70
+ },
+ {
+ "epoch": 0.9912739965095986,
+ "grad_norm": 0.24981893599033356,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.067540645599365,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.10255520790815353,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 4.425640106201172,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8861758708953857,
+ "eval_runtime": 58.2699,
+ "eval_samples_per_second": 41.908,
+ "eval_steps_per_second": 0.669,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139616055846423,
+ "grad_norm": 0.13374797999858856,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.073373317718506,
+ "step": 73
+ },
+ {
+ "epoch": 1.0279232111692844,
+ "grad_norm": 0.05079387128353119,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.0243306159973145,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418848167539267,
+ "grad_norm": 0.10056068748235703,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.014476299285889,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558464223385688,
+ "grad_norm": 0.06995758414268494,
+ "learning_rate": 0.00045,
+ "loss": 7.011931419372559,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698080279232112,
+ "grad_norm": 0.0483892522752285,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 6.9822797775268555,
+ "step": 77
+ },
+ {
+ "epoch": 1.0837696335078535,
+ "grad_norm": 0.08408170193433762,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 6.964348793029785,
+ "step": 78
+ },
+ {
+ "epoch": 1.0977312390924956,
+ "grad_norm": 0.08027710020542145,
+ "learning_rate": 0.000468,
+ "loss": 6.96852445602417,
+ "step": 79
+ },
+ {
+ "epoch": 1.111692844677138,
+ "grad_norm": 0.05647159367799759,
+ "learning_rate": 0.000474,
+ "loss": 6.944484710693359,
+ "step": 80
+ },
+ {
+ "epoch": 1.12565445026178,
+ "grad_norm": 0.07464330643415451,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 6.9246745109558105,
+ "step": 81
+ },
+ {
+ "epoch": 1.1396160558464223,
+ "grad_norm": 0.04451674595475197,
+ "learning_rate": 0.000486,
+ "loss": 6.895403861999512,
+ "step": 82
+ },
+ {
+ "epoch": 1.1535776614310647,
+ "grad_norm": 0.05820256471633911,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 6.887338638305664,
+ "step": 83
+ },
+ {
+ "epoch": 1.1675392670157068,
+ "grad_norm": 0.02925296127796173,
+ "learning_rate": 0.000498,
+ "loss": 6.857797622680664,
+ "step": 84
+ },
+ {
+ "epoch": 1.181500872600349,
+ "grad_norm": 0.05096987634897232,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 6.854822158813477,
+ "step": 85
+ },
+ {
+ "epoch": 1.1954624781849912,
+ "grad_norm": 0.03483197093009949,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 6.840914726257324,
+ "step": 86
+ },
+ {
+ "epoch": 1.2094240837696335,
+ "grad_norm": 0.03323264792561531,
+ "learning_rate": 0.000516,
+ "loss": 6.834497451782227,
+ "step": 87
+ },
+ {
+ "epoch": 1.2233856893542758,
+ "grad_norm": 0.03290868178009987,
+ "learning_rate": 0.000522,
+ "loss": 6.820746898651123,
+ "step": 88
+ },
+ {
+ "epoch": 1.237347294938918,
+ "grad_norm": 0.025398844853043556,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 6.795746803283691,
+ "step": 89
+ },
+ {
+ "epoch": 1.2513089005235603,
+ "grad_norm": 0.030700596049427986,
+ "learning_rate": 0.000534,
+ "loss": 6.7867889404296875,
+ "step": 90
+ },
+ {
+ "epoch": 1.2652705061082026,
+ "grad_norm": 0.019589420408010483,
+ "learning_rate": 0.00054,
+ "loss": 6.77443790435791,
+ "step": 91
+ },
+ {
+ "epoch": 1.2792321116928447,
+ "grad_norm": 0.03313920646905899,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 6.762271881103516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2931937172774868,
+ "grad_norm": 0.025389468297362328,
+ "learning_rate": 0.000552,
+ "loss": 6.750258445739746,
+ "step": 93
+ },
+ {
+ "epoch": 1.307155322862129,
+ "grad_norm": 0.040296655148267746,
+ "learning_rate": 0.000558,
+ "loss": 6.729808807373047,
+ "step": 94
+ },
+ {
+ "epoch": 1.3211169284467714,
+ "grad_norm": 0.04735933244228363,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 6.737982749938965,
+ "step": 95
+ },
+ {
+ "epoch": 1.3350785340314135,
+ "grad_norm": 0.05529680848121643,
+ "learning_rate": 0.00057,
+ "loss": 6.735714912414551,
+ "step": 96
+ },
+ {
+ "epoch": 1.3490401396160558,
+ "grad_norm": 0.07148554176092148,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 6.708859920501709,
+ "step": 97
+ },
+ {
+ "epoch": 1.3630017452006982,
+ "grad_norm": 0.059708479791879654,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 6.708860397338867,
+ "step": 98
+ },
+ {
+ "epoch": 1.3769633507853403,
+ "grad_norm": 0.027111195027828217,
+ "learning_rate": 0.000588,
+ "loss": 6.686685085296631,
+ "step": 99
+ },
+ {
+ "epoch": 1.3909249563699826,
+ "grad_norm": 0.0500967763364315,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 6.676325798034668,
+ "step": 100
+ },
+ {
+ "epoch": 1.404886561954625,
+ "grad_norm": 0.044565796852111816,
+ "learning_rate": 0.0006,
+ "loss": 6.674530029296875,
+ "step": 101
+ },
+ {
+ "epoch": 1.418848167539267,
+ "grad_norm": 0.03608180209994316,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 6.653741359710693,
+ "step": 102
+ },
+ {
+ "epoch": 1.4328097731239091,
+ "grad_norm": 0.030394982546567917,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 6.643152236938477,
+ "step": 103
+ },
+ {
+ "epoch": 1.4467713787085514,
+ "grad_norm": 0.040185656398534775,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 6.628686904907227,
+ "step": 104
+ },
+ {
+ "epoch": 1.4607329842931938,
+ "grad_norm": 0.021049322560429573,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 6.610864639282227,
+ "step": 105
+ },
+ {
+ "epoch": 1.4746945898778359,
+ "grad_norm": 0.04164277762174606,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 6.602880954742432,
+ "step": 106
+ },
+ {
+ "epoch": 1.4886561954624782,
+ "grad_norm": 0.0389864556491375,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 6.599456787109375,
+ "step": 107
+ },
+ {
+ "epoch": 1.5026178010471205,
+ "grad_norm": 0.05805973336100578,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 6.588700294494629,
+ "step": 108
+ },
+ {
+ "epoch": 1.5165794066317626,
+ "grad_norm": 0.08825036883354187,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 6.587496757507324,
+ "step": 109
+ },
+ {
+ "epoch": 1.530541012216405,
+ "grad_norm": 0.12521523237228394,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 6.626501083374023,
+ "step": 110
+ },
+ {
+ "epoch": 1.5445026178010473,
+ "grad_norm": 0.07429645210504532,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 6.590241432189941,
+ "step": 111
+ },
+ {
+ "epoch": 1.5584642233856894,
+ "grad_norm": 0.06463009119033813,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 6.578714370727539,
+ "step": 112
+ },
+ {
+ "epoch": 1.5724258289703315,
+ "grad_norm": 0.07508203387260437,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 6.561270713806152,
+ "step": 113
+ },
+ {
+ "epoch": 1.5863874345549738,
+ "grad_norm": 0.04125197231769562,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 6.558821678161621,
+ "step": 114
+ },
+ {
+ "epoch": 1.600349040139616,
+ "grad_norm": 0.05457386001944542,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.541611671447754,
+ "step": 115
+ },
+ {
+ "epoch": 1.6143106457242582,
+ "grad_norm": 0.03684385493397713,
+ "learning_rate": 0.000599973080557108,
+ "loss": 6.547738075256348,
+ "step": 116
+ },
+ {
+ "epoch": 1.6282722513089005,
+ "grad_norm": 0.05888746306300163,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.529491424560547,
+ "step": 117
+ },
+ {
+ "epoch": 1.6422338568935428,
+ "grad_norm": 0.056668370962142944,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.519444942474365,
+ "step": 118
+ },
+ {
+ "epoch": 1.656195462478185,
+ "grad_norm": 0.09123896062374115,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.522984981536865,
+ "step": 119
+ },
+ {
+ "epoch": 1.6701570680628273,
+ "grad_norm": 0.12380023300647736,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.515375137329102,
+ "step": 120
+ },
+ {
+ "epoch": 1.6841186736474696,
+ "grad_norm": 0.06953322887420654,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.500491142272949,
+ "step": 121
+ },
+ {
+ "epoch": 1.6980802792321117,
+ "grad_norm": 0.08613869547843933,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.5097198486328125,
+ "step": 122
+ },
+ {
+ "epoch": 1.7120418848167538,
+ "grad_norm": 0.08296383172273636,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.487521171569824,
+ "step": 123
+ },
+ {
+ "epoch": 1.7260034904013961,
+ "grad_norm": 0.05744783207774162,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.485134124755859,
+ "step": 124
+ },
+ {
+ "epoch": 1.7399650959860384,
+ "grad_norm": 0.04108566418290138,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.466182708740234,
+ "step": 125
+ },
+ {
+ "epoch": 1.7539267015706805,
+ "grad_norm": 0.05374931916594505,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.479011535644531,
+ "step": 126
+ },
+ {
+ "epoch": 1.7678883071553229,
+ "grad_norm": 0.03867914527654648,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.465902328491211,
+ "step": 127
+ },
+ {
+ "epoch": 1.7818499127399652,
+ "grad_norm": 0.05477374792098999,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.453082084655762,
+ "step": 128
+ },
+ {
+ "epoch": 1.7958115183246073,
+ "grad_norm": 0.06508298963308334,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.446089744567871,
+ "step": 129
+ },
+ {
+ "epoch": 1.8097731239092496,
+ "grad_norm": 0.06388624757528305,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.451748847961426,
+ "step": 130
+ },
+ {
+ "epoch": 1.823734729493892,
+ "grad_norm": 0.04223188757896423,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.4492387771606445,
+ "step": 131
+ },
+ {
+ "epoch": 1.837696335078534,
+ "grad_norm": 0.04528658092021942,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.409390926361084,
+ "step": 132
+ },
+ {
+ "epoch": 1.8516579406631761,
+ "grad_norm": 0.04962967336177826,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.415774345397949,
+ "step": 133
+ },
+ {
+ "epoch": 1.8656195462478184,
+ "grad_norm": 0.031944390386343,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.411528587341309,
+ "step": 134
+ },
+ {
+ "epoch": 1.8795811518324608,
+ "grad_norm": 0.028411149978637695,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.392329216003418,
+ "step": 135
+ },
+ {
+ "epoch": 1.8935427574171029,
+ "grad_norm": 0.03516148403286934,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.392238616943359,
+ "step": 136
+ },
+ {
+ "epoch": 1.9075043630017452,
+ "grad_norm": 0.027800120413303375,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.39179801940918,
+ "step": 137
+ },
+ {
+ "epoch": 1.9214659685863875,
+ "grad_norm": 0.02362987957894802,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.378960609436035,
+ "step": 138
+ },
+ {
+ "epoch": 1.9354275741710296,
+ "grad_norm": 0.033999308943748474,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.36802339553833,
+ "step": 139
+ },
+ {
+ "epoch": 1.949389179755672,
+ "grad_norm": 0.03290879726409912,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.366308689117432,
+ "step": 140
+ },
+ {
+ "epoch": 1.9633507853403143,
+ "grad_norm": 0.0286843404173851,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.350399971008301,
+ "step": 141
+ },
+ {
+ "epoch": 1.9773123909249564,
+ "grad_norm": 0.027145877480506897,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.33477783203125,
+ "step": 142
+ },
+ {
+ "epoch": 1.9912739965095985,
+ "grad_norm": 0.03530878946185112,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.322859764099121,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.028133150190114975,
+ "learning_rate": 0.000599778806120077,
+ "loss": 3.955288887023926,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7927101254463196,
+ "eval_runtime": 60.8372,
+ "eval_samples_per_second": 40.14,
+ "eval_steps_per_second": 0.641,
+ "step": 144
+ },
+ {
+ "epoch": 2.013961605584642,
+ "grad_norm": 0.03054557554423809,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.311975479125977,
+ "step": 145
+ },
+ {
+ "epoch": 2.0279232111692846,
+ "grad_norm": 0.05406459793448448,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.3030853271484375,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418848167539267,
+ "grad_norm": 0.10911019891500473,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.329056739807129,
+ "step": 147
+ },
+ {
+ "epoch": 2.055846422338569,
+ "grad_norm": 0.13948315382003784,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.362421035766602,
+ "step": 148
+ },
+ {
+ "epoch": 2.0698080279232114,
+ "grad_norm": 0.13176991045475006,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.3888139724731445,
+ "step": 149
+ },
+ {
+ "epoch": 2.0837696335078535,
+ "grad_norm": 0.09615885466337204,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.35154914855957,
+ "step": 150
+ },
+ {
+ "epoch": 2.0977312390924956,
+ "grad_norm": 0.06462561339139938,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.335790634155273,
+ "step": 151
+ },
+ {
+ "epoch": 2.1116928446771377,
+ "grad_norm": 0.05186621472239494,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.324396133422852,
+ "step": 152
+ },
+ {
+ "epoch": 2.1256544502617802,
+ "grad_norm": 0.06095981225371361,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.318549156188965,
+ "step": 153
+ },
+ {
+ "epoch": 2.1396160558464223,
+ "grad_norm": 0.036121442914009094,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.300607681274414,
+ "step": 154
+ },
+ {
+ "epoch": 2.1535776614310644,
+ "grad_norm": 0.055716633796691895,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.307400703430176,
+ "step": 155
+ },
+ {
+ "epoch": 2.167539267015707,
+ "grad_norm": 0.037626396864652634,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.287364959716797,
+ "step": 156
+ },
+ {
+ "epoch": 2.181500872600349,
+ "grad_norm": 0.04329098388552666,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.28244686126709,
+ "step": 157
+ },
+ {
+ "epoch": 2.195462478184991,
+ "grad_norm": 0.031868450343608856,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.2602410316467285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2094240837696333,
+ "grad_norm": 0.03647344186902046,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.263988971710205,
+ "step": 159
+ },
+ {
+ "epoch": 2.223385689354276,
+ "grad_norm": 0.03141765296459198,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.263205051422119,
+ "step": 160
+ },
+ {
+ "epoch": 2.237347294938918,
+ "grad_norm": 0.028253229334950447,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.253107070922852,
+ "step": 161
+ },
+ {
+ "epoch": 2.25130890052356,
+ "grad_norm": 0.03608971834182739,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.2571702003479,
+ "step": 162
+ },
+ {
+ "epoch": 2.2652705061082026,
+ "grad_norm": 0.02644292451441288,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.241732597351074,
+ "step": 163
+ },
+ {
+ "epoch": 2.2792321116928447,
+ "grad_norm": 0.028604455292224884,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.230289459228516,
+ "step": 164
+ },
+ {
+ "epoch": 2.2931937172774868,
+ "grad_norm": 0.023704614490270615,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.226101398468018,
+ "step": 165
+ },
+ {
+ "epoch": 2.3071553228621293,
+ "grad_norm": 0.020691027864813805,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.2096405029296875,
+ "step": 166
+ },
+ {
+ "epoch": 2.3211169284467714,
+ "grad_norm": 0.025598427280783653,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.22075080871582,
+ "step": 167
+ },
+ {
+ "epoch": 2.3350785340314135,
+ "grad_norm": 0.030546002089977264,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.216886520385742,
+ "step": 168
+ },
+ {
+ "epoch": 2.349040139616056,
+ "grad_norm": 0.026066245511174202,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.2022504806518555,
+ "step": 169
+ },
+ {
+ "epoch": 2.363001745200698,
+ "grad_norm": 0.023802239447832108,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.189509391784668,
+ "step": 170
+ },
+ {
+ "epoch": 2.3769633507853403,
+ "grad_norm": 0.029805002734065056,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.186277389526367,
+ "step": 171
+ },
+ {
+ "epoch": 2.3909249563699824,
+ "grad_norm": 0.043327830731868744,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.164876937866211,
+ "step": 172
+ },
+ {
+ "epoch": 2.404886561954625,
+ "grad_norm": 0.058787211775779724,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.179936408996582,
+ "step": 173
+ },
+ {
+ "epoch": 2.418848167539267,
+ "grad_norm": 0.0787397176027298,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.1830244064331055,
+ "step": 174
+ },
+ {
+ "epoch": 2.432809773123909,
+ "grad_norm": 0.0899854376912117,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.190343856811523,
+ "step": 175
+ },
+ {
+ "epoch": 2.4467713787085517,
+ "grad_norm": 0.06520707905292511,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.185057163238525,
+ "step": 176
+ },
+ {
+ "epoch": 2.4607329842931938,
+ "grad_norm": 0.1002504974603653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.171279430389404,
+ "step": 177
+ },
+ {
+ "epoch": 2.474694589877836,
+ "grad_norm": 0.09591004997491837,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.174382209777832,
+ "step": 178
+ },
+ {
+ "epoch": 2.488656195462478,
+ "grad_norm": 0.12127121537923813,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.185717582702637,
+ "step": 179
+ },
+ {
+ "epoch": 2.5026178010471205,
+ "grad_norm": 0.09640590846538544,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.196681976318359,
+ "step": 180
+ },
+ {
+ "epoch": 2.5165794066317626,
+ "grad_norm": 0.048341453075408936,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.158774375915527,
+ "step": 181
+ },
+ {
+ "epoch": 2.530541012216405,
+ "grad_norm": 0.08206334710121155,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.17445707321167,
+ "step": 182
+ },
+ {
+ "epoch": 2.5445026178010473,
+ "grad_norm": 0.038934968411922455,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.170527458190918,
+ "step": 183
+ },
+ {
+ "epoch": 2.5584642233856894,
+ "grad_norm": 0.060987457633018494,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.1501994132995605,
+ "step": 184
+ },
+ {
+ "epoch": 2.5724258289703315,
+ "grad_norm": 0.033269546926021576,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.144186019897461,
+ "step": 185
+ },
+ {
+ "epoch": 2.5863874345549736,
+ "grad_norm": 0.046137236058712006,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.1236677169799805,
+ "step": 186
+ },
+ {
+ "epoch": 2.600349040139616,
+ "grad_norm": 0.038265109062194824,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.121979236602783,
+ "step": 187
+ },
+ {
+ "epoch": 2.614310645724258,
+ "grad_norm": 0.039128001779317856,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.124156951904297,
+ "step": 188
+ },
+ {
+ "epoch": 2.6282722513089007,
+ "grad_norm": 0.025303540751338005,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.108036994934082,
+ "step": 189
+ },
+ {
+ "epoch": 2.642233856893543,
+ "grad_norm": 0.04441308230161667,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.1211161613464355,
+ "step": 190
+ },
+ {
+ "epoch": 2.656195462478185,
+ "grad_norm": 0.03083532303571701,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.094561576843262,
+ "step": 191
+ },
+ {
+ "epoch": 2.670157068062827,
+ "grad_norm": 0.02951296605169773,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.091608047485352,
+ "step": 192
+ },
+ {
+ "epoch": 2.6841186736474696,
+ "grad_norm": 0.02385987527668476,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.087080001831055,
+ "step": 193
+ },
+ {
+ "epoch": 2.6980802792321117,
+ "grad_norm": 0.03315780311822891,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.088132381439209,
+ "step": 194
+ },
+ {
+ "epoch": 2.712041884816754,
+ "grad_norm": 0.02410835586488247,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.081460952758789,
+ "step": 195
+ },
+ {
+ "epoch": 2.7260034904013963,
+ "grad_norm": 0.033069487661123276,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.0696539878845215,
+ "step": 196
+ },
+ {
+ "epoch": 2.7399650959860384,
+ "grad_norm": 0.031288471072912216,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.047612190246582,
+ "step": 197
+ },
+ {
+ "epoch": 2.7539267015706805,
+ "grad_norm": 0.03399677947163582,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.019026756286621,
+ "step": 198
+ },
+ {
+ "epoch": 2.7678883071553226,
+ "grad_norm": 0.031948331743478775,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.050133228302002,
+ "step": 199
+ },
+ {
+ "epoch": 2.781849912739965,
+ "grad_norm": 0.049206435680389404,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.045332431793213,
+ "step": 200
+ },
+ {
+ "epoch": 2.7958115183246073,
+ "grad_norm": 0.07528849691152573,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.059202194213867,
+ "step": 201
+ },
+ {
+ "epoch": 2.80977312390925,
+ "grad_norm": 0.11815470457077026,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.077486038208008,
+ "step": 202
+ },
+ {
+ "epoch": 2.823734729493892,
+ "grad_norm": 0.0867348462343216,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.045670032501221,
+ "step": 203
+ },
+ {
+ "epoch": 2.837696335078534,
+ "grad_norm": 0.05716705322265625,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.034373760223389,
+ "step": 204
+ },
+ {
+ "epoch": 2.851657940663176,
+ "grad_norm": 0.05944284051656723,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.049911975860596,
+ "step": 205
+ },
+ {
+ "epoch": 2.8656195462478182,
+ "grad_norm": 0.03220470994710922,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.039861679077148,
+ "step": 206
+ },
+ {
+ "epoch": 2.8795811518324608,
+ "grad_norm": 0.057081177830696106,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.010714530944824,
+ "step": 207
+ },
+ {
+ "epoch": 2.893542757417103,
+ "grad_norm": 0.04037516191601753,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.018877983093262,
+ "step": 208
+ },
+ {
+ "epoch": 2.9075043630017454,
+ "grad_norm": 0.036492880433797836,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.009493827819824,
+ "step": 209
+ },
+ {
+ "epoch": 2.9214659685863875,
+ "grad_norm": 0.04785023257136345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.0140862464904785,
+ "step": 210
+ },
+ {
+ "epoch": 2.9354275741710296,
+ "grad_norm": 0.07721036672592163,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.030359268188477,
+ "step": 211
+ },
+ {
+ "epoch": 2.9493891797556717,
+ "grad_norm": 0.14220190048217773,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.045989990234375,
+ "step": 212
+ },
+ {
+ "epoch": 2.9633507853403143,
+ "grad_norm": 0.10265158116817474,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.043968200683594,
+ "step": 213
+ },
+ {
+ "epoch": 2.9773123909249564,
+ "grad_norm": 0.051398951560258865,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.022278785705566,
+ "step": 214
+ },
+ {
+ "epoch": 2.9912739965095985,
+ "grad_norm": 0.06044618785381317,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.017296314239502,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.03284750506281853,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 3.7544798851013184,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7531244158744812,
+ "eval_runtime": 60.3986,
+ "eval_samples_per_second": 40.431,
+ "eval_steps_per_second": 0.646,
+ "step": 216
+ },
+ {
+ "epoch": 3.013961605584642,
+ "grad_norm": 0.03895840048789978,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 5.9985857009887695,
+ "step": 217
+ },
+ {
+ "epoch": 3.0279232111692846,
+ "grad_norm": 0.04554992541670799,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 5.984710693359375,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418848167539267,
+ "grad_norm": 0.03359116613864899,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 5.984078407287598,
+ "step": 219
+ },
+ {
+ "epoch": 3.055846422338569,
+ "grad_norm": 0.035171881318092346,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 5.979424476623535,
+ "step": 220
+ },
+ {
+ "epoch": 3.0698080279232114,
+ "grad_norm": 0.032774072140455246,
+ "learning_rate": 0.000598278794687186,
+ "loss": 5.94467830657959,
+ "step": 221
+ },
+ {
+ "epoch": 3.0837696335078535,
+ "grad_norm": 0.032842908054590225,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 5.95986270904541,
+ "step": 222
+ },
+ {
+ "epoch": 3.0977312390924956,
+ "grad_norm": 0.02870488539338112,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 5.951186656951904,
+ "step": 223
+ },
+ {
+ "epoch": 3.1116928446771377,
+ "grad_norm": 0.030777374282479286,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 5.942493438720703,
+ "step": 224
+ },
+ {
+ "epoch": 3.1256544502617802,
+ "grad_norm": 0.026439018547534943,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 5.938275337219238,
+ "step": 225
+ },
+ {
+ "epoch": 3.1396160558464223,
+ "grad_norm": 0.028689472004771233,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 5.931982040405273,
+ "step": 226
+ },
+ {
+ "epoch": 3.1535776614310644,
+ "grad_norm": 0.028100362047553062,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 5.919692039489746,
+ "step": 227
+ },
+ {
+ "epoch": 3.167539267015707,
+ "grad_norm": 0.02407177910208702,
+ "learning_rate": 0.00059807235435179,
+ "loss": 5.909029006958008,
+ "step": 228
+ },
+ {
+ "epoch": 3.181500872600349,
+ "grad_norm": 0.0237748846411705,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 5.9131598472595215,
+ "step": 229
+ },
+ {
+ "epoch": 3.195462478184991,
+ "grad_norm": 0.022418556734919548,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 5.909396171569824,
+ "step": 230
+ },
+ {
+ "epoch": 3.2094240837696333,
+ "grad_norm": 0.024656571447849274,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 5.885243892669678,
+ "step": 231
+ },
+ {
+ "epoch": 3.223385689354276,
+ "grad_norm": 0.02136118710041046,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 5.898958206176758,
+ "step": 232
+ },
+ {
+ "epoch": 3.237347294938918,
+ "grad_norm": 0.025290116667747498,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 5.886986255645752,
+ "step": 233
+ },
+ {
+ "epoch": 3.25130890052356,
+ "grad_norm": 0.02656404674053192,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 5.897647857666016,
+ "step": 234
+ },
+ {
+ "epoch": 3.2652705061082026,
+ "grad_norm": 0.019188793376088142,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 5.875659942626953,
+ "step": 235
+ },
+ {
+ "epoch": 3.2792321116928447,
+ "grad_norm": 0.02308926172554493,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 5.859380722045898,
+ "step": 236
+ },
+ {
+ "epoch": 3.2931937172774868,
+ "grad_norm": 0.029775725677609444,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 5.873961448669434,
+ "step": 237
+ },
+ {
+ "epoch": 3.3071553228621293,
+ "grad_norm": 0.03875259310007095,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 5.863171100616455,
+ "step": 238
+ },
+ {
+ "epoch": 3.3211169284467714,
+ "grad_norm": 0.052446842193603516,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 5.886547088623047,
+ "step": 239
+ },
+ {
+ "epoch": 3.3350785340314135,
+ "grad_norm": 0.05583837628364563,
+ "learning_rate": 0.000597691359743243,
+ "loss": 5.8849287033081055,
+ "step": 240
+ },
+ {
+ "epoch": 3.349040139616056,
+ "grad_norm": 0.034729067236185074,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 5.863419532775879,
+ "step": 241
+ },
+ {
+ "epoch": 3.363001745200698,
+ "grad_norm": 0.03590013459324837,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 5.852169990539551,
+ "step": 242
+ },
+ {
+ "epoch": 3.3769633507853403,
+ "grad_norm": 0.035041507333517075,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 5.836690902709961,
+ "step": 243
+ },
+ {
+ "epoch": 3.3909249563699824,
+ "grad_norm": 0.02857939340174198,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 5.836485862731934,
+ "step": 244
+ },
+ {
+ "epoch": 3.404886561954625,
+ "grad_norm": 0.03176642581820488,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 5.860719203948975,
+ "step": 245
+ },
+ {
+ "epoch": 3.418848167539267,
+ "grad_norm": 0.026105748489499092,
+ "learning_rate": 0.000597488038559129,
+ "loss": 5.8354949951171875,
+ "step": 246
+ },
+ {
+ "epoch": 3.432809773123909,
+ "grad_norm": 0.029094981029629707,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 5.846247673034668,
+ "step": 247
+ },
+ {
+ "epoch": 3.4467713787085517,
+ "grad_norm": 0.02388785220682621,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 5.837607383728027,
+ "step": 248
+ },
+ {
+ "epoch": 3.4607329842931938,
+ "grad_norm": 0.03094472736120224,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 5.832090377807617,
+ "step": 249
+ },
+ {
+ "epoch": 3.474694589877836,
+ "grad_norm": 0.03227658197283745,
+ "learning_rate": 0.000597347745466109,
+ "loss": 5.83722448348999,
+ "step": 250
+ },
+ {
+ "epoch": 3.488656195462478,
+ "grad_norm": 0.042813923209905624,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 5.851722717285156,
+ "step": 251
+ },
+ {
+ "epoch": 3.5026178010471205,
+ "grad_norm": 0.05825525149703026,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.795719146728516,
+ "step": 252
+ },
+ {
+ "epoch": 3.5165794066317626,
+ "grad_norm": 0.0867447778582573,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.813750267028809,
+ "step": 253
+ },
+ {
+ "epoch": 3.530541012216405,
+ "grad_norm": 0.09192710369825363,
+ "learning_rate": 0.000597203657966319,
+ "loss": 5.823488235473633,
+ "step": 254
+ },
+ {
+ "epoch": 3.5445026178010473,
+ "grad_norm": 0.06069248169660568,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 5.814342021942139,
+ "step": 255
+ },
+ {
+ "epoch": 3.5584642233856894,
+ "grad_norm": 0.06259952485561371,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 5.834272384643555,
+ "step": 256
+ },
+ {
+ "epoch": 3.5724258289703315,
+ "grad_norm": 0.0513639897108078,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.807951927185059,
+ "step": 257
+ },
+ {
+ "epoch": 3.5863874345549736,
+ "grad_norm": 0.03327798470854759,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.790238857269287,
+ "step": 258
+ },
+ {
+ "epoch": 3.600349040139616,
+ "grad_norm": 0.03963128477334976,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.7956695556640625,
+ "step": 259
+ },
+ {
+ "epoch": 3.614310645724258,
+ "grad_norm": 0.031153453513979912,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 5.7874531745910645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6282722513089007,
+ "grad_norm": 0.031419433653354645,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 5.779650688171387,
+ "step": 261
+ },
+ {
+ "epoch": 3.642233856893543,
+ "grad_norm": 0.0288684144616127,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.776577949523926,
+ "step": 262
+ },
+ {
+ "epoch": 3.656195462478185,
+ "grad_norm": 0.028315918520092964,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.778571128845215,
+ "step": 263
+ },
+ {
+ "epoch": 3.670157068062827,
+ "grad_norm": 0.024977795779705048,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 5.776470184326172,
+ "step": 264
+ },
+ {
+ "epoch": 3.6841186736474696,
+ "grad_norm": 0.026822632178664207,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.755108833312988,
+ "step": 265
+ },
+ {
+ "epoch": 3.6980802792321117,
+ "grad_norm": 0.02738630585372448,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.74041748046875,
+ "step": 266
+ },
+ {
+ "epoch": 3.712041884816754,
+ "grad_norm": 0.02361280657351017,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.764027118682861,
+ "step": 267
+ },
+ {
+ "epoch": 3.7260034904013963,
+ "grad_norm": 0.029926009476184845,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.780390739440918,
+ "step": 268
+ },
+ {
+ "epoch": 3.7399650959860384,
+ "grad_norm": 0.03519507870078087,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.758105278015137,
+ "step": 269
+ },
+ {
+ "epoch": 3.7539267015706805,
+ "grad_norm": 0.030114633962512016,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.757230758666992,
+ "step": 270
+ },
+ {
+ "epoch": 3.7678883071553226,
+ "grad_norm": 0.02467748522758484,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.7615461349487305,
+ "step": 271
+ },
+ {
+ "epoch": 3.781849912739965,
+ "grad_norm": 0.03425183147192001,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.749448776245117,
+ "step": 272
+ },
+ {
+ "epoch": 3.7958115183246073,
+ "grad_norm": 0.046307582408189774,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.7356977462768555,
+ "step": 273
+ },
+ {
+ "epoch": 3.80977312390925,
+ "grad_norm": 0.054102856665849686,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.741869926452637,
+ "step": 274
+ },
+ {
+ "epoch": 3.823734729493892,
+ "grad_norm": 0.07026689499616623,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.721103668212891,
+ "step": 275
+ },
+ {
+ "epoch": 3.837696335078534,
+ "grad_norm": 0.0733039602637291,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.731414794921875,
+ "step": 276
+ },
+ {
+ "epoch": 3.851657940663176,
+ "grad_norm": 0.05342920497059822,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.734922885894775,
+ "step": 277
+ },
+ {
+ "epoch": 3.8656195462478182,
+ "grad_norm": 0.0504680834710598,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.732746124267578,
+ "step": 278
+ },
+ {
+ "epoch": 3.8795811518324608,
+ "grad_norm": 0.038991592824459076,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.719473838806152,
+ "step": 279
+ },
+ {
+ "epoch": 3.893542757417103,
+ "grad_norm": 0.04434404894709587,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.709481239318848,
+ "step": 280
+ },
+ {
+ "epoch": 3.9075043630017454,
+ "grad_norm": 0.047285184264183044,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.693531036376953,
+ "step": 281
+ },
+ {
+ "epoch": 3.9214659685863875,
+ "grad_norm": 0.045081138610839844,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.728959083557129,
+ "step": 282
+ },
+ {
+ "epoch": 3.9354275741710296,
+ "grad_norm": 0.025133663788437843,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.719696998596191,
+ "step": 283
+ },
+ {
+ "epoch": 3.9493891797556717,
+ "grad_norm": 0.03353596478700638,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.700235366821289,
+ "step": 284
+ },
+ {
+ "epoch": 3.9633507853403143,
+ "grad_norm": 0.035162411630153656,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.684836387634277,
+ "step": 285
+ },
+ {
+ "epoch": 3.9773123909249564,
+ "grad_norm": 0.04025450348854065,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.692097187042236,
+ "step": 286
+ },
+ {
+ "epoch": 3.9912739965095985,
+ "grad_norm": 0.04995397478342056,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.6766862869262695,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.031619708985090256,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 3.5519747734069824,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.714687705039978,
+ "eval_runtime": 59.637,
+ "eval_samples_per_second": 40.948,
+ "eval_steps_per_second": 0.654,
+ "step": 288
+ },
+ {
+ "epoch": 4.013961605584642,
+ "grad_norm": 0.036328598856925964,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.666294097900391,
+ "step": 289
+ },
+ {
+ "epoch": 4.027923211169284,
+ "grad_norm": 0.07326336950063705,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.6879167556762695,
+ "step": 290
+ },
+ {
+ "epoch": 4.041884816753926,
+ "grad_norm": 0.07947526127099991,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.674883842468262,
+ "step": 291
+ },
+ {
+ "epoch": 4.055846422338569,
+ "grad_norm": 0.04599885269999504,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.688086986541748,
+ "step": 292
+ },
+ {
+ "epoch": 4.069808027923211,
+ "grad_norm": 0.0465150885283947,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.649483680725098,
+ "step": 293
+ },
+ {
+ "epoch": 4.0837696335078535,
+ "grad_norm": 0.038542453199625015,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.653055667877197,
+ "step": 294
+ },
+ {
+ "epoch": 4.097731239092496,
+ "grad_norm": 0.03659164905548096,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.65955924987793,
+ "step": 295
+ },
+ {
+ "epoch": 4.111692844677138,
+ "grad_norm": 0.03824038803577423,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.650853157043457,
+ "step": 296
+ },
+ {
+ "epoch": 4.12565445026178,
+ "grad_norm": 0.03591204807162285,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.643499374389648,
+ "step": 297
+ },
+ {
+ "epoch": 4.139616055846423,
+ "grad_norm": 0.03269682824611664,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.657600402832031,
+ "step": 298
+ },
+ {
+ "epoch": 4.153577661431065,
+ "grad_norm": 0.03343537449836731,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.6352081298828125,
+ "step": 299
+ },
+ {
+ "epoch": 4.167539267015707,
+ "grad_norm": 0.034283801913261414,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.626757621765137,
+ "step": 300
+ },
+ {
+ "epoch": 4.181500872600349,
+ "grad_norm": 0.034614723175764084,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.6316680908203125,
+ "step": 301
+ },
+ {
+ "epoch": 4.195462478184991,
+ "grad_norm": 0.0340128131210804,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.648160934448242,
+ "step": 302
+ },
+ {
+ "epoch": 4.209424083769633,
+ "grad_norm": 0.03068280965089798,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.619738578796387,
+ "step": 303
+ },
+ {
+ "epoch": 4.223385689354275,
+ "grad_norm": 0.030269058421254158,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.609238624572754,
+ "step": 304
+ },
+ {
+ "epoch": 4.237347294938918,
+ "grad_norm": 0.03612440824508667,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.602194786071777,
+ "step": 305
+ },
+ {
+ "epoch": 4.2513089005235605,
+ "grad_norm": 0.044292811304330826,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.61159610748291,
+ "step": 306
+ },
+ {
+ "epoch": 4.265270506108203,
+ "grad_norm": 0.04449291527271271,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.6274237632751465,
+ "step": 307
+ },
+ {
+ "epoch": 4.279232111692845,
+ "grad_norm": 0.05014493316411972,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.589034080505371,
+ "step": 308
+ },
+ {
+ "epoch": 4.293193717277487,
+ "grad_norm": 0.05286206305027008,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.600770950317383,
+ "step": 309
+ },
+ {
+ "epoch": 4.307155322862129,
+ "grad_norm": 0.048459846526384354,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.5796098709106445,
+ "step": 310
+ },
+ {
+ "epoch": 4.321116928446771,
+ "grad_norm": 0.05329663306474686,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.590407371520996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335078534031414,
+ "grad_norm": 0.052595119923353195,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.583693504333496,
+ "step": 312
+ },
+ {
+ "epoch": 4.349040139616056,
+ "grad_norm": 0.05164601653814316,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.603281021118164,
+ "step": 313
+ },
+ {
+ "epoch": 4.363001745200698,
+ "grad_norm": 0.04657137766480446,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.591344833374023,
+ "step": 314
+ },
+ {
+ "epoch": 4.37696335078534,
+ "grad_norm": 0.03804599866271019,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.580578804016113,
+ "step": 315
+ },
+ {
+ "epoch": 4.390924956369982,
+ "grad_norm": 0.03835320845246315,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.597836971282959,
+ "step": 316
+ },
+ {
+ "epoch": 4.4048865619546245,
+ "grad_norm": 0.034059133380651474,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.567502021789551,
+ "step": 317
+ },
+ {
+ "epoch": 4.418848167539267,
+ "grad_norm": 0.03350386768579483,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.552022933959961,
+ "step": 318
+ },
+ {
+ "epoch": 4.43280977312391,
+ "grad_norm": 0.032574303448200226,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.559177875518799,
+ "step": 319
+ },
+ {
+ "epoch": 4.446771378708552,
+ "grad_norm": 0.030893854796886444,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.549288272857666,
+ "step": 320
+ },
+ {
+ "epoch": 4.460732984293194,
+ "grad_norm": 0.03255036100745201,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.5400390625,
+ "step": 321
+ },
+ {
+ "epoch": 4.474694589877836,
+ "grad_norm": 0.03009258769452572,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.5328474044799805,
+ "step": 322
+ },
+ {
+ "epoch": 4.488656195462478,
+ "grad_norm": 0.03178320452570915,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.5594482421875,
+ "step": 323
+ },
+ {
+ "epoch": 4.50261780104712,
+ "grad_norm": 0.03360465541481972,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.5402727127075195,
+ "step": 324
+ },
+ {
+ "epoch": 4.516579406631763,
+ "grad_norm": 0.04275049269199371,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.534474849700928,
+ "step": 325
+ },
+ {
+ "epoch": 4.530541012216405,
+ "grad_norm": 0.04140641540288925,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.5324859619140625,
+ "step": 326
+ },
+ {
+ "epoch": 4.544502617801047,
+ "grad_norm": 0.04769204929471016,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.4919352531433105,
+ "step": 327
+ },
+ {
+ "epoch": 4.558464223385689,
+ "grad_norm": 0.04402285814285278,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.520858287811279,
+ "step": 328
+ },
+ {
+ "epoch": 4.5724258289703315,
+ "grad_norm": 0.038526229560375214,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.523778438568115,
+ "step": 329
+ },
+ {
+ "epoch": 4.5863874345549736,
+ "grad_norm": 0.05422377213835716,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.517973899841309,
+ "step": 330
+ },
+ {
+ "epoch": 4.600349040139616,
+ "grad_norm": 0.07236721366643906,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.548083782196045,
+ "step": 331
+ },
+ {
+ "epoch": 4.614310645724259,
+ "grad_norm": 0.06861061602830887,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.535577774047852,
+ "step": 332
+ },
+ {
+ "epoch": 4.628272251308901,
+ "grad_norm": 0.07148340344429016,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.54155158996582,
+ "step": 333
+ },
+ {
+ "epoch": 4.642233856893543,
+ "grad_norm": 0.07838302850723267,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.552852630615234,
+ "step": 334
+ },
+ {
+ "epoch": 4.656195462478185,
+ "grad_norm": 0.10946758091449738,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.555066108703613,
+ "step": 335
+ },
+ {
+ "epoch": 4.670157068062827,
+ "grad_norm": 0.10143530368804932,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.584932804107666,
+ "step": 336
+ },
+ {
+ "epoch": 4.684118673647469,
+ "grad_norm": 0.08926606178283691,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.570462226867676,
+ "step": 337
+ },
+ {
+ "epoch": 4.698080279232112,
+ "grad_norm": 0.07060826569795609,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.5716753005981445,
+ "step": 338
+ },
+ {
+ "epoch": 4.712041884816754,
+ "grad_norm": 0.0418456569314003,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.556496620178223,
+ "step": 339
+ },
+ {
+ "epoch": 4.726003490401396,
+ "grad_norm": 0.05952411890029907,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.567801475524902,
+ "step": 340
+ },
+ {
+ "epoch": 4.739965095986038,
+ "grad_norm": 0.042999885976314545,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.537403106689453,
+ "step": 341
+ },
+ {
+ "epoch": 4.7539267015706805,
+ "grad_norm": 0.04426449537277222,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.552387237548828,
+ "step": 342
+ },
+ {
+ "epoch": 4.767888307155323,
+ "grad_norm": 0.040835145860910416,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.53051233291626,
+ "step": 343
+ },
+ {
+ "epoch": 4.781849912739965,
+ "grad_norm": 0.03819291293621063,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.531174659729004,
+ "step": 344
+ },
+ {
+ "epoch": 4.795811518324607,
+ "grad_norm": 0.030088599771261215,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.499328136444092,
+ "step": 345
+ },
+ {
+ "epoch": 4.80977312390925,
+ "grad_norm": 0.029984362423419952,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.504540920257568,
+ "step": 346
+ },
+ {
+ "epoch": 4.823734729493892,
+ "grad_norm": 0.027402915060520172,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.50899076461792,
+ "step": 347
+ },
+ {
+ "epoch": 4.837696335078534,
+ "grad_norm": 0.025096213445067406,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.480479717254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.851657940663176,
+ "grad_norm": 0.026828132569789886,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.501729965209961,
+ "step": 349
+ },
+ {
+ "epoch": 4.865619546247818,
+ "grad_norm": 0.023715168237686157,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.466860771179199,
+ "step": 350
+ },
+ {
+ "epoch": 4.879581151832461,
+ "grad_norm": 0.024503640830516815,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.491827964782715,
+ "step": 351
+ },
+ {
+ "epoch": 4.893542757417103,
+ "grad_norm": 0.020891105756163597,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.478384971618652,
+ "step": 352
+ },
+ {
+ "epoch": 4.907504363001745,
+ "grad_norm": 0.021373765543103218,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.4612135887146,
+ "step": 353
+ },
+ {
+ "epoch": 4.9214659685863875,
+ "grad_norm": 0.021773340180516243,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.4547319412231445,
+ "step": 354
+ },
+ {
+ "epoch": 4.93542757417103,
+ "grad_norm": 0.019637586548924446,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.431741714477539,
+ "step": 355
+ },
+ {
+ "epoch": 4.949389179755672,
+ "grad_norm": 0.023108696565032005,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.448612213134766,
+ "step": 356
+ },
+ {
+ "epoch": 4.963350785340314,
+ "grad_norm": 0.021935103461146355,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.447112083435059,
+ "step": 357
+ },
+ {
+ "epoch": 4.977312390924956,
+ "grad_norm": 0.023260753601789474,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.436091423034668,
+ "step": 358
+ },
+ {
+ "epoch": 4.991273996509599,
+ "grad_norm": 0.021355057135224342,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.408421993255615,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.014992325566709042,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 3.4004223346710205,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6834986805915833,
+ "eval_runtime": 60.4148,
+ "eval_samples_per_second": 40.421,
+ "eval_steps_per_second": 0.646,
+ "step": 360
+ },
+ {
+ "epoch": 5.013961605584642,
+ "grad_norm": 0.02471218630671501,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.396238803863525,
+ "step": 361
+ },
+ {
+ "epoch": 5.027923211169284,
+ "grad_norm": 0.027672411873936653,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.4048919677734375,
+ "step": 362
+ },
+ {
+ "epoch": 5.041884816753926,
+ "grad_norm": 0.031560689210891724,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.399665832519531,
+ "step": 363
+ },
+ {
+ "epoch": 5.055846422338569,
+ "grad_norm": 0.03459061309695244,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.396360397338867,
+ "step": 364
+ },
+ {
+ "epoch": 5.069808027923211,
+ "grad_norm": 0.04040960222482681,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.405309677124023,
+ "step": 365
+ },
+ {
+ "epoch": 5.0837696335078535,
+ "grad_norm": 0.05444670096039772,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.408894062042236,
+ "step": 366
+ },
+ {
+ "epoch": 5.097731239092496,
+ "grad_norm": 0.05816745012998581,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.403566360473633,
+ "step": 367
+ },
+ {
+ "epoch": 5.111692844677138,
+ "grad_norm": 0.06330069154500961,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.395961761474609,
+ "step": 368
+ },
+ {
+ "epoch": 5.12565445026178,
+ "grad_norm": 0.07458917796611786,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.400968551635742,
+ "step": 369
+ },
+ {
+ "epoch": 5.139616055846423,
+ "grad_norm": 0.06231356039643288,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.412624359130859,
+ "step": 370
+ },
+ {
+ "epoch": 5.153577661431065,
+ "grad_norm": 0.05212346464395523,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.394794464111328,
+ "step": 371
+ },
+ {
+ "epoch": 5.167539267015707,
+ "grad_norm": 0.050003375858068466,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.380125045776367,
+ "step": 372
+ },
+ {
+ "epoch": 5.181500872600349,
+ "grad_norm": 0.048393815755844116,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.390806198120117,
+ "step": 373
+ },
+ {
+ "epoch": 5.195462478184991,
+ "grad_norm": 0.04200794920325279,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.387467384338379,
+ "step": 374
+ },
+ {
+ "epoch": 5.209424083769633,
+ "grad_norm": 0.04931448772549629,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.37552547454834,
+ "step": 375
+ },
+ {
+ "epoch": 5.223385689354275,
+ "grad_norm": 0.03851857781410217,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.373126983642578,
+ "step": 376
+ },
+ {
+ "epoch": 5.237347294938918,
+ "grad_norm": 0.03458717465400696,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.346826553344727,
+ "step": 377
+ },
+ {
+ "epoch": 5.2513089005235605,
+ "grad_norm": 0.031107677146792412,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.353341102600098,
+ "step": 378
+ },
+ {
+ "epoch": 5.265270506108203,
+ "grad_norm": 0.03077058307826519,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.358076095581055,
+ "step": 379
+ },
+ {
+ "epoch": 5.279232111692845,
+ "grad_norm": 0.028170401230454445,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.354019641876221,
+ "step": 380
+ },
+ {
+ "epoch": 5.293193717277487,
+ "grad_norm": 0.02871990203857422,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.368258476257324,
+ "step": 381
+ },
+ {
+ "epoch": 5.307155322862129,
+ "grad_norm": 0.03046584501862526,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.342505931854248,
+ "step": 382
+ },
+ {
+ "epoch": 5.321116928446771,
+ "grad_norm": 0.02781563811004162,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.339411735534668,
+ "step": 383
+ },
+ {
+ "epoch": 5.335078534031414,
+ "grad_norm": 0.028825288638472557,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.341948509216309,
+ "step": 384
+ },
+ {
+ "epoch": 5.349040139616056,
+ "grad_norm": 0.026684079319238663,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.340881824493408,
+ "step": 385
+ },
+ {
+ "epoch": 5.363001745200698,
+ "grad_norm": 0.02429167740046978,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.342733383178711,
+ "step": 386
+ },
+ {
+ "epoch": 5.37696335078534,
+ "grad_norm": 0.02855239436030388,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.313878536224365,
+ "step": 387
+ },
+ {
+ "epoch": 5.390924956369982,
+ "grad_norm": 0.029651548713445663,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.312665939331055,
+ "step": 388
+ },
+ {
+ "epoch": 5.4048865619546245,
+ "grad_norm": 0.027191132307052612,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3086838722229,
+ "step": 389
+ },
+ {
+ "epoch": 5.418848167539267,
+ "grad_norm": 0.026719525456428528,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.305359840393066,
+ "step": 390
+ },
+ {
+ "epoch": 5.43280977312391,
+ "grad_norm": 0.026763100177049637,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.3164381980896,
+ "step": 391
+ },
+ {
+ "epoch": 5.446771378708552,
+ "grad_norm": 0.03125109151005745,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.3083953857421875,
+ "step": 392
+ },
+ {
+ "epoch": 5.460732984293194,
+ "grad_norm": 0.0394827201962471,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.275996208190918,
+ "step": 393
+ },
+ {
+ "epoch": 5.474694589877836,
+ "grad_norm": 0.040579475462436676,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.292995929718018,
+ "step": 394
+ },
+ {
+ "epoch": 5.488656195462478,
+ "grad_norm": 0.037223003804683685,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.298526763916016,
+ "step": 395
+ },
+ {
+ "epoch": 5.50261780104712,
+ "grad_norm": 0.04207765311002731,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.287985801696777,
+ "step": 396
+ },
+ {
+ "epoch": 5.516579406631763,
+ "grad_norm": 0.048299964517354965,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.244608402252197,
+ "step": 397
+ },
+ {
+ "epoch": 5.530541012216405,
+ "grad_norm": 0.04650307446718216,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.285453796386719,
+ "step": 398
+ },
+ {
+ "epoch": 5.544502617801047,
+ "grad_norm": 0.045339711010456085,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.312187671661377,
+ "step": 399
+ },
+ {
+ "epoch": 5.558464223385689,
+ "grad_norm": 0.041067708283662796,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.292692184448242,
+ "step": 400
+ },
+ {
+ "epoch": 5.5724258289703315,
+ "grad_norm": 0.05455026403069496,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.253364562988281,
+ "step": 401
+ },
+ {
+ "epoch": 5.5863874345549736,
+ "grad_norm": 0.06660915911197662,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.262646675109863,
+ "step": 402
+ },
+ {
+ "epoch": 5.600349040139616,
+ "grad_norm": 0.06634793430566788,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.274322986602783,
+ "step": 403
+ },
+ {
+ "epoch": 5.614310645724259,
+ "grad_norm": 0.08996330946683884,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.287951469421387,
+ "step": 404
+ },
+ {
+ "epoch": 5.628272251308901,
+ "grad_norm": 0.08526470512151718,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.3007097244262695,
+ "step": 405
+ },
+ {
+ "epoch": 5.642233856893543,
+ "grad_norm": 0.08766836673021317,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.332538604736328,
+ "step": 406
+ },
+ {
+ "epoch": 5.656195462478185,
+ "grad_norm": 0.07055048644542694,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.296169281005859,
+ "step": 407
+ },
+ {
+ "epoch": 5.670157068062827,
+ "grad_norm": 0.07520575076341629,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.309683322906494,
+ "step": 408
+ },
+ {
+ "epoch": 5.684118673647469,
+ "grad_norm": 0.06115756183862686,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.311301231384277,
+ "step": 409
+ },
+ {
+ "epoch": 5.698080279232112,
+ "grad_norm": 0.05862593650817871,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.28066349029541,
+ "step": 410
+ },
+ {
+ "epoch": 5.712041884816754,
+ "grad_norm": 0.06169601157307625,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.286840438842773,
+ "step": 411
+ },
+ {
+ "epoch": 5.726003490401396,
+ "grad_norm": 0.05270719155669212,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.302594184875488,
+ "step": 412
+ },
+ {
+ "epoch": 5.739965095986038,
+ "grad_norm": 0.03912634402513504,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.262563705444336,
+ "step": 413
+ },
+ {
+ "epoch": 5.7539267015706805,
+ "grad_norm": 0.03215811401605606,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.269310474395752,
+ "step": 414
+ },
+ {
+ "epoch": 5.767888307155323,
+ "grad_norm": 0.03236516937613487,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.268766403198242,
+ "step": 415
+ },
+ {
+ "epoch": 5.781849912739965,
+ "grad_norm": 0.027429452165961266,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.283801555633545,
+ "step": 416
+ },
+ {
+ "epoch": 5.795811518324607,
+ "grad_norm": 0.02990538813173771,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.22291374206543,
+ "step": 417
+ },
+ {
+ "epoch": 5.80977312390925,
+ "grad_norm": 0.030410146340727806,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.2281599044799805,
+ "step": 418
+ },
+ {
+ "epoch": 5.823734729493892,
+ "grad_norm": 0.030068187043070793,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.250294208526611,
+ "step": 419
+ },
+ {
+ "epoch": 5.837696335078534,
+ "grad_norm": 0.02800513058900833,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.226508617401123,
+ "step": 420
+ },
+ {
+ "epoch": 5.851657940663176,
+ "grad_norm": 0.023823332041502,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.247838020324707,
+ "step": 421
+ },
+ {
+ "epoch": 5.865619546247818,
+ "grad_norm": 0.024974249303340912,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.2392473220825195,
+ "step": 422
+ },
+ {
+ "epoch": 5.879581151832461,
+ "grad_norm": 0.023320546373724937,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.209543704986572,
+ "step": 423
+ },
+ {
+ "epoch": 5.893542757417103,
+ "grad_norm": 0.024201491847634315,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.219725608825684,
+ "step": 424
+ },
+ {
+ "epoch": 5.907504363001745,
+ "grad_norm": 0.021649666130542755,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.203141212463379,
+ "step": 425
+ },
+ {
+ "epoch": 5.9214659685863875,
+ "grad_norm": 0.0248002577573061,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.179058074951172,
+ "step": 426
+ },
+ {
+ "epoch": 5.93542757417103,
+ "grad_norm": 0.028549587354063988,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.202126502990723,
+ "step": 427
+ },
+ {
+ "epoch": 5.949389179755672,
+ "grad_norm": 0.0331057533621788,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.208107948303223,
+ "step": 428
+ },
+ {
+ "epoch": 5.963350785340314,
+ "grad_norm": 0.031843364238739014,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.184683799743652,
+ "step": 429
+ },
+ {
+ "epoch": 5.977312390924956,
+ "grad_norm": 0.024927377700805664,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.180486679077148,
+ "step": 430
+ },
+ {
+ "epoch": 5.991273996509599,
+ "grad_norm": 0.026685267686843872,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.186746597290039,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.018545351922512054,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 3.217043876647949,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.653471827507019,
+ "eval_runtime": 61.1586,
+ "eval_samples_per_second": 39.929,
+ "eval_steps_per_second": 0.638,
+ "step": 432
+ },
+ {
+ "epoch": 6.013961605584642,
+ "grad_norm": 0.025291118770837784,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.156534194946289,
+ "step": 433
+ },
+ {
+ "epoch": 6.027923211169284,
+ "grad_norm": 0.03774752840399742,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.144041538238525,
+ "step": 434
+ },
+ {
+ "epoch": 6.041884816753926,
+ "grad_norm": 0.04244692996144295,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.158862113952637,
+ "step": 435
+ },
+ {
+ "epoch": 6.055846422338569,
+ "grad_norm": 0.039782389998435974,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.166165351867676,
+ "step": 436
+ },
+ {
+ "epoch": 6.069808027923211,
+ "grad_norm": 0.05489896610379219,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.155766487121582,
+ "step": 437
+ },
+ {
+ "epoch": 6.0837696335078535,
+ "grad_norm": 0.059844326227903366,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.153733253479004,
+ "step": 438
+ },
+ {
+ "epoch": 6.097731239092496,
+ "grad_norm": 0.05539938807487488,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.130112648010254,
+ "step": 439
+ },
+ {
+ "epoch": 6.111692844677138,
+ "grad_norm": 0.05283183231949806,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.136910915374756,
+ "step": 440
+ },
+ {
+ "epoch": 6.12565445026178,
+ "grad_norm": 0.054407261312007904,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.165216445922852,
+ "step": 441
+ },
+ {
+ "epoch": 6.139616055846423,
+ "grad_norm": 0.0509788915514946,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.140888214111328,
+ "step": 442
+ },
+ {
+ "epoch": 6.153577661431065,
+ "grad_norm": 0.05018748342990875,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.13115930557251,
+ "step": 443
+ },
+ {
+ "epoch": 6.167539267015707,
+ "grad_norm": 0.047012437134981155,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.146507263183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181500872600349,
+ "grad_norm": 0.04943335801362991,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.135608673095703,
+ "step": 445
+ },
+ {
+ "epoch": 6.195462478184991,
+ "grad_norm": 0.05304260924458504,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.136394023895264,
+ "step": 446
+ },
+ {
+ "epoch": 6.209424083769633,
+ "grad_norm": 0.04554047808051109,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.124661922454834,
+ "step": 447
+ },
+ {
+ "epoch": 6.223385689354275,
+ "grad_norm": 0.040672432631254196,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.129773139953613,
+ "step": 448
+ },
+ {
+ "epoch": 6.237347294938918,
+ "grad_norm": 0.03442879393696785,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.128028392791748,
+ "step": 449
+ },
+ {
+ "epoch": 6.2513089005235605,
+ "grad_norm": 0.03927934914827347,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.092485427856445,
+ "step": 450
+ },
+ {
+ "epoch": 6.265270506108203,
+ "grad_norm": 0.029380977153778076,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.077669143676758,
+ "step": 451
+ },
+ {
+ "epoch": 6.279232111692845,
+ "grad_norm": 0.03331821784377098,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.138857841491699,
+ "step": 452
+ },
+ {
+ "epoch": 6.293193717277487,
+ "grad_norm": 0.02939118817448616,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.082937240600586,
+ "step": 453
+ },
+ {
+ "epoch": 6.307155322862129,
+ "grad_norm": 0.02730412967503071,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.093183517456055,
+ "step": 454
+ },
+ {
+ "epoch": 6.321116928446771,
+ "grad_norm": 0.02964780665934086,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.1028151512146,
+ "step": 455
+ },
+ {
+ "epoch": 6.335078534031414,
+ "grad_norm": 0.026350129395723343,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.066362380981445,
+ "step": 456
+ },
+ {
+ "epoch": 6.349040139616056,
+ "grad_norm": 0.027010295540094376,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.070037841796875,
+ "step": 457
+ },
+ {
+ "epoch": 6.363001745200698,
+ "grad_norm": 0.02906128577888012,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.107708930969238,
+ "step": 458
+ },
+ {
+ "epoch": 6.37696335078534,
+ "grad_norm": 0.028619443997740746,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.073182582855225,
+ "step": 459
+ },
+ {
+ "epoch": 6.390924956369982,
+ "grad_norm": 0.029981834813952446,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.076602935791016,
+ "step": 460
+ },
+ {
+ "epoch": 6.4048865619546245,
+ "grad_norm": 0.03255423903465271,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.063301086425781,
+ "step": 461
+ },
+ {
+ "epoch": 6.418848167539267,
+ "grad_norm": 0.03863866627216339,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.041812896728516,
+ "step": 462
+ },
+ {
+ "epoch": 6.43280977312391,
+ "grad_norm": 0.03860306739807129,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.06298828125,
+ "step": 463
+ },
+ {
+ "epoch": 6.446771378708552,
+ "grad_norm": 0.03378435596823692,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.0434370040893555,
+ "step": 464
+ },
+ {
+ "epoch": 6.460732984293194,
+ "grad_norm": 0.03276759013533592,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.036187171936035,
+ "step": 465
+ },
+ {
+ "epoch": 6.474694589877836,
+ "grad_norm": 0.04130513221025467,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.039098739624023,
+ "step": 466
+ },
+ {
+ "epoch": 6.488656195462478,
+ "grad_norm": 0.05389940366148949,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.062798976898193,
+ "step": 467
+ },
+ {
+ "epoch": 6.50261780104712,
+ "grad_norm": 0.05881704390048981,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.067028522491455,
+ "step": 468
+ },
+ {
+ "epoch": 6.516579406631763,
+ "grad_norm": 0.061897799372673035,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.056947231292725,
+ "step": 469
+ },
+ {
+ "epoch": 6.530541012216405,
+ "grad_norm": 0.06549323350191116,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.089262008666992,
+ "step": 470
+ },
+ {
+ "epoch": 6.544502617801047,
+ "grad_norm": 0.048082590103149414,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.055979251861572,
+ "step": 471
+ },
+ {
+ "epoch": 6.558464223385689,
+ "grad_norm": 0.03768027946352959,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.0660600662231445,
+ "step": 472
+ },
+ {
+ "epoch": 6.5724258289703315,
+ "grad_norm": 0.043922603130340576,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 5.036745071411133,
+ "step": 473
+ },
+ {
+ "epoch": 6.5863874345549736,
+ "grad_norm": 0.044848836958408356,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.073650360107422,
+ "step": 474
+ },
+ {
+ "epoch": 6.600349040139616,
+ "grad_norm": 0.05841026455163956,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.053045272827148,
+ "step": 475
+ },
+ {
+ "epoch": 6.614310645724259,
+ "grad_norm": 0.0650017261505127,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 5.052206039428711,
+ "step": 476
+ },
+ {
+ "epoch": 6.628272251308901,
+ "grad_norm": 0.048979002982378006,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 5.054277420043945,
+ "step": 477
+ },
+ {
+ "epoch": 6.642233856893543,
+ "grad_norm": 0.05085799843072891,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 5.063844680786133,
+ "step": 478
+ },
+ {
+ "epoch": 6.656195462478185,
+ "grad_norm": 0.04336017370223999,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 5.045474052429199,
+ "step": 479
+ },
+ {
+ "epoch": 6.670157068062827,
+ "grad_norm": 0.04113380238413811,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 5.031338214874268,
+ "step": 480
+ },
+ {
+ "epoch": 6.684118673647469,
+ "grad_norm": 0.03199538215994835,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 5.030025959014893,
+ "step": 481
+ },
+ {
+ "epoch": 6.698080279232112,
+ "grad_norm": 0.03043113648891449,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 5.033163070678711,
+ "step": 482
+ },
+ {
+ "epoch": 6.712041884816754,
+ "grad_norm": 0.031694378703832626,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 5.004739761352539,
+ "step": 483
+ },
+ {
+ "epoch": 6.726003490401396,
+ "grad_norm": 0.030243348330259323,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 5.012566089630127,
+ "step": 484
+ },
+ {
+ "epoch": 6.739965095986038,
+ "grad_norm": 0.02937416359782219,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 5.039803504943848,
+ "step": 485
+ },
+ {
+ "epoch": 6.7539267015706805,
+ "grad_norm": 0.029616443440318108,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 5.026117324829102,
+ "step": 486
+ },
+ {
+ "epoch": 6.767888307155323,
+ "grad_norm": 0.029567131772637367,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 5.006569862365723,
+ "step": 487
+ },
+ {
+ "epoch": 6.781849912739965,
+ "grad_norm": 0.029443370178341866,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.9941487312316895,
+ "step": 488
+ },
+ {
+ "epoch": 6.795811518324607,
+ "grad_norm": 0.03434352949261665,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.976431846618652,
+ "step": 489
+ },
+ {
+ "epoch": 6.80977312390925,
+ "grad_norm": 0.03845234215259552,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.993168354034424,
+ "step": 490
+ },
+ {
+ "epoch": 6.823734729493892,
+ "grad_norm": 0.05058208107948303,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.030162334442139,
+ "step": 491
+ },
+ {
+ "epoch": 6.837696335078534,
+ "grad_norm": 0.05961710214614868,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 5.007053375244141,
+ "step": 492
+ },
+ {
+ "epoch": 6.851657940663176,
+ "grad_norm": 0.056259628385305405,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 5.020356178283691,
+ "step": 493
+ },
+ {
+ "epoch": 6.865619546247818,
+ "grad_norm": 0.048628658056259155,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 5.037652969360352,
+ "step": 494
+ },
+ {
+ "epoch": 6.879581151832461,
+ "grad_norm": 0.040836453437805176,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.990390777587891,
+ "step": 495
+ },
+ {
+ "epoch": 6.893542757417103,
+ "grad_norm": 0.04112066701054573,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 5.017822742462158,
+ "step": 496
+ },
+ {
+ "epoch": 6.907504363001745,
+ "grad_norm": 0.03956829756498337,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.993971824645996,
+ "step": 497
+ },
+ {
+ "epoch": 6.9214659685863875,
+ "grad_norm": 0.03887917473912239,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.982889175415039,
+ "step": 498
+ },
+ {
+ "epoch": 6.93542757417103,
+ "grad_norm": 0.031012587249279022,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.961717128753662,
+ "step": 499
+ },
+ {
+ "epoch": 6.949389179755672,
+ "grad_norm": 0.03443489596247673,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.9904046058654785,
+ "step": 500
+ },
+ {
+ "epoch": 6.963350785340314,
+ "grad_norm": 0.035175010561943054,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.983619213104248,
+ "step": 501
+ },
+ {
+ "epoch": 6.977312390924956,
+ "grad_norm": 0.031724050641059875,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 5.000522613525391,
+ "step": 502
+ },
+ {
+ "epoch": 6.991273996509599,
+ "grad_norm": 0.032037995755672455,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.947064399719238,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.01883271522819996,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 3.112166166305542,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6296830177307129,
+ "eval_runtime": 59.5495,
+ "eval_samples_per_second": 41.008,
+ "eval_steps_per_second": 0.655,
+ "step": 504
+ },
+ {
+ "epoch": 7.013961605584642,
+ "grad_norm": 0.031091563403606415,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.919820785522461,
+ "step": 505
+ },
+ {
+ "epoch": 7.027923211169284,
+ "grad_norm": 0.04074421525001526,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.945765972137451,
+ "step": 506
+ },
+ {
+ "epoch": 7.041884816753926,
+ "grad_norm": 0.04225052148103714,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.935208797454834,
+ "step": 507
+ },
+ {
+ "epoch": 7.055846422338569,
+ "grad_norm": 0.0399051196873188,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.951314926147461,
+ "step": 508
+ },
+ {
+ "epoch": 7.069808027923211,
+ "grad_norm": 0.04785890504717827,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.92843770980835,
+ "step": 509
+ },
+ {
+ "epoch": 7.0837696335078535,
+ "grad_norm": 0.04922301694750786,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.901020050048828,
+ "step": 510
+ },
+ {
+ "epoch": 7.097731239092496,
+ "grad_norm": 0.05307295173406601,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.961151599884033,
+ "step": 511
+ },
+ {
+ "epoch": 7.111692844677138,
+ "grad_norm": 0.07003362476825714,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.9427876472473145,
+ "step": 512
+ },
+ {
+ "epoch": 7.12565445026178,
+ "grad_norm": 0.06498776376247406,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.947838306427002,
+ "step": 513
+ },
+ {
+ "epoch": 7.139616055846423,
+ "grad_norm": 0.058723628520965576,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.938859939575195,
+ "step": 514
+ },
+ {
+ "epoch": 7.153577661431065,
+ "grad_norm": 0.06590746343135834,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.962221145629883,
+ "step": 515
+ },
+ {
+ "epoch": 7.167539267015707,
+ "grad_norm": 0.052714601159095764,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.931011199951172,
+ "step": 516
+ },
+ {
+ "epoch": 7.181500872600349,
+ "grad_norm": 0.04296804964542389,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.93586540222168,
+ "step": 517
+ },
+ {
+ "epoch": 7.195462478184991,
+ "grad_norm": 0.03699173778295517,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.910763263702393,
+ "step": 518
+ },
+ {
+ "epoch": 7.209424083769633,
+ "grad_norm": 0.0317913293838501,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.9256439208984375,
+ "step": 519
+ },
+ {
+ "epoch": 7.223385689354275,
+ "grad_norm": 0.030313925817608833,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.938329696655273,
+ "step": 520
+ },
+ {
+ "epoch": 7.237347294938918,
+ "grad_norm": 0.02879229187965393,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.924966812133789,
+ "step": 521
+ },
+ {
+ "epoch": 7.2513089005235605,
+ "grad_norm": 0.02689451351761818,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.9079742431640625,
+ "step": 522
+ },
+ {
+ "epoch": 7.265270506108203,
+ "grad_norm": 0.028403930366039276,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.913238525390625,
+ "step": 523
+ },
+ {
+ "epoch": 7.279232111692845,
+ "grad_norm": 0.025912854820489883,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.896944046020508,
+ "step": 524
+ },
+ {
+ "epoch": 7.293193717277487,
+ "grad_norm": 0.024899760261178017,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.884920120239258,
+ "step": 525
+ },
+ {
+ "epoch": 7.307155322862129,
+ "grad_norm": 0.026912573724985123,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.9181060791015625,
+ "step": 526
+ },
+ {
+ "epoch": 7.321116928446771,
+ "grad_norm": 0.029539721086621284,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.914867401123047,
+ "step": 527
+ },
+ {
+ "epoch": 7.335078534031414,
+ "grad_norm": 0.028527872636914253,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.907217979431152,
+ "step": 528
+ },
+ {
+ "epoch": 7.349040139616056,
+ "grad_norm": 0.025879016146063805,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.905783176422119,
+ "step": 529
+ },
+ {
+ "epoch": 7.363001745200698,
+ "grad_norm": 0.026395199820399284,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.880505561828613,
+ "step": 530
+ },
+ {
+ "epoch": 7.37696335078534,
+ "grad_norm": 0.025072351098060608,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.870449066162109,
+ "step": 531
+ },
+ {
+ "epoch": 7.390924956369982,
+ "grad_norm": 0.02751019410789013,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.888952255249023,
+ "step": 532
+ },
+ {
+ "epoch": 7.4048865619546245,
+ "grad_norm": 0.025810658931732178,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.894552230834961,
+ "step": 533
+ },
+ {
+ "epoch": 7.418848167539267,
+ "grad_norm": 0.02399870567023754,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.8768229484558105,
+ "step": 534
+ },
+ {
+ "epoch": 7.43280977312391,
+ "grad_norm": 0.024639368057250977,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.9030914306640625,
+ "step": 535
+ },
+ {
+ "epoch": 7.446771378708552,
+ "grad_norm": 0.02543291077017784,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.86467170715332,
+ "step": 536
+ },
+ {
+ "epoch": 7.460732984293194,
+ "grad_norm": 0.02702544629573822,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.888283729553223,
+ "step": 537
+ },
+ {
+ "epoch": 7.474694589877836,
+ "grad_norm": 0.028904743492603302,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.866081237792969,
+ "step": 538
+ },
+ {
+ "epoch": 7.488656195462478,
+ "grad_norm": 0.03055848367512226,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.890739440917969,
+ "step": 539
+ },
+ {
+ "epoch": 7.50261780104712,
+ "grad_norm": 0.03410084918141365,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.862918853759766,
+ "step": 540
+ },
+ {
+ "epoch": 7.516579406631763,
+ "grad_norm": 0.03738495334982872,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.864728927612305,
+ "step": 541
+ },
+ {
+ "epoch": 7.530541012216405,
+ "grad_norm": 0.037646882236003876,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.865185737609863,
+ "step": 542
+ },
+ {
+ "epoch": 7.544502617801047,
+ "grad_norm": 0.03418618440628052,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.866114139556885,
+ "step": 543
+ },
+ {
+ "epoch": 7.558464223385689,
+ "grad_norm": 0.036924004554748535,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.883186340332031,
+ "step": 544
+ },
+ {
+ "epoch": 7.5724258289703315,
+ "grad_norm": 0.03590598329901695,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.863725662231445,
+ "step": 545
+ },
+ {
+ "epoch": 7.5863874345549736,
+ "grad_norm": 0.037363454699516296,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.841030120849609,
+ "step": 546
+ },
+ {
+ "epoch": 7.600349040139616,
+ "grad_norm": 0.046537864953279495,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.873837471008301,
+ "step": 547
+ },
+ {
+ "epoch": 7.614310645724259,
+ "grad_norm": 0.057226769626140594,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.865679740905762,
+ "step": 548
+ },
+ {
+ "epoch": 7.628272251308901,
+ "grad_norm": 0.06175961345434189,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.888645648956299,
+ "step": 549
+ },
+ {
+ "epoch": 7.642233856893543,
+ "grad_norm": 0.05063694715499878,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.877828598022461,
+ "step": 550
+ },
+ {
+ "epoch": 7.656195462478185,
+ "grad_norm": 0.03950628638267517,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.88637638092041,
+ "step": 551
+ },
+ {
+ "epoch": 7.670157068062827,
+ "grad_norm": 0.04183337092399597,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.8597917556762695,
+ "step": 552
+ },
+ {
+ "epoch": 7.684118673647469,
+ "grad_norm": 0.04642999917268753,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.880380153656006,
+ "step": 553
+ },
+ {
+ "epoch": 7.698080279232112,
+ "grad_norm": 0.05333106964826584,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.853932857513428,
+ "step": 554
+ },
+ {
+ "epoch": 7.712041884816754,
+ "grad_norm": 0.04764975979924202,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.870235443115234,
+ "step": 555
+ },
+ {
+ "epoch": 7.726003490401396,
+ "grad_norm": 0.043072231113910675,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.822697639465332,
+ "step": 556
+ },
+ {
+ "epoch": 7.739965095986038,
+ "grad_norm": 0.03913857042789459,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.849332809448242,
+ "step": 557
+ },
+ {
+ "epoch": 7.7539267015706805,
+ "grad_norm": 0.034836407750844955,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.842280387878418,
+ "step": 558
+ },
+ {
+ "epoch": 7.767888307155323,
+ "grad_norm": 0.033344294875860214,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.837811470031738,
+ "step": 559
+ },
+ {
+ "epoch": 7.781849912739965,
+ "grad_norm": 0.03342701122164726,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.876040458679199,
+ "step": 560
+ },
+ {
+ "epoch": 7.795811518324607,
+ "grad_norm": 0.034638747572898865,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.830340385437012,
+ "step": 561
+ },
+ {
+ "epoch": 7.80977312390925,
+ "grad_norm": 0.03358135372400284,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.828607559204102,
+ "step": 562
+ },
+ {
+ "epoch": 7.823734729493892,
+ "grad_norm": 0.036001041531562805,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.827905654907227,
+ "step": 563
+ },
+ {
+ "epoch": 7.837696335078534,
+ "grad_norm": 0.038079630583524704,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.866955280303955,
+ "step": 564
+ },
+ {
+ "epoch": 7.851657940663176,
+ "grad_norm": 0.03728470206260681,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.828039169311523,
+ "step": 565
+ },
+ {
+ "epoch": 7.865619546247818,
+ "grad_norm": 0.03920375928282738,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.841631889343262,
+ "step": 566
+ },
+ {
+ "epoch": 7.879581151832461,
+ "grad_norm": 0.032353926450014114,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.861110210418701,
+ "step": 567
+ },
+ {
+ "epoch": 7.893542757417103,
+ "grad_norm": 0.03241422027349472,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.818441390991211,
+ "step": 568
+ },
+ {
+ "epoch": 7.907504363001745,
+ "grad_norm": 0.03879080340266228,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.823749542236328,
+ "step": 569
+ },
+ {
+ "epoch": 7.9214659685863875,
+ "grad_norm": 0.03486498072743416,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.830697059631348,
+ "step": 570
+ },
+ {
+ "epoch": 7.93542757417103,
+ "grad_norm": 0.03279713913798332,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.815736770629883,
+ "step": 571
+ },
+ {
+ "epoch": 7.949389179755672,
+ "grad_norm": 0.03313925117254257,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.8364033699035645,
+ "step": 572
+ },
+ {
+ "epoch": 7.963350785340314,
+ "grad_norm": 0.03405062481760979,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.816010475158691,
+ "step": 573
+ },
+ {
+ "epoch": 7.977312390924956,
+ "grad_norm": 0.03259236365556717,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.813413619995117,
+ "step": 574
+ },
+ {
+ "epoch": 7.991273996509599,
+ "grad_norm": 0.030408985912799835,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.833840370178223,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.022680889815092087,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 2.99565052986145,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6126086115837097,
+ "eval_runtime": 60.2662,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 576
+ },
+ {
+ "epoch": 8.013961605584642,
+ "grad_norm": 0.029732126742601395,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.7816314697265625,
+ "step": 577
+ },
+ {
+ "epoch": 8.027923211169284,
+ "grad_norm": 0.04357453063130379,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.7575225830078125,
+ "step": 578
+ },
+ {
+ "epoch": 8.041884816753926,
+ "grad_norm": 0.051903221756219864,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.802519798278809,
+ "step": 579
+ },
+ {
+ "epoch": 8.055846422338568,
+ "grad_norm": 0.0539076030254364,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.799004554748535,
+ "step": 580
+ },
+ {
+ "epoch": 8.06980802792321,
+ "grad_norm": 0.043642807751894,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.7896528244018555,
+ "step": 581
+ },
+ {
+ "epoch": 8.083769633507853,
+ "grad_norm": 0.0469554178416729,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.765430927276611,
+ "step": 582
+ },
+ {
+ "epoch": 8.097731239092496,
+ "grad_norm": 0.042060062289237976,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.784132957458496,
+ "step": 583
+ },
+ {
+ "epoch": 8.111692844677139,
+ "grad_norm": 0.035101912915706635,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.767234802246094,
+ "step": 584
+ },
+ {
+ "epoch": 8.12565445026178,
+ "grad_norm": 0.042576879262924194,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.762052536010742,
+ "step": 585
+ },
+ {
+ "epoch": 8.139616055846423,
+ "grad_norm": 0.04231419041752815,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.752150535583496,
+ "step": 586
+ },
+ {
+ "epoch": 8.153577661431065,
+ "grad_norm": 0.046912599354982376,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.771881103515625,
+ "step": 587
+ },
+ {
+ "epoch": 8.167539267015707,
+ "grad_norm": 0.051271144300699234,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.773284912109375,
+ "step": 588
+ },
+ {
+ "epoch": 8.181500872600349,
+ "grad_norm": 0.04655958339571953,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.769917964935303,
+ "step": 589
+ },
+ {
+ "epoch": 8.195462478184991,
+ "grad_norm": 0.04463687539100647,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.77938175201416,
+ "step": 590
+ },
+ {
+ "epoch": 8.209424083769633,
+ "grad_norm": 0.047959886491298676,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.785962104797363,
+ "step": 591
+ },
+ {
+ "epoch": 8.223385689354275,
+ "grad_norm": 0.04414986073970795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.786281585693359,
+ "step": 592
+ },
+ {
+ "epoch": 8.237347294938917,
+ "grad_norm": 0.044150348752737045,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.777070045471191,
+ "step": 593
+ },
+ {
+ "epoch": 8.25130890052356,
+ "grad_norm": 0.037002332508563995,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.77547550201416,
+ "step": 594
+ },
+ {
+ "epoch": 8.265270506108202,
+ "grad_norm": 0.03417367860674858,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.782101154327393,
+ "step": 595
+ },
+ {
+ "epoch": 8.279232111692846,
+ "grad_norm": 0.03069210797548294,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.745595455169678,
+ "step": 596
+ },
+ {
+ "epoch": 8.293193717277488,
+ "grad_norm": 0.03228718787431717,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.7663655281066895,
+ "step": 597
+ },
+ {
+ "epoch": 8.30715532286213,
+ "grad_norm": 0.028603695333003998,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.767274856567383,
+ "step": 598
+ },
+ {
+ "epoch": 8.321116928446772,
+ "grad_norm": 0.03098377026617527,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.7853684425354,
+ "step": 599
+ },
+ {
+ "epoch": 8.335078534031414,
+ "grad_norm": 0.029242439195513725,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.738029956817627,
+ "step": 600
+ },
+ {
+ "epoch": 8.349040139616056,
+ "grad_norm": 0.02751081995666027,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.76801872253418,
+ "step": 601
+ },
+ {
+ "epoch": 8.363001745200698,
+ "grad_norm": 0.02838125079870224,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.743884086608887,
+ "step": 602
+ },
+ {
+ "epoch": 8.37696335078534,
+ "grad_norm": 0.029203062877058983,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.746668815612793,
+ "step": 603
+ },
+ {
+ "epoch": 8.390924956369982,
+ "grad_norm": 0.02679944597184658,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.760764122009277,
+ "step": 604
+ },
+ {
+ "epoch": 8.404886561954624,
+ "grad_norm": 0.028955848887562752,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.7353668212890625,
+ "step": 605
+ },
+ {
+ "epoch": 8.418848167539267,
+ "grad_norm": 0.02880851738154888,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.766837120056152,
+ "step": 606
+ },
+ {
+ "epoch": 8.432809773123909,
+ "grad_norm": 0.02495497092604637,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.78959321975708,
+ "step": 607
+ },
+ {
+ "epoch": 8.44677137870855,
+ "grad_norm": 0.026064753532409668,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.736433029174805,
+ "step": 608
+ },
+ {
+ "epoch": 8.460732984293193,
+ "grad_norm": 0.027566388249397278,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.782382965087891,
+ "step": 609
+ },
+ {
+ "epoch": 8.474694589877837,
+ "grad_norm": 0.02967359498143196,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.735286712646484,
+ "step": 610
+ },
+ {
+ "epoch": 8.488656195462479,
+ "grad_norm": 0.03079676441848278,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.725196838378906,
+ "step": 611
+ },
+ {
+ "epoch": 8.502617801047121,
+ "grad_norm": 0.032531846314668655,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.719333648681641,
+ "step": 612
+ },
+ {
+ "epoch": 8.516579406631763,
+ "grad_norm": 0.03120352514088154,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.723845958709717,
+ "step": 613
+ },
+ {
+ "epoch": 8.530541012216405,
+ "grad_norm": 0.03328424692153931,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.7669782638549805,
+ "step": 614
+ },
+ {
+ "epoch": 8.544502617801047,
+ "grad_norm": 0.035601649433374405,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.714925765991211,
+ "step": 615
+ },
+ {
+ "epoch": 8.55846422338569,
+ "grad_norm": 0.03675796464085579,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.715317726135254,
+ "step": 616
+ },
+ {
+ "epoch": 8.572425828970331,
+ "grad_norm": 0.03701988235116005,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.724606990814209,
+ "step": 617
+ },
+ {
+ "epoch": 8.586387434554974,
+ "grad_norm": 0.036719705909490585,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.746044635772705,
+ "step": 618
+ },
+ {
+ "epoch": 8.600349040139616,
+ "grad_norm": 0.034756604582071304,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.74885368347168,
+ "step": 619
+ },
+ {
+ "epoch": 8.614310645724258,
+ "grad_norm": 0.03237922117114067,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.721858501434326,
+ "step": 620
+ },
+ {
+ "epoch": 8.6282722513089,
+ "grad_norm": 0.028420142829418182,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.733814239501953,
+ "step": 621
+ },
+ {
+ "epoch": 8.642233856893542,
+ "grad_norm": 0.026203038170933723,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.726088523864746,
+ "step": 622
+ },
+ {
+ "epoch": 8.656195462478184,
+ "grad_norm": 0.02627231739461422,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.727700233459473,
+ "step": 623
+ },
+ {
+ "epoch": 8.670157068062828,
+ "grad_norm": 0.02920261025428772,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.731454849243164,
+ "step": 624
+ },
+ {
+ "epoch": 8.68411867364747,
+ "grad_norm": 0.02792907878756523,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.734154224395752,
+ "step": 625
+ },
+ {
+ "epoch": 8.698080279232112,
+ "grad_norm": 0.026770325377583504,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.715337753295898,
+ "step": 626
+ },
+ {
+ "epoch": 8.712041884816754,
+ "grad_norm": 0.027346540242433548,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.716512680053711,
+ "step": 627
+ },
+ {
+ "epoch": 8.726003490401396,
+ "grad_norm": 0.031965773552656174,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.7072038650512695,
+ "step": 628
+ },
+ {
+ "epoch": 8.739965095986038,
+ "grad_norm": 0.03708025813102722,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.749785900115967,
+ "step": 629
+ },
+ {
+ "epoch": 8.75392670157068,
+ "grad_norm": 0.04846490919589996,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.735586166381836,
+ "step": 630
+ },
+ {
+ "epoch": 8.767888307155323,
+ "grad_norm": 0.053352758288383484,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.720524787902832,
+ "step": 631
+ },
+ {
+ "epoch": 8.781849912739965,
+ "grad_norm": 0.040559202432632446,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.719901084899902,
+ "step": 632
+ },
+ {
+ "epoch": 8.795811518324607,
+ "grad_norm": 0.03333032503724098,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.708139419555664,
+ "step": 633
+ },
+ {
+ "epoch": 8.809773123909249,
+ "grad_norm": 0.030532941222190857,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.728184223175049,
+ "step": 634
+ },
+ {
+ "epoch": 8.823734729493891,
+ "grad_norm": 0.03337649628520012,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.726020336151123,
+ "step": 635
+ },
+ {
+ "epoch": 8.837696335078533,
+ "grad_norm": 0.03352045267820358,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.7196855545043945,
+ "step": 636
+ },
+ {
+ "epoch": 8.851657940663177,
+ "grad_norm": 0.03662872314453125,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.732823371887207,
+ "step": 637
+ },
+ {
+ "epoch": 8.86561954624782,
+ "grad_norm": 0.037085745483636856,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.728443145751953,
+ "step": 638
+ },
+ {
+ "epoch": 8.879581151832461,
+ "grad_norm": 0.036635689437389374,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.713988780975342,
+ "step": 639
+ },
+ {
+ "epoch": 8.893542757417103,
+ "grad_norm": 0.03719659522175789,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.715950965881348,
+ "step": 640
+ },
+ {
+ "epoch": 8.907504363001745,
+ "grad_norm": 0.03719159588217735,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.7124924659729,
+ "step": 641
+ },
+ {
+ "epoch": 8.921465968586388,
+ "grad_norm": 0.04275721684098244,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.708404064178467,
+ "step": 642
+ },
+ {
+ "epoch": 8.93542757417103,
+ "grad_norm": 0.04204598814249039,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.715372085571289,
+ "step": 643
+ },
+ {
+ "epoch": 8.949389179755672,
+ "grad_norm": 0.03938259929418564,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.729745864868164,
+ "step": 644
+ },
+ {
+ "epoch": 8.963350785340314,
+ "grad_norm": 0.036226850003004074,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.733749866485596,
+ "step": 645
+ },
+ {
+ "epoch": 8.977312390924956,
+ "grad_norm": 0.03542874753475189,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.706825256347656,
+ "step": 646
+ },
+ {
+ "epoch": 8.991273996509598,
+ "grad_norm": 0.029055381193757057,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.718567848205566,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.01990380510687828,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 2.943183183670044,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.601699709892273,
+ "eval_runtime": 60.0221,
+ "eval_samples_per_second": 40.685,
+ "eval_steps_per_second": 0.65,
+ "step": 648
+ },
+ {
+ "epoch": 9.013961605584642,
+ "grad_norm": 0.030904127284884453,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.660327911376953,
+ "step": 649
+ },
+ {
+ "epoch": 9.027923211169284,
+ "grad_norm": 0.0441313236951828,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.701565742492676,
+ "step": 650
+ },
+ {
+ "epoch": 9.041884816753926,
+ "grad_norm": 0.047547727823257446,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.666440486907959,
+ "step": 651
+ },
+ {
+ "epoch": 9.055846422338568,
+ "grad_norm": 0.04549463838338852,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.660548210144043,
+ "step": 652
+ },
+ {
+ "epoch": 9.06980802792321,
+ "grad_norm": 0.050346389412879944,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.640229225158691,
+ "step": 653
+ },
+ {
+ "epoch": 9.083769633507853,
+ "grad_norm": 0.058505453169345856,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.673881530761719,
+ "step": 654
+ },
+ {
+ "epoch": 9.097731239092496,
+ "grad_norm": 0.051989227533340454,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.693943977355957,
+ "step": 655
+ },
+ {
+ "epoch": 9.111692844677139,
+ "grad_norm": 0.052564945071935654,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.644315242767334,
+ "step": 656
+ },
+ {
+ "epoch": 9.12565445026178,
+ "grad_norm": 0.05647841840982437,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.681768417358398,
+ "step": 657
+ },
+ {
+ "epoch": 9.139616055846423,
+ "grad_norm": 0.058640867471694946,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.676457405090332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153577661431065,
+ "grad_norm": 0.05386025458574295,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.659588813781738,
+ "step": 659
+ },
+ {
+ "epoch": 9.167539267015707,
+ "grad_norm": 0.048699021339416504,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.680031776428223,
+ "step": 660
+ },
+ {
+ "epoch": 9.181500872600349,
+ "grad_norm": 0.04374176263809204,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.672671318054199,
+ "step": 661
+ },
+ {
+ "epoch": 9.195462478184991,
+ "grad_norm": 0.042337436228990555,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.673431873321533,
+ "step": 662
+ },
+ {
+ "epoch": 9.209424083769633,
+ "grad_norm": 0.04564661160111427,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.632595062255859,
+ "step": 663
+ },
+ {
+ "epoch": 9.223385689354275,
+ "grad_norm": 0.04404835030436516,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.678891181945801,
+ "step": 664
+ },
+ {
+ "epoch": 9.237347294938917,
+ "grad_norm": 0.040642790496349335,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.633418560028076,
+ "step": 665
+ },
+ {
+ "epoch": 9.25130890052356,
+ "grad_norm": 0.036492060869932175,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.670109748840332,
+ "step": 666
+ },
+ {
+ "epoch": 9.265270506108202,
+ "grad_norm": 0.03735879808664322,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.608570098876953,
+ "step": 667
+ },
+ {
+ "epoch": 9.279232111692846,
+ "grad_norm": 0.034563641995191574,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.632505416870117,
+ "step": 668
+ },
+ {
+ "epoch": 9.293193717277488,
+ "grad_norm": 0.03522484749555588,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.6475629806518555,
+ "step": 669
+ },
+ {
+ "epoch": 9.30715532286213,
+ "grad_norm": 0.03216472268104553,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.627873420715332,
+ "step": 670
+ },
+ {
+ "epoch": 9.321116928446772,
+ "grad_norm": 0.029155759140849113,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.652706146240234,
+ "step": 671
+ },
+ {
+ "epoch": 9.335078534031414,
+ "grad_norm": 0.031060541048645973,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.644545078277588,
+ "step": 672
+ },
+ {
+ "epoch": 9.349040139616056,
+ "grad_norm": 0.034979186952114105,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.642645359039307,
+ "step": 673
+ },
+ {
+ "epoch": 9.363001745200698,
+ "grad_norm": 0.035258810967206955,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.647383689880371,
+ "step": 674
+ },
+ {
+ "epoch": 9.37696335078534,
+ "grad_norm": 0.031552575528621674,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.667779445648193,
+ "step": 675
+ },
+ {
+ "epoch": 9.390924956369982,
+ "grad_norm": 0.03193951025605202,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.63625431060791,
+ "step": 676
+ },
+ {
+ "epoch": 9.404886561954624,
+ "grad_norm": 0.029811711981892586,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.658000946044922,
+ "step": 677
+ },
+ {
+ "epoch": 9.418848167539267,
+ "grad_norm": 0.028464138507843018,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.6545515060424805,
+ "step": 678
+ },
+ {
+ "epoch": 9.432809773123909,
+ "grad_norm": 0.0315735749900341,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.641153335571289,
+ "step": 679
+ },
+ {
+ "epoch": 9.44677137870855,
+ "grad_norm": 0.033412106335163116,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.626931667327881,
+ "step": 680
+ },
+ {
+ "epoch": 9.460732984293193,
+ "grad_norm": 0.03757626190781593,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.671254634857178,
+ "step": 681
+ },
+ {
+ "epoch": 9.474694589877837,
+ "grad_norm": 0.03910793736577034,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.651954174041748,
+ "step": 682
+ },
+ {
+ "epoch": 9.488656195462479,
+ "grad_norm": 0.04015549272298813,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.613324165344238,
+ "step": 683
+ },
+ {
+ "epoch": 9.502617801047121,
+ "grad_norm": 0.040689144283533096,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.639243125915527,
+ "step": 684
+ },
+ {
+ "epoch": 9.516579406631763,
+ "grad_norm": 0.04075361415743828,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.649250030517578,
+ "step": 685
+ },
+ {
+ "epoch": 9.530541012216405,
+ "grad_norm": 0.03914607688784599,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.66444730758667,
+ "step": 686
+ },
+ {
+ "epoch": 9.544502617801047,
+ "grad_norm": 0.033253733068704605,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.647335052490234,
+ "step": 687
+ },
+ {
+ "epoch": 9.55846422338569,
+ "grad_norm": 0.03179008886218071,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.6727800369262695,
+ "step": 688
+ },
+ {
+ "epoch": 9.572425828970331,
+ "grad_norm": 0.03477025404572487,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.655778884887695,
+ "step": 689
+ },
+ {
+ "epoch": 9.586387434554974,
+ "grad_norm": 0.030094800516963005,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.680851936340332,
+ "step": 690
+ },
+ {
+ "epoch": 9.600349040139616,
+ "grad_norm": 0.030153103172779083,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.654638290405273,
+ "step": 691
+ },
+ {
+ "epoch": 9.614310645724258,
+ "grad_norm": 0.029009278863668442,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.625916481018066,
+ "step": 692
+ },
+ {
+ "epoch": 9.6282722513089,
+ "grad_norm": 0.027248440310359,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.656370162963867,
+ "step": 693
+ },
+ {
+ "epoch": 9.642233856893542,
+ "grad_norm": 0.027727462351322174,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.631473541259766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656195462478184,
+ "grad_norm": 0.026710307225584984,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.636524200439453,
+ "step": 695
+ },
+ {
+ "epoch": 9.670157068062828,
+ "grad_norm": 0.02726181596517563,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.629701614379883,
+ "step": 696
+ },
+ {
+ "epoch": 9.68411867364747,
+ "grad_norm": 0.027146652340888977,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.640240669250488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698080279232112,
+ "grad_norm": 0.027165887877345085,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.64462947845459,
+ "step": 698
+ },
+ {
+ "epoch": 9.712041884816754,
+ "grad_norm": 0.03025982715189457,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.6271820068359375,
+ "step": 699
+ },
+ {
+ "epoch": 9.726003490401396,
+ "grad_norm": 0.02837025746703148,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.629697799682617,
+ "step": 700
+ },
+ {
+ "epoch": 9.739965095986038,
+ "grad_norm": 0.026777995750308037,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.646146774291992,
+ "step": 701
+ },
+ {
+ "epoch": 9.75392670157068,
+ "grad_norm": 0.024626802653074265,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.641671657562256,
+ "step": 702
+ },
+ {
+ "epoch": 9.767888307155323,
+ "grad_norm": 0.025329895317554474,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.626345157623291,
+ "step": 703
+ },
+ {
+ "epoch": 9.781849912739965,
+ "grad_norm": 0.027804844081401825,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.621457099914551,
+ "step": 704
+ },
+ {
+ "epoch": 9.795811518324607,
+ "grad_norm": 0.028763437643647194,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.6124114990234375,
+ "step": 705
+ },
+ {
+ "epoch": 9.809773123909249,
+ "grad_norm": 0.030769916251301765,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.604399681091309,
+ "step": 706
+ },
+ {
+ "epoch": 9.823734729493891,
+ "grad_norm": 0.03060949593782425,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.640804290771484,
+ "step": 707
+ },
+ {
+ "epoch": 9.837696335078533,
+ "grad_norm": 0.029832778498530388,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.593376159667969,
+ "step": 708
+ },
+ {
+ "epoch": 9.851657940663177,
+ "grad_norm": 0.02952776849269867,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.619533538818359,
+ "step": 709
+ },
+ {
+ "epoch": 9.86561954624782,
+ "grad_norm": 0.02894212119281292,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.640516757965088,
+ "step": 710
+ },
+ {
+ "epoch": 9.879581151832461,
+ "grad_norm": 0.026435624808073044,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.6342997550964355,
+ "step": 711
+ },
+ {
+ "epoch": 9.893542757417103,
+ "grad_norm": 0.02591666206717491,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.630884170532227,
+ "step": 712
+ },
+ {
+ "epoch": 9.907504363001745,
+ "grad_norm": 0.02702309936285019,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.632085800170898,
+ "step": 713
+ },
+ {
+ "epoch": 9.921465968586388,
+ "grad_norm": 0.026766657829284668,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.6231231689453125,
+ "step": 714
+ },
+ {
+ "epoch": 9.93542757417103,
+ "grad_norm": 0.026666559278964996,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.601619243621826,
+ "step": 715
+ },
+ {
+ "epoch": 9.949389179755672,
+ "grad_norm": 0.026801064610481262,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.641724109649658,
+ "step": 716
+ },
+ {
+ "epoch": 9.963350785340314,
+ "grad_norm": 0.03152798116207123,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.642511367797852,
+ "step": 717
+ },
+ {
+ "epoch": 9.977312390924956,
+ "grad_norm": 0.035406943410634995,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.635779857635498,
+ "step": 718
+ },
+ {
+ "epoch": 9.991273996509598,
+ "grad_norm": 0.037504684180021286,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.629463195800781,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02263174019753933,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 2.876767635345459,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.5942904949188232,
+ "eval_runtime": 60.2086,
+ "eval_samples_per_second": 40.559,
+ "eval_steps_per_second": 0.648,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03707852587103844,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.5765533447265625,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.05454680323600769,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.576363563537598,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.04832114651799202,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.586915969848633,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.050469402223825455,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.5265092849731445,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0622565895318985,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.592833518981934,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0640949010848999,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.5855560302734375,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.06502196192741394,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.591236114501953,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.06167149916291237,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.56747579574585,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.06747374683618546,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.60452938079834,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.07210271060466766,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.561588764190674,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.0733448714017868,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.610766410827637,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.0731528028845787,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.589570045471191,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.0822535902261734,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.631077289581299,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.06921929866075516,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.628293991088867,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.07443586736917496,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.627321243286133,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.07153813540935516,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.641068935394287,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.069196417927742,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.615875720977783,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06483697891235352,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.6165313720703125,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.05482150986790657,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.599890232086182,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.05364019051194191,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.61989164352417,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.053216829895973206,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.564305305480957,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05740506947040558,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.582683563232422,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.05587657541036606,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.61264705657959,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.050252825021743774,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.648050308227539,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.050955042243003845,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.615367889404297,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04903746768832207,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.617378234863281,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04252410680055618,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.609130859375,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.03801053389906883,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.625149250030518,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.03422372043132782,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.606106281280518,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.030507879331707954,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.6125993728637695,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.0340375155210495,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.5690202713012695,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.030507374554872513,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.6042985916137695,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027383938431739807,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.585570812225342,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.028792303055524826,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.566786766052246,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02866457775235176,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.566376686096191,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.028657952323555946,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.581750869750977,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.028227943927049637,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.576788902282715,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.024757765233516693,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.59373664855957,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.027442635968327522,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.570978164672852,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.02590034157037735,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.571286201477051,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.025696570053696632,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.542209625244141,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.025478003546595573,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.5900421142578125,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.02563696913421154,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.544609546661377,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.025326019152998924,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.5808515548706055,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.024502627551555634,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.560579776763916,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.02687518671154976,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.567154884338379,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.025280235335230827,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.560011863708496,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.026266315951943398,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.5758442878723145,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.026058891788125038,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.583309173583984,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.02620471641421318,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.547737121582031,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02626362442970276,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.556955337524414,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.024503158405423164,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.580735206604004,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02580532431602478,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.564000129699707,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.024437161162495613,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.545594215393066,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.026664750650525093,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.56415319442749,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.02499818615615368,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.550689697265625,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.027514580637216568,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.5548248291015625,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.026799403131008148,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.548295021057129,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.02685476280748844,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.5660600662231445,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.026104001328349113,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.556082248687744,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.027479661628603935,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.5580244064331055,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.02638247236609459,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.550073623657227,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.025853842496871948,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.578088760375977,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.02636682614684105,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.58807373046875,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.02684398740530014,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.561017036437988,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.028538964688777924,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.548762321472168,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.028327809646725655,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.548969745635986,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.025131452828645706,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.564047813415527,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.026498543098568916,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.553997039794922,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.030081963166594505,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.5787787437438965,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.03201213851571083,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.566424369812012,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.020518779754638672,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8421616554260254,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.5899484157562256,
+ "eval_runtime": 60.2671,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03692612797021866,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.492715835571289,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.050258126109838486,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.480323791503906,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.04179152473807335,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.484371185302734,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.04160204157233238,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.4905524253845215,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.03974860534071922,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.503803253173828,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.037031810730695724,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.494629859924316,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.038762833923101425,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.495111465454102,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.0412563718855381,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.490113735198975,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.041371624916791916,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.490643501281738,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04151557758450508,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.49881649017334,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04037577286362648,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.492178916931152,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.03676290065050125,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.477460861206055,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.038055796176195145,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.509150981903076,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.041152238845825195,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.47133731842041,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.03728092834353447,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.508072376251221,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.035338856279850006,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.505619049072266,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.037813492119312286,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.509489059448242,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.03789111226797104,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.508646488189697,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.03880828619003296,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.52567720413208,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.03611127287149429,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.498889446258545,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.036903638392686844,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.498601913452148,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03620576485991478,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.490840911865234,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03555903211236,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.506166458129883,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.03441924974322319,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.498001575469971,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.0334339402616024,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.491082191467285,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.031418606638908386,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.527408599853516,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03073921985924244,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.514107704162598,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.029589885845780373,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.48323917388916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.02974501997232437,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.51338529586792,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.029247552156448364,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.504205226898193,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.03132032975554466,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.505674362182617,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03088037669658661,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.477198600769043,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.031054604798555374,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.498721122741699,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03293661028146744,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.527826309204102,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.03181406110525131,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.495999336242676,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.029380083084106445,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.530351638793945,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.029296347871422768,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.492367744445801,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.02958036959171295,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.493037700653076,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03078395500779152,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.47504186630249,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03280661255121231,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.518409729003906,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.03323954716324806,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.460753917694092,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.032254140824079514,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.474285125732422,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.03018898330628872,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.5162353515625,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.02802245132625103,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.538684844970703,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.02821732498705387,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.4850287437438965,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.029189350083470345,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.50770378112793,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.02851196378469467,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.499988555908203,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.02741030789911747,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.499852180480957,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.028708625584840775,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.465211868286133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.028519680723547935,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.502464294433594,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.027846841141581535,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.503248691558838,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030456233769655228,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.490391731262207,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029952052980661392,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.50056266784668,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.02925778552889824,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.49215030670166,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.029679695144295692,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.523841857910156,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.031178541481494904,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.499871730804443,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.031264498829841614,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.50546932220459,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03482109308242798,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.494480133056641,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.03341105580329895,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.532899856567383,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028623664751648903,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.5016279220581055,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028600146993994713,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.511589050292969,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.029127901419997215,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.502058982849121,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.027898237109184265,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.493182182312012,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.028683282434940338,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.519793510437012,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.02887023612856865,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.457289695739746,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.02811296097934246,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.49398946762085,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.028485111892223358,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.478344440460205,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02914245054125786,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.487666130065918,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.028030497953295708,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.518488883972168,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.027973609045147896,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.515271186828613,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.027458522468805313,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.490856647491455,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.01909293793141842,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.822120189666748,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5866912603378296,
+ "eval_runtime": 61.3185,
+ "eval_samples_per_second": 39.825,
+ "eval_steps_per_second": 0.636,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03683992475271225,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.438363075256348,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.05042923241853714,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.443560600280762,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.055914171040058136,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.422282695770264,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.057007886469364166,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.442644119262695,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.05999710038304329,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.44780969619751,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06472347676753998,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.485074996948242,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.06923902779817581,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.457515716552734,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.0710320770740509,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.465476989746094,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.06942574679851532,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.426226615905762,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.06867868453264236,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.438092231750488,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.0688796266913414,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.446664810180664,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.0690719485282898,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.456482887268066,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.06068912148475647,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.4730730056762695,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.05411486327648163,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.4596452713012695,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.056623928248882294,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.418197154998779,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.05634097009897232,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.482892036437988,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.05056057870388031,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.457589149475098,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.04460986703634262,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.451439380645752,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.045022401958703995,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.452855110168457,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.04054102674126625,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.417323112487793,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.03682882338762283,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.465713977813721,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.03528819605708122,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.4502339363098145,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.03568802773952484,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.422399520874023,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.0363088883459568,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.447678089141846,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.03460463136434555,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.437664985656738,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.034751344472169876,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.489151954650879,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.033658601343631744,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.4471635818481445,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.03433239087462425,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.455332279205322,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.032074104994535446,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.435157775878906,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.032168325036764145,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.435118675231934,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.03007678873836994,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.42781925201416,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03094375506043434,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.433105945587158,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.032186444848775864,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.383000373840332,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.030610304325819016,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.473257064819336,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.030257435515522957,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.453041076660156,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.0301323514431715,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.396578788757324,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.030934512615203857,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.445223808288574,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.029907869175076485,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.442287445068359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03163841366767883,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.434650421142578,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.029627414420247078,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.454374313354492,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.031026875600218773,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.413784027099609,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03099491260945797,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.458438873291016,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.02981731668114662,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.438560962677002,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.02927570603787899,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.44493293762207,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.029779717326164246,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.445121765136719,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.02909068390727043,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.4736127853393555,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.029217200353741646,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.434604644775391,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.029872911050915718,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.454591751098633,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.028742222115397453,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.468222141265869,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.030313383787870407,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.45759391784668,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.03014662116765976,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.446331977844238,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029481127858161926,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.478209495544434,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.028485266491770744,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.444445610046387,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.02857700176537037,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.430604934692383,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.029038602486252785,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.43759298324585,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.029576119035482407,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.474579811096191,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.02897879108786583,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.440483093261719,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.029799485579133034,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.4839277267456055,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.030710609629750252,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.466696739196777,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.030270954594016075,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.455197334289551,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.03170182928442955,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.454664707183838,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030684730038046837,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.437619209289551,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.030972573906183243,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.431796073913574,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028914963826537132,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.460618019104004,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.029810264706611633,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.407312393188477,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.029806939885020256,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.466520309448242,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.03020886518061161,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.427349090576172,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030598826706409454,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.427891731262207,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03192012757062912,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.448113918304443,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.030492989346385002,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.450060844421387,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03188254311680794,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.43007230758667,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02081947773694992,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.782536506652832,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5860040187835693,
+ "eval_runtime": 60.6725,
+ "eval_samples_per_second": 40.249,
+ "eval_steps_per_second": 0.643,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.036721356213092804,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.358441352844238,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.056633755564689636,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.361527919769287,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.06175022944808006,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.3827924728393555,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05159056559205055,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.354856491088867,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.04885871335864067,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.359966278076172,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.04407693073153496,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.364612579345703,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.04891050234436989,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.344021320343018,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.04643657058477402,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.366066932678223,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.05241481214761734,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.3390607833862305,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.05837438255548477,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.390261650085449,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.058482322841882706,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.3856201171875,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.06507553160190582,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.414223670959473,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.0575365275144577,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.402576446533203,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05055592209100723,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.389563083648682,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.047167956829071045,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.402750015258789,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.04473405331373215,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.358315467834473,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.042440690100193024,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.392551898956299,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04465719312429428,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.373558044433594,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.04337528720498085,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.37944221496582,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.04322915896773338,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.38493013381958,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.039873406291007996,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.389793395996094,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.0413178876042366,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.427964210510254,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.04567280039191246,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.404172897338867,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04154926910996437,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.391359329223633,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.03774876892566681,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.381457328796387,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03588850423693657,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.392526626586914,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.03757261857390404,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.39273738861084,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03849692642688751,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.409083366394043,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03787294402718544,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.3776044845581055,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.034496866166591644,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.413462162017822,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.03973205015063286,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.359318733215332,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.036192040890455246,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.392844200134277,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.03436632081866264,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.421384811401367,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.035658642649650574,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.386757850646973,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.03434896469116211,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.364978313446045,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03642290085554123,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.385326385498047,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.03427736461162567,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.406352996826172,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.034845270216464996,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.370953559875488,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03390975296497345,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.3799943923950195,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.03134961053729057,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.419156551361084,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.03512260690331459,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.404660701751709,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03319067507982254,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.388535022735596,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03113507106900215,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.403522968292236,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03559132665395737,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.382323265075684,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.03579750657081604,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.37620210647583,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.033675212413072586,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.411443710327148,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.03220067173242569,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.443470001220703,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031610481441020966,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.406038284301758,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.02958764135837555,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.41604471206665,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.0318237729370594,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4253387451171875,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03260074555873871,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.370792388916016,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.03359333053231239,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.397639274597168,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032516930252313614,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.4140777587890625,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.03203396871685982,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.408872604370117,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.03201935440301895,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.402975082397461,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.03330211713910103,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.394018173217773,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.0348181314766407,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.380587100982666,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.03174913674592972,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.372378349304199,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.029864462092518806,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.42018985748291,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.033312153071165085,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.385973930358887,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03203144669532776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.395167827606201,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.031578030437231064,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.3982062339782715,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.030885562300682068,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.400731086730957,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03235667943954468,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.389830589294434,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.032664328813552856,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.401277542114258,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.030354196205735207,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.422520637512207,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.031162403523921967,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.393032073974609,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.032187070697546005,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.464164733886719,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.03445067256689072,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.368991851806641,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03398612141609192,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.418752670288086,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.029730670154094696,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.41429328918457,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022440090775489807,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7315595149993896,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5853753089904785,
+ "eval_runtime": 60.7174,
+ "eval_samples_per_second": 40.219,
+ "eval_steps_per_second": 0.642,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.03754177689552307,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.304028511047363,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.051616180688142776,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.34659481048584,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055258627980947495,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.310445785522461,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.049489911645650864,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3416619300842285,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05377533286809921,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.3143310546875,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.05443088337779045,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.320600986480713,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.050222691148519516,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.334789752960205,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.051904670894145966,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.316756248474121,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.05376426503062248,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.274116516113281,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.053313180804252625,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.321177005767822,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.058234598487615585,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.3169264793396,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05730101093649864,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.307013511657715,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.058140840381383896,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.332961082458496,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05654815956950188,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.315462112426758,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.05129384249448776,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.338476181030273,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.047652702778577805,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.321232795715332,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.050912003964185715,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.318971633911133,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.0514022521674633,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.3334150314331055,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.05457981303334236,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.325802803039551,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04678885266184807,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.307281494140625,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.05048161372542381,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.299831390380859,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.05334261804819107,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.313231468200684,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.05055771768093109,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.36859130859375,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04955561086535454,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.301530361175537,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.04548008739948273,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.339517593383789,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.046585116535425186,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.333036422729492,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.0503716766834259,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.364710807800293,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.040569406002759933,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.369089126586914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.03754054754972458,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.3349432945251465,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03775954991579056,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.333134174346924,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.04066266492009163,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.354358673095703,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.040490977466106415,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.360981464385986,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.0426657609641552,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.357726097106934,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.0415605828166008,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.363457679748535,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.04020509123802185,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.372447967529297,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.04101516678929329,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.292452812194824,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.03546427935361862,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.33892297744751,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.03661105036735535,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.329778671264648,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03565476834774017,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.351012229919434,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.03591311350464821,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.376980781555176,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.03665630891919136,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.354767799377441,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.0371236577630043,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.374916076660156,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.03464849665760994,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.3713059425354,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03345152735710144,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.350670337677002,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.03357568010687828,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.3573713302612305,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03503486514091492,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.362038612365723,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.03455005958676338,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.3263773918151855,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.03173859417438507,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.351345062255859,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03301836550235748,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.398364067077637,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.033733855932950974,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.335268020629883,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034080155193805695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.371492385864258,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.034396376460790634,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.382076740264893,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.03303401917219162,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.36623477935791,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03297369182109833,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3148393630981445,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.034819211810827255,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.358110427856445,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.033389780670404434,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.367427825927734,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.032363444566726685,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.3813676834106445,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03509577363729477,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.386520862579346,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.034685730934143066,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.33738374710083,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.034755002707242966,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.360980987548828,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03461688384413719,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.396650314331055,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03392560034990311,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.356130123138428,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.03342147544026375,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.359145164489746,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.032852694392204285,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.3796162605285645,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03472859039902687,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.361110687255859,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03277098760008812,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.361811637878418,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.03341882303357124,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.369524955749512,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.032347049564123154,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.369096755981445,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03161193057894707,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.339844703674316,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.03322301432490349,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.376675605773926,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03419092297554016,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.346864223480225,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02461826801300049,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7236275672912598,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.5859372019767761,
+ "eval_runtime": 59.0139,
+ "eval_samples_per_second": 41.38,
+ "eval_steps_per_second": 0.661,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013961605584642,
+ "grad_norm": 0.039559684693813324,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 4.289841651916504,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027923211169284,
+ "grad_norm": 0.05554218962788582,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 4.255925178527832,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041884816753926,
+ "grad_norm": 0.05330996215343475,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 4.272125244140625,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055846422338568,
+ "grad_norm": 0.053247954696416855,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 4.290765285491943,
+ "step": 1084
+ },
+ {
+ "epoch": 15.06980802792321,
+ "grad_norm": 0.05471150949597359,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 4.271744728088379,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083769633507853,
+ "grad_norm": 0.05976719409227371,
+ "learning_rate": 0.000491285979372924,
+ "loss": 4.28203821182251,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097731239092496,
+ "grad_norm": 0.05369880795478821,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 4.275313377380371,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111692844677139,
+ "grad_norm": 0.052529267966747284,
+ "learning_rate": 0.000490873338088198,
+ "loss": 4.25678825378418,
+ "step": 1088
+ },
+ {
+ "epoch": 15.12565445026178,
+ "grad_norm": 0.05184522643685341,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 4.260510444641113,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139616055846423,
+ "grad_norm": 0.051606856286525726,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 4.283233642578125,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153577661431065,
+ "grad_norm": 0.05318903923034668,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 4.286595344543457,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167539267015707,
+ "grad_norm": 0.058525051921606064,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 4.282374382019043,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181500872600349,
+ "grad_norm": 0.054776228964328766,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 4.2913126945495605,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195462478184991,
+ "grad_norm": 0.04907633364200592,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 4.297187805175781,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209424083769633,
+ "grad_norm": 0.05163136497139931,
+ "learning_rate": 0.000489424334303338,
+ "loss": 4.2968645095825195,
+ "step": 1095
+ },
+ {
+ "epoch": 15.223385689354275,
+ "grad_norm": 0.053157489746809006,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 4.285839080810547,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237347294938917,
+ "grad_norm": 0.05251535773277283,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 4.311246871948242,
+ "step": 1097
+ },
+ {
+ "epoch": 15.25130890052356,
+ "grad_norm": 0.048785097897052765,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 4.300364971160889,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265270506108202,
+ "grad_norm": 0.050393108278512955,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 4.299890995025635,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279232111692846,
+ "grad_norm": 0.04894037917256355,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 4.312449932098389,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293193717277488,
+ "grad_norm": 0.05274952948093414,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 4.2889814376831055,
+ "step": 1101
+ },
+ {
+ "epoch": 15.30715532286213,
+ "grad_norm": 0.05946570262312889,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 4.2731170654296875,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321116928446772,
+ "grad_norm": 0.054245397448539734,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 4.272463798522949,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335078534031414,
+ "grad_norm": 0.042092181742191315,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 4.299953937530518,
+ "step": 1104
+ },
+ {
+ "epoch": 15.349040139616056,
+ "grad_norm": 0.040222618728876114,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 4.292524814605713,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363001745200698,
+ "grad_norm": 0.04298696666955948,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 4.31342887878418,
+ "step": 1106
+ },
+ {
+ "epoch": 15.37696335078534,
+ "grad_norm": 0.04271050915122032,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 4.277663707733154,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390924956369982,
+ "grad_norm": 0.042364370077848434,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 4.311102390289307,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404886561954624,
+ "grad_norm": 0.04343319311738014,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 4.311610221862793,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418848167539267,
+ "grad_norm": 0.041321512311697006,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 4.29677677154541,
+ "step": 1110
+ },
+ {
+ "epoch": 15.432809773123909,
+ "grad_norm": 0.04107942432165146,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 4.3163018226623535,
+ "step": 1111
+ },
+ {
+ "epoch": 15.44677137870855,
+ "grad_norm": 0.03844515234231949,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 4.311301231384277,
+ "step": 1112
+ },
+ {
+ "epoch": 15.460732984293193,
+ "grad_norm": 0.03731007128953934,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 4.31325626373291,
+ "step": 1113
+ },
+ {
+ "epoch": 15.474694589877837,
+ "grad_norm": 0.03807826340198517,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 4.296201705932617,
+ "step": 1114
+ },
+ {
+ "epoch": 15.488656195462479,
+ "grad_norm": 0.040137600153684616,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 4.320395469665527,
+ "step": 1115
+ },
+ {
+ "epoch": 15.502617801047121,
+ "grad_norm": 0.038484394550323486,
+ "learning_rate": 0.000485033394934934,
+ "loss": 4.313858985900879,
+ "step": 1116
+ },
+ {
+ "epoch": 15.516579406631763,
+ "grad_norm": 0.03603338822722435,
+ "learning_rate": 0.000484822676912974,
+ "loss": 4.297666549682617,
+ "step": 1117
+ },
+ {
+ "epoch": 15.530541012216405,
+ "grad_norm": 0.038647182285785675,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 4.305002689361572,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544502617801047,
+ "grad_norm": 0.03907139599323273,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 4.329583644866943,
+ "step": 1119
+ },
+ {
+ "epoch": 15.55846422338569,
+ "grad_norm": 0.03621622920036316,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 4.290896415710449,
+ "step": 1120
+ },
+ {
+ "epoch": 15.572425828970331,
+ "grad_norm": 0.037496887147426605,
+ "learning_rate": 0.000483978341612154,
+ "loss": 4.33336877822876,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586387434554974,
+ "grad_norm": 0.03778311610221863,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 4.285354137420654,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600349040139616,
+ "grad_norm": 0.03782997280359268,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 4.29697847366333,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614310645724258,
+ "grad_norm": 0.03665657714009285,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 4.320103168487549,
+ "step": 1124
+ },
+ {
+ "epoch": 15.6282722513089,
+ "grad_norm": 0.03796127438545227,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 4.317153453826904,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642233856893542,
+ "grad_norm": 0.036425214260816574,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 4.293642520904541,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656195462478184,
+ "grad_norm": 0.03629784658551216,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 4.291100025177002,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670157068062828,
+ "grad_norm": 0.03674762696027756,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 4.295144557952881,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68411867364747,
+ "grad_norm": 0.036291178315877914,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 4.313133239746094,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698080279232112,
+ "grad_norm": 0.03817860782146454,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 4.294533729553223,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712041884816754,
+ "grad_norm": 0.036819685250520706,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 4.299263954162598,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726003490401396,
+ "grad_norm": 0.035103823989629745,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 4.287391185760498,
+ "step": 1132
+ },
+ {
+ "epoch": 15.739965095986038,
+ "grad_norm": 0.035701096057891846,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 4.318216323852539,
+ "step": 1133
+ },
+ {
+ "epoch": 15.75392670157068,
+ "grad_norm": 0.03639856353402138,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 4.301465034484863,
+ "step": 1134
+ },
+ {
+ "epoch": 15.767888307155323,
+ "grad_norm": 0.03593013063073158,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 4.311532974243164,
+ "step": 1135
+ },
+ {
+ "epoch": 15.781849912739965,
+ "grad_norm": 0.03564770147204399,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 4.3400726318359375,
+ "step": 1136
+ },
+ {
+ "epoch": 15.795811518324607,
+ "grad_norm": 0.03598494827747345,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 4.30976676940918,
+ "step": 1137
+ },
+ {
+ "epoch": 15.809773123909249,
+ "grad_norm": 0.034391652792692184,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 4.328464508056641,
+ "step": 1138
+ },
+ {
+ "epoch": 15.823734729493891,
+ "grad_norm": 0.03445672616362572,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 4.310579299926758,
+ "step": 1139
+ },
+ {
+ "epoch": 15.837696335078533,
+ "grad_norm": 0.03643282502889633,
+ "learning_rate": 0.000479936075927473,
+ "loss": 4.320905685424805,
+ "step": 1140
+ },
+ {
+ "epoch": 15.851657940663177,
+ "grad_norm": 0.03489850088953972,
+ "learning_rate": 0.000479721889242305,
+ "loss": 4.317229270935059,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86561954624782,
+ "grad_norm": 0.03609171137213707,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 4.309203147888184,
+ "step": 1142
+ },
+ {
+ "epoch": 15.879581151832461,
+ "grad_norm": 0.03705338388681412,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 4.287387371063232,
+ "step": 1143
+ },
+ {
+ "epoch": 15.893542757417103,
+ "grad_norm": 0.03697717934846878,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 4.302860260009766,
+ "step": 1144
+ },
+ {
+ "epoch": 15.907504363001745,
+ "grad_norm": 0.035338208079338074,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 4.3152265548706055,
+ "step": 1145
+ },
+ {
+ "epoch": 15.921465968586388,
+ "grad_norm": 0.03548634052276611,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 4.327335357666016,
+ "step": 1146
+ },
+ {
+ "epoch": 15.93542757417103,
+ "grad_norm": 0.035512328147888184,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 4.3067803382873535,
+ "step": 1147
+ },
+ {
+ "epoch": 15.949389179755672,
+ "grad_norm": 0.03799289092421532,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 4.320701599121094,
+ "step": 1148
+ },
+ {
+ "epoch": 15.963350785340314,
+ "grad_norm": 0.03866339474916458,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 4.321087837219238,
+ "step": 1149
+ },
+ {
+ "epoch": 15.977312390924956,
+ "grad_norm": 0.036002084612846375,
+ "learning_rate": 0.000477787822648007,
+ "loss": 4.287045478820801,
+ "step": 1150
+ },
+ {
+ "epoch": 15.991273996509598,
+ "grad_norm": 0.03433441370725632,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 4.2562665939331055,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.02580358274281025,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 2.67783260345459,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.5870338082313538,
+ "eval_runtime": 60.0469,
+ "eval_samples_per_second": 40.668,
+ "eval_steps_per_second": 0.649,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013961605584644,
+ "grad_norm": 0.04105810821056366,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 4.224295616149902,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027923211169284,
+ "grad_norm": 0.049538832157850266,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 4.222184181213379,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041884816753928,
+ "grad_norm": 0.05089569091796875,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 4.2067060470581055,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05584642233857,
+ "grad_norm": 0.04600589722394943,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 4.238527774810791,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069808027923212,
+ "grad_norm": 0.043311409652233124,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 4.242960453033447,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083769633507853,
+ "grad_norm": 0.047897014766931534,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 4.199069023132324,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097731239092496,
+ "grad_norm": 0.05633604899048805,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 4.211665153503418,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111692844677137,
+ "grad_norm": 0.056110311299562454,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 4.20163631439209,
+ "step": 1160
+ },
+ {
+ "epoch": 16.12565445026178,
+ "grad_norm": 0.05426819995045662,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 4.21889591217041,
+ "step": 1161
+ },
+ {
+ "epoch": 16.13961605584642,
+ "grad_norm": 0.05489835515618324,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 4.220406532287598,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153577661431065,
+ "grad_norm": 0.053941354155540466,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 4.231909275054932,
+ "step": 1163
+ },
+ {
+ "epoch": 16.167539267015705,
+ "grad_norm": 0.058548808097839355,
+ "learning_rate": 0.000474756648870413,
+ "loss": 4.237003326416016,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18150087260035,
+ "grad_norm": 0.05370597913861275,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 4.240085124969482,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195462478184993,
+ "grad_norm": 0.05684991553425789,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 4.203842639923096,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209424083769633,
+ "grad_norm": 0.060745254158973694,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 4.247998237609863,
+ "step": 1167
+ },
+ {
+ "epoch": 16.223385689354277,
+ "grad_norm": 0.058296725153923035,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 4.226161003112793,
+ "step": 1168
+ },
+ {
+ "epoch": 16.237347294938917,
+ "grad_norm": 0.05310584232211113,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 4.229870796203613,
+ "step": 1169
+ },
+ {
+ "epoch": 16.25130890052356,
+ "grad_norm": 0.04710390046238899,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 4.240084648132324,
+ "step": 1170
+ },
+ {
+ "epoch": 16.2652705061082,
+ "grad_norm": 0.05003240332007408,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 4.240972518920898,
+ "step": 1171
+ },
+ {
+ "epoch": 16.279232111692846,
+ "grad_norm": 0.05552851781249046,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 4.248024940490723,
+ "step": 1172
+ },
+ {
+ "epoch": 16.293193717277486,
+ "grad_norm": 0.049967192113399506,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 4.279229164123535,
+ "step": 1173
+ },
+ {
+ "epoch": 16.30715532286213,
+ "grad_norm": 0.04883681237697601,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 4.269807815551758,
+ "step": 1174
+ },
+ {
+ "epoch": 16.32111692844677,
+ "grad_norm": 0.04931480810046196,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 4.232419490814209,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335078534031414,
+ "grad_norm": 0.048746511340141296,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 4.267911911010742,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349040139616054,
+ "grad_norm": 0.04678189381957054,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 4.262979507446289,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363001745200698,
+ "grad_norm": 0.048893529921770096,
+ "learning_rate": 0.000471698352726896,
+ "loss": 4.254408359527588,
+ "step": 1178
+ },
+ {
+ "epoch": 16.376963350785342,
+ "grad_norm": 0.045027732849121094,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 4.2445902824401855,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390924956369982,
+ "grad_norm": 0.04324870929121971,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 4.2685136795043945,
+ "step": 1180
+ },
+ {
+ "epoch": 16.404886561954626,
+ "grad_norm": 0.0470522940158844,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 4.237505912780762,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418848167539267,
+ "grad_norm": 0.04896962642669678,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 4.2523345947265625,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43280977312391,
+ "grad_norm": 0.04562787711620331,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 4.2192792892456055,
+ "step": 1183
+ },
+ {
+ "epoch": 16.44677137870855,
+ "grad_norm": 0.04414575919508934,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 4.250238418579102,
+ "step": 1184
+ },
+ {
+ "epoch": 16.460732984293195,
+ "grad_norm": 0.04363945871591568,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 4.256250381469727,
+ "step": 1185
+ },
+ {
+ "epoch": 16.474694589877835,
+ "grad_norm": 0.04159797728061676,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 4.274903297424316,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48865619546248,
+ "grad_norm": 0.04248544201254845,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 4.278222560882568,
+ "step": 1187
+ },
+ {
+ "epoch": 16.50261780104712,
+ "grad_norm": 0.04379890486598015,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 4.267505645751953,
+ "step": 1188
+ },
+ {
+ "epoch": 16.516579406631763,
+ "grad_norm": 0.04319411888718605,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 4.2607269287109375,
+ "step": 1189
+ },
+ {
+ "epoch": 16.530541012216403,
+ "grad_norm": 0.04143810644745827,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 4.267977714538574,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544502617801047,
+ "grad_norm": 0.03914840891957283,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 4.260498523712158,
+ "step": 1191
+ },
+ {
+ "epoch": 16.55846422338569,
+ "grad_norm": 0.04107543081045151,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 4.2550482749938965,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57242582897033,
+ "grad_norm": 0.041316065937280655,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 4.3075456619262695,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586387434554975,
+ "grad_norm": 0.04069342464208603,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 4.234392166137695,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600349040139616,
+ "grad_norm": 0.03816857561469078,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 4.251038551330566,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61431064572426,
+ "grad_norm": 0.04095374792814255,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 4.284790992736816,
+ "step": 1196
+ },
+ {
+ "epoch": 16.6282722513089,
+ "grad_norm": 0.03788160905241966,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 4.255716323852539,
+ "step": 1197
+ },
+ {
+ "epoch": 16.642233856893544,
+ "grad_norm": 0.03784061595797539,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 4.269606590270996,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656195462478184,
+ "grad_norm": 0.04087168723344803,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 4.231165885925293,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670157068062828,
+ "grad_norm": 0.040546853095293045,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 4.288422584533691,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68411867364747,
+ "grad_norm": 0.03777262568473816,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 4.277492523193359,
+ "step": 1201
+ },
+ {
+ "epoch": 16.698080279232112,
+ "grad_norm": 0.039857733994722366,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 4.239383697509766,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712041884816752,
+ "grad_norm": 0.04387792944908142,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 4.300265312194824,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726003490401396,
+ "grad_norm": 0.04214833676815033,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 4.255545616149902,
+ "step": 1204
+ },
+ {
+ "epoch": 16.739965095986037,
+ "grad_norm": 0.04290056228637695,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 4.275530815124512,
+ "step": 1205
+ },
+ {
+ "epoch": 16.75392670157068,
+ "grad_norm": 0.03975554555654526,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 4.282685279846191,
+ "step": 1206
+ },
+ {
+ "epoch": 16.767888307155324,
+ "grad_norm": 0.039464011788368225,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 4.285504341125488,
+ "step": 1207
+ },
+ {
+ "epoch": 16.781849912739965,
+ "grad_norm": 0.039714012295007706,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 4.273477554321289,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79581151832461,
+ "grad_norm": 0.03936111181974411,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 4.304205894470215,
+ "step": 1209
+ },
+ {
+ "epoch": 16.80977312390925,
+ "grad_norm": 0.03782064467668533,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 4.307382583618164,
+ "step": 1210
+ },
+ {
+ "epoch": 16.823734729493893,
+ "grad_norm": 0.040340717881917953,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 4.308620452880859,
+ "step": 1211
+ },
+ {
+ "epoch": 16.837696335078533,
+ "grad_norm": 0.0403442308306694,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 4.248626708984375,
+ "step": 1212
+ },
+ {
+ "epoch": 16.851657940663177,
+ "grad_norm": 0.038351960480213165,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 4.261248588562012,
+ "step": 1213
+ },
+ {
+ "epoch": 16.865619546247817,
+ "grad_norm": 0.03965333476662636,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 4.278291702270508,
+ "step": 1214
+ },
+ {
+ "epoch": 16.87958115183246,
+ "grad_norm": 0.040909092873334885,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 4.241847991943359,
+ "step": 1215
+ },
+ {
+ "epoch": 16.8935427574171,
+ "grad_norm": 0.038684047758579254,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 4.281346321105957,
+ "step": 1216
+ },
+ {
+ "epoch": 16.907504363001745,
+ "grad_norm": 0.03698103874921799,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 4.3085432052612305,
+ "step": 1217
+ },
+ {
+ "epoch": 16.921465968586386,
+ "grad_norm": 0.03908613696694374,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 4.24220085144043,
+ "step": 1218
+ },
+ {
+ "epoch": 16.93542757417103,
+ "grad_norm": 0.03873388469219208,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 4.282486915588379,
+ "step": 1219
+ },
+ {
+ "epoch": 16.949389179755673,
+ "grad_norm": 0.03709883987903595,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 4.279009819030762,
+ "step": 1220
+ },
+ {
+ "epoch": 16.963350785340314,
+ "grad_norm": 0.03937513753771782,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 4.273211479187012,
+ "step": 1221
+ },
+ {
+ "epoch": 16.977312390924958,
+ "grad_norm": 0.038478728383779526,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 4.281154155731201,
+ "step": 1222
+ },
+ {
+ "epoch": 16.991273996509598,
+ "grad_norm": 0.03913680836558342,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 4.269218444824219,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.02641749382019043,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 2.6905035972595215,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.5879416465759277,
+ "eval_runtime": 60.2807,
+ "eval_samples_per_second": 40.511,
+ "eval_steps_per_second": 0.647,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013961605584644,
+ "grad_norm": 0.048571839928627014,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 4.2008538246154785,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027923211169284,
+ "grad_norm": 0.0640900656580925,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 4.180826663970947,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041884816753928,
+ "grad_norm": 0.06458134204149246,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 4.199562072753906,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05584642233857,
+ "grad_norm": 0.0563119500875473,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 4.171557426452637,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069808027923212,
+ "grad_norm": 0.050914082676172256,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 4.183539390563965,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083769633507853,
+ "grad_norm": 0.053884491324424744,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 4.210450172424316,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097731239092496,
+ "grad_norm": 0.05301345884799957,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 4.1553192138671875,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111692844677137,
+ "grad_norm": 0.051138825714588165,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 4.158510208129883,
+ "step": 1232
+ },
+ {
+ "epoch": 17.12565445026178,
+ "grad_norm": 0.05355675891041756,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 4.167689323425293,
+ "step": 1233
+ },
+ {
+ "epoch": 17.13961605584642,
+ "grad_norm": 0.06062275916337967,
+ "learning_rate": 0.000459203665939781,
+ "loss": 4.170838356018066,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153577661431065,
+ "grad_norm": 0.058666616678237915,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 4.187441825866699,
+ "step": 1235
+ },
+ {
+ "epoch": 17.167539267015705,
+ "grad_norm": 0.057622168213129044,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 4.190746784210205,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18150087260035,
+ "grad_norm": 0.069090336561203,
+ "learning_rate": 0.000458522886595554,
+ "loss": 4.224946022033691,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195462478184993,
+ "grad_norm": 0.06552847474813461,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 4.204067230224609,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209424083769633,
+ "grad_norm": 0.06602177768945694,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 4.189286708831787,
+ "step": 1239
+ },
+ {
+ "epoch": 17.223385689354277,
+ "grad_norm": 0.06751416623592377,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 4.208927154541016,
+ "step": 1240
+ },
+ {
+ "epoch": 17.237347294938917,
+ "grad_norm": 0.0706794261932373,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 4.212923049926758,
+ "step": 1241
+ },
+ {
+ "epoch": 17.25130890052356,
+ "grad_norm": 0.05346836522221565,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 4.2221784591674805,
+ "step": 1242
+ },
+ {
+ "epoch": 17.2652705061082,
+ "grad_norm": 0.05395734682679176,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 4.196428298950195,
+ "step": 1243
+ },
+ {
+ "epoch": 17.279232111692846,
+ "grad_norm": 0.05496550351381302,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 4.208031177520752,
+ "step": 1244
+ },
+ {
+ "epoch": 17.293193717277486,
+ "grad_norm": 0.05320895463228226,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 4.232810974121094,
+ "step": 1245
+ },
+ {
+ "epoch": 17.30715532286213,
+ "grad_norm": 0.05064759403467178,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 4.207826614379883,
+ "step": 1246
+ },
+ {
+ "epoch": 17.32111692844677,
+ "grad_norm": 0.04287239536643028,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 4.195648670196533,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335078534031414,
+ "grad_norm": 0.04527301341295242,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 4.19954776763916,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349040139616054,
+ "grad_norm": 0.04421805590391159,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 4.248671054840088,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363001745200698,
+ "grad_norm": 0.04487175494432449,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 4.2072954177856445,
+ "step": 1250
+ },
+ {
+ "epoch": 17.376963350785342,
+ "grad_norm": 0.04415794834494591,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 4.222837448120117,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390924956369982,
+ "grad_norm": 0.04474013298749924,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 4.213104248046875,
+ "step": 1252
+ },
+ {
+ "epoch": 17.404886561954626,
+ "grad_norm": 0.042631037533283234,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 4.209726333618164,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418848167539267,
+ "grad_norm": 0.04424271360039711,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 4.202699661254883,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43280977312391,
+ "grad_norm": 0.04151812940835953,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 4.190732955932617,
+ "step": 1255
+ },
+ {
+ "epoch": 17.44677137870855,
+ "grad_norm": 0.041588131338357925,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 4.222360610961914,
+ "step": 1256
+ },
+ {
+ "epoch": 17.460732984293195,
+ "grad_norm": 0.042378414422273636,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 4.198424339294434,
+ "step": 1257
+ },
+ {
+ "epoch": 17.474694589877835,
+ "grad_norm": 0.041907574981451035,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 4.224804401397705,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48865619546248,
+ "grad_norm": 0.043774355202913284,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 4.2276611328125,
+ "step": 1259
+ },
+ {
+ "epoch": 17.50261780104712,
+ "grad_norm": 0.04238573834300041,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 4.228109359741211,
+ "step": 1260
+ },
+ {
+ "epoch": 17.516579406631763,
+ "grad_norm": 0.045351386070251465,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 4.208807468414307,
+ "step": 1261
+ },
+ {
+ "epoch": 17.530541012216403,
+ "grad_norm": 0.04300729185342789,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 4.257306098937988,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544502617801047,
+ "grad_norm": 0.041442908346652985,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 4.208004951477051,
+ "step": 1263
+ },
+ {
+ "epoch": 17.55846422338569,
+ "grad_norm": 0.04337337985634804,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 4.23393440246582,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57242582897033,
+ "grad_norm": 0.04335462674498558,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 4.228087425231934,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586387434554975,
+ "grad_norm": 0.03987223282456398,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 4.213971138000488,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600349040139616,
+ "grad_norm": 0.041760094463825226,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 4.248366355895996,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61431064572426,
+ "grad_norm": 0.04420867934823036,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 4.278865814208984,
+ "step": 1268
+ },
+ {
+ "epoch": 17.6282722513089,
+ "grad_norm": 0.04244982823729515,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 4.223642349243164,
+ "step": 1269
+ },
+ {
+ "epoch": 17.642233856893544,
+ "grad_norm": 0.04280664771795273,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 4.231023788452148,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656195462478184,
+ "grad_norm": 0.04294133558869362,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 4.222380638122559,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670157068062828,
+ "grad_norm": 0.04179249703884125,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 4.215641975402832,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68411867364747,
+ "grad_norm": 0.04038798063993454,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 4.228580474853516,
+ "step": 1273
+ },
+ {
+ "epoch": 17.698080279232112,
+ "grad_norm": 0.04211528226733208,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 4.245234489440918,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712041884816752,
+ "grad_norm": 0.04102989658713341,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 4.218495845794678,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726003490401396,
+ "grad_norm": 0.04038692265748978,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 4.2466511726379395,
+ "step": 1276
+ },
+ {
+ "epoch": 17.739965095986037,
+ "grad_norm": 0.03958791121840477,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 4.237762451171875,
+ "step": 1277
+ },
+ {
+ "epoch": 17.75392670157068,
+ "grad_norm": 0.04129907861351967,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 4.238739013671875,
+ "step": 1278
+ },
+ {
+ "epoch": 17.767888307155324,
+ "grad_norm": 0.042581796646118164,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 4.235558032989502,
+ "step": 1279
+ },
+ {
+ "epoch": 17.781849912739965,
+ "grad_norm": 0.04339978098869324,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 4.226593971252441,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79581151832461,
+ "grad_norm": 0.03808023780584335,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 4.238213062286377,
+ "step": 1281
+ },
+ {
+ "epoch": 17.80977312390925,
+ "grad_norm": 0.04022229090332985,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 4.214224815368652,
+ "step": 1282
+ },
+ {
+ "epoch": 17.823734729493893,
+ "grad_norm": 0.04379156231880188,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 4.253068447113037,
+ "step": 1283
+ },
+ {
+ "epoch": 17.837696335078533,
+ "grad_norm": 0.04232964664697647,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 4.249903202056885,
+ "step": 1284
+ },
+ {
+ "epoch": 17.851657940663177,
+ "grad_norm": 0.040980905294418335,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 4.252640724182129,
+ "step": 1285
+ },
+ {
+ "epoch": 17.865619546247817,
+ "grad_norm": 0.04015526548027992,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 4.234862327575684,
+ "step": 1286
+ },
+ {
+ "epoch": 17.87958115183246,
+ "grad_norm": 0.04273828864097595,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 4.216028213500977,
+ "step": 1287
+ },
+ {
+ "epoch": 17.8935427574171,
+ "grad_norm": 0.04210563376545906,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 4.23359489440918,
+ "step": 1288
+ },
+ {
+ "epoch": 17.907504363001745,
+ "grad_norm": 0.038766685873270035,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 4.220360279083252,
+ "step": 1289
+ },
+ {
+ "epoch": 17.921465968586386,
+ "grad_norm": 0.03921609744429588,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 4.253805160522461,
+ "step": 1290
+ },
+ {
+ "epoch": 17.93542757417103,
+ "grad_norm": 0.04071544483304024,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 4.211221694946289,
+ "step": 1291
+ },
+ {
+ "epoch": 17.949389179755673,
+ "grad_norm": 0.03996877372264862,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 4.241739273071289,
+ "step": 1292
+ },
+ {
+ "epoch": 17.963350785340314,
+ "grad_norm": 0.03878853842616081,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 4.237001419067383,
+ "step": 1293
+ },
+ {
+ "epoch": 17.977312390924958,
+ "grad_norm": 0.03889666125178337,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 4.244929790496826,
+ "step": 1294
+ },
+ {
+ "epoch": 17.991273996509598,
+ "grad_norm": 0.03876441717147827,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 4.236083030700684,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.027286000549793243,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 2.636739730834961,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.5896694660186768,
+ "eval_runtime": 60.4742,
+ "eval_samples_per_second": 40.381,
+ "eval_steps_per_second": 0.645,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013961605584644,
+ "grad_norm": 0.04279542714357376,
+ "learning_rate": 0.000444677103403854,
+ "loss": 4.136401653289795,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027923211169284,
+ "grad_norm": 0.051685452461242676,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 4.159584045410156,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041884816753928,
+ "grad_norm": 0.05411457270383835,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 4.160367965698242,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05584642233857,
+ "grad_norm": 0.048574138432741165,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 4.134189128875732,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069808027923212,
+ "grad_norm": 0.04528478533029556,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 4.1022725105285645,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083769633507853,
+ "grad_norm": 0.04474090039730072,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 4.16152286529541,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097731239092496,
+ "grad_norm": 0.04882791265845299,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 4.1553144454956055,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111692844677137,
+ "grad_norm": 0.04831654578447342,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 4.147022724151611,
+ "step": 1304
+ },
+ {
+ "epoch": 18.12565445026178,
+ "grad_norm": 0.047365427017211914,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 4.139405250549316,
+ "step": 1305
+ },
+ {
+ "epoch": 18.13961605584642,
+ "grad_norm": 0.0514012910425663,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 4.115568161010742,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153577661431065,
+ "grad_norm": 0.05225161090493202,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 4.166429042816162,
+ "step": 1307
+ },
+ {
+ "epoch": 18.167539267015705,
+ "grad_norm": 0.05575249716639519,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 4.1414289474487305,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18150087260035,
+ "grad_norm": 0.05405383184552193,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 4.138680934906006,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195462478184993,
+ "grad_norm": 0.05185031518340111,
+ "learning_rate": 0.000441621628895411,
+ "loss": 4.159243583679199,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209424083769633,
+ "grad_norm": 0.05070916935801506,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 4.1844892501831055,
+ "step": 1311
+ },
+ {
+ "epoch": 18.223385689354277,
+ "grad_norm": 0.05037331581115723,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 4.132781028747559,
+ "step": 1312
+ },
+ {
+ "epoch": 18.237347294938917,
+ "grad_norm": 0.05082111805677414,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 4.143031120300293,
+ "step": 1313
+ },
+ {
+ "epoch": 18.25130890052356,
+ "grad_norm": 0.04742169380187988,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 4.137580871582031,
+ "step": 1314
+ },
+ {
+ "epoch": 18.2652705061082,
+ "grad_norm": 0.049042731523513794,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 4.159721374511719,
+ "step": 1315
+ },
+ {
+ "epoch": 18.279232111692846,
+ "grad_norm": 0.045649368315935135,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 4.144008636474609,
+ "step": 1316
+ },
+ {
+ "epoch": 18.293193717277486,
+ "grad_norm": 0.0473787859082222,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 4.164243698120117,
+ "step": 1317
+ },
+ {
+ "epoch": 18.30715532286213,
+ "grad_norm": 0.047626055777072906,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 4.164441108703613,
+ "step": 1318
+ },
+ {
+ "epoch": 18.32111692844677,
+ "grad_norm": 0.04738573357462883,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 4.180338382720947,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335078534031414,
+ "grad_norm": 0.04712498560547829,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 4.146285533905029,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349040139616054,
+ "grad_norm": 0.045505691319704056,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 4.162572860717773,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363001745200698,
+ "grad_norm": 0.0489528588950634,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 4.160312175750732,
+ "step": 1322
+ },
+ {
+ "epoch": 18.376963350785342,
+ "grad_norm": 0.047517575323581696,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 4.164361953735352,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390924956369982,
+ "grad_norm": 0.044214535504579544,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 4.182154655456543,
+ "step": 1324
+ },
+ {
+ "epoch": 18.404886561954626,
+ "grad_norm": 0.043775733560323715,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 4.130522727966309,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418848167539267,
+ "grad_norm": 0.044364530593156815,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 4.151247024536133,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43280977312391,
+ "grad_norm": 0.046320103108882904,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 4.204159736633301,
+ "step": 1327
+ },
+ {
+ "epoch": 18.44677137870855,
+ "grad_norm": 0.04802565276622772,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 4.182550430297852,
+ "step": 1328
+ },
+ {
+ "epoch": 18.460732984293195,
+ "grad_norm": 0.049340926110744476,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 4.163658142089844,
+ "step": 1329
+ },
+ {
+ "epoch": 18.474694589877835,
+ "grad_norm": 0.046701837331056595,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 4.170445442199707,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48865619546248,
+ "grad_norm": 0.04746273159980774,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 4.220232963562012,
+ "step": 1331
+ },
+ {
+ "epoch": 18.50261780104712,
+ "grad_norm": 0.04644489288330078,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 4.190027236938477,
+ "step": 1332
+ },
+ {
+ "epoch": 18.516579406631763,
+ "grad_norm": 0.04394262284040451,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 4.159570693969727,
+ "step": 1333
+ },
+ {
+ "epoch": 18.530541012216403,
+ "grad_norm": 0.04609063267707825,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 4.1657586097717285,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544502617801047,
+ "grad_norm": 0.0450211800634861,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 4.184865474700928,
+ "step": 1335
+ },
+ {
+ "epoch": 18.55846422338569,
+ "grad_norm": 0.0475253090262413,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 4.185857772827148,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57242582897033,
+ "grad_norm": 0.04332203418016434,
+ "learning_rate": 0.000435215814386134,
+ "loss": 4.160165309906006,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586387434554975,
+ "grad_norm": 0.0427570678293705,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 4.1727495193481445,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600349040139616,
+ "grad_norm": 0.04651174694299698,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 4.204797744750977,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61431064572426,
+ "grad_norm": 0.04324701428413391,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 4.186931610107422,
+ "step": 1340
+ },
+ {
+ "epoch": 18.6282722513089,
+ "grad_norm": 0.044644393026828766,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 4.209554195404053,
+ "step": 1341
+ },
+ {
+ "epoch": 18.642233856893544,
+ "grad_norm": 0.042982228100299835,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 4.205570220947266,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656195462478184,
+ "grad_norm": 0.04385090991854668,
+ "learning_rate": 0.000433781621332752,
+ "loss": 4.177936553955078,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670157068062828,
+ "grad_norm": 0.04771510139107704,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 4.1795806884765625,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68411867364747,
+ "grad_norm": 0.04551433399319649,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 4.192194938659668,
+ "step": 1345
+ },
+ {
+ "epoch": 18.698080279232112,
+ "grad_norm": 0.043501488864421844,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 4.177781105041504,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712041884816752,
+ "grad_norm": 0.04394034296274185,
+ "learning_rate": 0.000432823382695327,
+ "loss": 4.206354141235352,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726003490401396,
+ "grad_norm": 0.04911153391003609,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 4.205874919891357,
+ "step": 1348
+ },
+ {
+ "epoch": 18.739965095986037,
+ "grad_norm": 0.04602045193314552,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 4.1947021484375,
+ "step": 1349
+ },
+ {
+ "epoch": 18.75392670157068,
+ "grad_norm": 0.04393904283642769,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 4.188337802886963,
+ "step": 1350
+ },
+ {
+ "epoch": 18.767888307155324,
+ "grad_norm": 0.04526393488049507,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 4.175235748291016,
+ "step": 1351
+ },
+ {
+ "epoch": 18.781849912739965,
+ "grad_norm": 0.041930247098207474,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 4.210986137390137,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79581151832461,
+ "grad_norm": 0.044262565672397614,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 4.176301002502441,
+ "step": 1353
+ },
+ {
+ "epoch": 18.80977312390925,
+ "grad_norm": 0.044190581887960434,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 4.1888837814331055,
+ "step": 1354
+ },
+ {
+ "epoch": 18.823734729493893,
+ "grad_norm": 0.045371655374765396,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 4.182244777679443,
+ "step": 1355
+ },
+ {
+ "epoch": 18.837696335078533,
+ "grad_norm": 0.04302282631397247,
+ "learning_rate": 0.000430661245733797,
+ "loss": 4.189350128173828,
+ "step": 1356
+ },
+ {
+ "epoch": 18.851657940663177,
+ "grad_norm": 0.04402372986078262,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 4.233372211456299,
+ "step": 1357
+ },
+ {
+ "epoch": 18.865619546247817,
+ "grad_norm": 0.04397045075893402,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 4.21682071685791,
+ "step": 1358
+ },
+ {
+ "epoch": 18.87958115183246,
+ "grad_norm": 0.04319734871387482,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 4.204183578491211,
+ "step": 1359
+ },
+ {
+ "epoch": 18.8935427574171,
+ "grad_norm": 0.04600509628653526,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 4.206524848937988,
+ "step": 1360
+ },
+ {
+ "epoch": 18.907504363001745,
+ "grad_norm": 0.04364705830812454,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 4.213656425476074,
+ "step": 1361
+ },
+ {
+ "epoch": 18.921465968586386,
+ "grad_norm": 0.044004205614328384,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 4.1841583251953125,
+ "step": 1362
+ },
+ {
+ "epoch": 18.93542757417103,
+ "grad_norm": 0.04480595886707306,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 4.166214942932129,
+ "step": 1363
+ },
+ {
+ "epoch": 18.949389179755673,
+ "grad_norm": 0.04260887950658798,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 4.198726654052734,
+ "step": 1364
+ },
+ {
+ "epoch": 18.963350785340314,
+ "grad_norm": 0.04513927176594734,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 4.2122392654418945,
+ "step": 1365
+ },
+ {
+ "epoch": 18.977312390924958,
+ "grad_norm": 0.04286042973399162,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 4.192383766174316,
+ "step": 1366
+ },
+ {
+ "epoch": 18.991273996509598,
+ "grad_norm": 0.04287346079945564,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 4.221860408782959,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.030670644715428352,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 2.624823570251465,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.5910831689834595,
+ "eval_runtime": 60.5517,
+ "eval_samples_per_second": 40.329,
+ "eval_steps_per_second": 0.644,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013961605584644,
+ "grad_norm": 0.046196095645427704,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 4.104501724243164,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027923211169284,
+ "grad_norm": 0.05348595231771469,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 4.109098434448242,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041884816753928,
+ "grad_norm": 0.05256905406713486,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 4.115832328796387,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05584642233857,
+ "grad_norm": 0.05241430550813675,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 4.079610824584961,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069808027923212,
+ "grad_norm": 0.04983370006084442,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 4.095335006713867,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083769633507853,
+ "grad_norm": 0.04900007322430611,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 4.093306541442871,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097731239092496,
+ "grad_norm": 0.04995666816830635,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 4.111114501953125,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111692844677137,
+ "grad_norm": 0.04875681549310684,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 4.091944694519043,
+ "step": 1376
+ },
+ {
+ "epoch": 19.12565445026178,
+ "grad_norm": 0.05283811315894127,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 4.112549304962158,
+ "step": 1377
+ },
+ {
+ "epoch": 19.13961605584642,
+ "grad_norm": 0.04801992326974869,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 4.108818054199219,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153577661431065,
+ "grad_norm": 0.04751930385828018,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 4.089441299438477,
+ "step": 1379
+ },
+ {
+ "epoch": 19.167539267015705,
+ "grad_norm": 0.045653726905584335,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 4.07722282409668,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18150087260035,
+ "grad_norm": 0.045616209506988525,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 4.097578048706055,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195462478184993,
+ "grad_norm": 0.04929365590214729,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 4.082221984863281,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209424083769633,
+ "grad_norm": 0.04915834963321686,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 4.123330116271973,
+ "step": 1383
+ },
+ {
+ "epoch": 19.223385689354277,
+ "grad_norm": 0.0470132976770401,
+ "learning_rate": 0.000423881957237287,
+ "loss": 4.113927841186523,
+ "step": 1384
+ },
+ {
+ "epoch": 19.237347294938917,
+ "grad_norm": 0.04695259779691696,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 4.096814155578613,
+ "step": 1385
+ },
+ {
+ "epoch": 19.25130890052356,
+ "grad_norm": 0.04730398207902908,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 4.122525215148926,
+ "step": 1386
+ },
+ {
+ "epoch": 19.2652705061082,
+ "grad_norm": 0.050295207649469376,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 4.117993354797363,
+ "step": 1387
+ },
+ {
+ "epoch": 19.279232111692846,
+ "grad_norm": 0.05082492157816887,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 4.14082145690918,
+ "step": 1388
+ },
+ {
+ "epoch": 19.293193717277486,
+ "grad_norm": 0.04960232973098755,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 4.139887809753418,
+ "step": 1389
+ },
+ {
+ "epoch": 19.30715532286213,
+ "grad_norm": 0.04851173236966133,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 4.138492584228516,
+ "step": 1390
+ },
+ {
+ "epoch": 19.32111692844677,
+ "grad_norm": 0.047838468104600906,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 4.118064880371094,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335078534031414,
+ "grad_norm": 0.05148220434784889,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 4.136533737182617,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349040139616054,
+ "grad_norm": 0.053177181631326675,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 4.1125168800354,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363001745200698,
+ "grad_norm": 0.04861171543598175,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 4.1437153816223145,
+ "step": 1394
+ },
+ {
+ "epoch": 19.376963350785342,
+ "grad_norm": 0.05194924771785736,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 4.097813606262207,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390924956369982,
+ "grad_norm": 0.049609847366809845,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 4.068940162658691,
+ "step": 1396
+ },
+ {
+ "epoch": 19.404886561954626,
+ "grad_norm": 0.047042325139045715,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 4.121333122253418,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418848167539267,
+ "grad_norm": 0.047646693885326385,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 4.114943981170654,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43280977312391,
+ "grad_norm": 0.04676460474729538,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 4.139252185821533,
+ "step": 1399
+ },
+ {
+ "epoch": 19.44677137870855,
+ "grad_norm": 0.04905956983566284,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 4.135970115661621,
+ "step": 1400
+ },
+ {
+ "epoch": 19.460732984293195,
+ "grad_norm": 0.050025034695863724,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 4.15662956237793,
+ "step": 1401
+ },
+ {
+ "epoch": 19.474694589877835,
+ "grad_norm": 0.04740425944328308,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 4.140295505523682,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48865619546248,
+ "grad_norm": 0.046904709190130234,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 4.14083194732666,
+ "step": 1403
+ },
+ {
+ "epoch": 19.50261780104712,
+ "grad_norm": 0.048393651843070984,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 4.161823272705078,
+ "step": 1404
+ },
+ {
+ "epoch": 19.516579406631763,
+ "grad_norm": 0.04766222462058067,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 4.151817798614502,
+ "step": 1405
+ },
+ {
+ "epoch": 19.530541012216403,
+ "grad_norm": 0.04797709733247757,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 4.173160552978516,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544502617801047,
+ "grad_norm": 0.05012711137533188,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 4.161989688873291,
+ "step": 1407
+ },
+ {
+ "epoch": 19.55846422338569,
+ "grad_norm": 0.04726000875234604,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 4.136838436126709,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57242582897033,
+ "grad_norm": 0.046251531690359116,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 4.1526994705200195,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586387434554975,
+ "grad_norm": 0.047958049923181534,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 4.13817024230957,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600349040139616,
+ "grad_norm": 0.046333350241184235,
+ "learning_rate": 0.000417272427439646,
+ "loss": 4.160187721252441,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61431064572426,
+ "grad_norm": 0.0488678440451622,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 4.166190147399902,
+ "step": 1412
+ },
+ {
+ "epoch": 19.6282722513089,
+ "grad_norm": 0.04784323647618294,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 4.164414405822754,
+ "step": 1413
+ },
+ {
+ "epoch": 19.642233856893544,
+ "grad_norm": 0.05064673349261284,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 4.149814128875732,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656195462478184,
+ "grad_norm": 0.05072085186839104,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 4.167759895324707,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670157068062828,
+ "grad_norm": 0.0499395951628685,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 4.162792205810547,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68411867364747,
+ "grad_norm": 0.046766918152570724,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 4.1416521072387695,
+ "step": 1417
+ },
+ {
+ "epoch": 19.698080279232112,
+ "grad_norm": 0.04523273929953575,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 4.161751747131348,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712041884816752,
+ "grad_norm": 0.046106480062007904,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 4.166394233703613,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726003490401396,
+ "grad_norm": 0.04666793718934059,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 4.124961853027344,
+ "step": 1420
+ },
+ {
+ "epoch": 19.739965095986037,
+ "grad_norm": 0.047407012432813644,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 4.1794328689575195,
+ "step": 1421
+ },
+ {
+ "epoch": 19.75392670157068,
+ "grad_norm": 0.04690873995423317,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 4.160114765167236,
+ "step": 1422
+ },
+ {
+ "epoch": 19.767888307155324,
+ "grad_norm": 0.04589436575770378,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 4.137969970703125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.781849912739965,
+ "grad_norm": 0.04641961678862572,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 4.161533355712891,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79581151832461,
+ "grad_norm": 0.045799050480127335,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 4.187039375305176,
+ "step": 1425
+ },
+ {
+ "epoch": 19.80977312390925,
+ "grad_norm": 0.04881894588470459,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 4.158107757568359,
+ "step": 1426
+ },
+ {
+ "epoch": 19.823734729493893,
+ "grad_norm": 0.04506978392601013,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 4.143360137939453,
+ "step": 1427
+ },
+ {
+ "epoch": 19.837696335078533,
+ "grad_norm": 0.04504861310124397,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 4.174564361572266,
+ "step": 1428
+ },
+ {
+ "epoch": 19.851657940663177,
+ "grad_norm": 0.04639774188399315,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 4.162899971008301,
+ "step": 1429
+ },
+ {
+ "epoch": 19.865619546247817,
+ "grad_norm": 0.047444432973861694,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 4.143390655517578,
+ "step": 1430
+ },
+ {
+ "epoch": 19.87958115183246,
+ "grad_norm": 0.04461613669991493,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 4.171284198760986,
+ "step": 1431
+ },
+ {
+ "epoch": 19.8935427574171,
+ "grad_norm": 0.046383798122406006,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 4.14809513092041,
+ "step": 1432
+ },
+ {
+ "epoch": 19.907504363001745,
+ "grad_norm": 0.04642366245388985,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 4.178903579711914,
+ "step": 1433
+ },
+ {
+ "epoch": 19.921465968586386,
+ "grad_norm": 0.0451771542429924,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 4.171825408935547,
+ "step": 1434
+ },
+ {
+ "epoch": 19.93542757417103,
+ "grad_norm": 0.04767893627285957,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 4.169270038604736,
+ "step": 1435
+ },
+ {
+ "epoch": 19.949389179755673,
+ "grad_norm": 0.04673412814736366,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 4.175506591796875,
+ "step": 1436
+ },
+ {
+ "epoch": 19.963350785340314,
+ "grad_norm": 0.04465145245194435,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 4.149707794189453,
+ "step": 1437
+ },
+ {
+ "epoch": 19.977312390924958,
+ "grad_norm": 0.04734131321310997,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 4.166044235229492,
+ "step": 1438
+ },
+ {
+ "epoch": 19.991273996509598,
+ "grad_norm": 0.04706178605556488,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 4.152081489562988,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.031556785106658936,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 2.60109543800354,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013961605584644,
+ "grad_norm": 0.05152537673711777,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 4.084830284118652,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027923211169284,
+ "grad_norm": 0.05918274074792862,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 4.045302391052246,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041884816753928,
+ "grad_norm": 0.05093073844909668,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 4.070842742919922,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05584642233857,
+ "grad_norm": 0.05095985531806946,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 4.083273410797119,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069808027923212,
+ "grad_norm": 0.05502687394618988,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 4.072626113891602,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083769633507853,
+ "grad_norm": 0.05037364363670349,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 4.075478553771973,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097731239092496,
+ "grad_norm": 0.04983789846301079,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 4.066207408905029,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111692844677137,
+ "grad_norm": 0.049478255212306976,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 4.069969654083252,
+ "step": 1448
+ },
+ {
+ "epoch": 20.12565445026178,
+ "grad_norm": 0.051651593297719955,
+ "learning_rate": 0.000407857329622967,
+ "loss": 4.050713062286377,
+ "step": 1449
+ },
+ {
+ "epoch": 20.13961605584642,
+ "grad_norm": 0.05436555668711662,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 4.069849967956543,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153577661431065,
+ "grad_norm": 0.055190350860357285,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 4.081624984741211,
+ "step": 1451
+ },
+ {
+ "epoch": 20.167539267015705,
+ "grad_norm": 0.053246211260557175,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 4.076076507568359,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18150087260035,
+ "grad_norm": 0.051757775247097015,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 4.049139022827148,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195462478184993,
+ "grad_norm": 0.05064203217625618,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 4.068076133728027,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209424083769633,
+ "grad_norm": 0.05114085227251053,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 4.092901229858398,
+ "step": 1455
+ },
+ {
+ "epoch": 20.223385689354277,
+ "grad_norm": 0.047997843474149704,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 4.077084541320801,
+ "step": 1456
+ },
+ {
+ "epoch": 20.237347294938917,
+ "grad_norm": 0.04880451038479805,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 4.059187889099121,
+ "step": 1457
+ },
+ {
+ "epoch": 20.25130890052356,
+ "grad_norm": 0.05036284029483795,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 4.102115154266357,
+ "step": 1458
+ },
+ {
+ "epoch": 20.2652705061082,
+ "grad_norm": 0.04916979745030403,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 4.098356246948242,
+ "step": 1459
+ },
+ {
+ "epoch": 20.279232111692846,
+ "grad_norm": 0.04713984206318855,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 4.057218074798584,
+ "step": 1460
+ },
+ {
+ "epoch": 20.293193717277486,
+ "grad_norm": 0.04982447996735573,
+ "learning_rate": 0.000404858275823277,
+ "loss": 4.070363998413086,
+ "step": 1461
+ },
+ {
+ "epoch": 20.30715532286213,
+ "grad_norm": 0.0475616492331028,
+ "learning_rate": 0.000404607816444578,
+ "loss": 4.074314117431641,
+ "step": 1462
+ },
+ {
+ "epoch": 20.32111692844677,
+ "grad_norm": 0.046537093818187714,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 4.079501152038574,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335078534031414,
+ "grad_norm": 0.04889644682407379,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 4.0731425285339355,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349040139616054,
+ "grad_norm": 0.05095166340470314,
+ "learning_rate": 0.000403855947934478,
+ "loss": 4.08765983581543,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363001745200698,
+ "grad_norm": 0.049677830189466476,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 4.128052711486816,
+ "step": 1466
+ },
+ {
+ "epoch": 20.376963350785342,
+ "grad_norm": 0.05221942439675331,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 4.094407081604004,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390924956369982,
+ "grad_norm": 0.052042894065380096,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 4.088282585144043,
+ "step": 1468
+ },
+ {
+ "epoch": 20.404886561954626,
+ "grad_norm": 0.04822041094303131,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 4.1070990562438965,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418848167539267,
+ "grad_norm": 0.047207772731781006,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 4.102565765380859,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43280977312391,
+ "grad_norm": 0.049633774906396866,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 4.097216606140137,
+ "step": 1471
+ },
+ {
+ "epoch": 20.44677137870855,
+ "grad_norm": 0.05135185644030571,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 4.095193862915039,
+ "step": 1472
+ },
+ {
+ "epoch": 20.460732984293195,
+ "grad_norm": 0.05215826258063316,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 4.085207939147949,
+ "step": 1473
+ },
+ {
+ "epoch": 20.474694589877835,
+ "grad_norm": 0.050653353333473206,
+ "learning_rate": 0.000401595976318565,
+ "loss": 4.085142612457275,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48865619546248,
+ "grad_norm": 0.04795852303504944,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 4.1158270835876465,
+ "step": 1475
+ },
+ {
+ "epoch": 20.50261780104712,
+ "grad_norm": 0.047416940331459045,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 4.096989154815674,
+ "step": 1476
+ },
+ {
+ "epoch": 20.516579406631763,
+ "grad_norm": 0.049753669649362564,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 4.118251800537109,
+ "step": 1477
+ },
+ {
+ "epoch": 20.530541012216403,
+ "grad_norm": 0.04889055714011192,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 4.081287384033203,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544502617801047,
+ "grad_norm": 0.04857790097594261,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 4.084896564483643,
+ "step": 1479
+ },
+ {
+ "epoch": 20.55846422338569,
+ "grad_norm": 0.047250498086214066,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 4.113656997680664,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57242582897033,
+ "grad_norm": 0.048608481884002686,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 4.113809108734131,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586387434554975,
+ "grad_norm": 0.048189081251621246,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 4.113949775695801,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600349040139616,
+ "grad_norm": 0.049929242581129074,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 4.104221820831299,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61431064572426,
+ "grad_norm": 0.048482850193977356,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 4.091039180755615,
+ "step": 1484
+ },
+ {
+ "epoch": 20.6282722513089,
+ "grad_norm": 0.04810645803809166,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 4.130088806152344,
+ "step": 1485
+ },
+ {
+ "epoch": 20.642233856893544,
+ "grad_norm": 0.04745674133300781,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 4.106161117553711,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656195462478184,
+ "grad_norm": 0.048513609915971756,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 4.119222164154053,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670157068062828,
+ "grad_norm": 0.04506325349211693,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 4.115543842315674,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68411867364747,
+ "grad_norm": 0.04726741835474968,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 4.084345817565918,
+ "step": 1489
+ },
+ {
+ "epoch": 20.698080279232112,
+ "grad_norm": 0.04873545467853546,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 4.116580009460449,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712041884816752,
+ "grad_norm": 0.051252465695142746,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 4.129733085632324,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726003490401396,
+ "grad_norm": 0.0489567406475544,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 4.103477954864502,
+ "step": 1492
+ },
+ {
+ "epoch": 20.739965095986037,
+ "grad_norm": 0.04539056867361069,
+ "learning_rate": 0.000396803974909638,
+ "loss": 4.1327667236328125,
+ "step": 1493
+ },
+ {
+ "epoch": 20.75392670157068,
+ "grad_norm": 0.04954095557332039,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 4.084133625030518,
+ "step": 1494
+ },
+ {
+ "epoch": 20.767888307155324,
+ "grad_norm": 0.05041591823101044,
+ "learning_rate": 0.000396297942393265,
+ "loss": 4.12288761138916,
+ "step": 1495
+ },
+ {
+ "epoch": 20.781849912739965,
+ "grad_norm": 0.046549517661333084,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 4.1259918212890625,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79581151832461,
+ "grad_norm": 0.04704999923706055,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 4.102563858032227,
+ "step": 1497
+ },
+ {
+ "epoch": 20.80977312390925,
+ "grad_norm": 0.05025320500135422,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 4.135492324829102,
+ "step": 1498
+ },
+ {
+ "epoch": 20.823734729493893,
+ "grad_norm": 0.04559441655874252,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 4.118548393249512,
+ "step": 1499
+ },
+ {
+ "epoch": 20.837696335078533,
+ "grad_norm": 0.04913092032074928,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 4.129302024841309,
+ "step": 1500
+ },
+ {
+ "epoch": 20.851657940663177,
+ "grad_norm": 0.04846089705824852,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 4.166059970855713,
+ "step": 1501
+ },
+ {
+ "epoch": 20.865619546247817,
+ "grad_norm": 0.04710553213953972,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 4.104712009429932,
+ "step": 1502
+ },
+ {
+ "epoch": 20.87958115183246,
+ "grad_norm": 0.04896444454789162,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 4.140282154083252,
+ "step": 1503
+ },
+ {
+ "epoch": 20.8935427574171,
+ "grad_norm": 0.0469740591943264,
+ "learning_rate": 0.000394017098798633,
+ "loss": 4.125306129455566,
+ "step": 1504
+ },
+ {
+ "epoch": 20.907504363001745,
+ "grad_norm": 0.04638523980975151,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 4.133161544799805,
+ "step": 1505
+ },
+ {
+ "epoch": 20.921465968586386,
+ "grad_norm": 0.04814442619681358,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 4.093092918395996,
+ "step": 1506
+ },
+ {
+ "epoch": 20.93542757417103,
+ "grad_norm": 0.049911532551050186,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 4.1001715660095215,
+ "step": 1507
+ },
+ {
+ "epoch": 20.949389179755673,
+ "grad_norm": 0.0446670837700367,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 4.140350341796875,
+ "step": 1508
+ },
+ {
+ "epoch": 20.963350785340314,
+ "grad_norm": 0.04789397493004799,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 4.163155555725098,
+ "step": 1509
+ },
+ {
+ "epoch": 20.977312390924958,
+ "grad_norm": 0.050148315727710724,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 4.118380069732666,
+ "step": 1510
+ },
+ {
+ "epoch": 20.991273996509598,
+ "grad_norm": 0.047446198761463165,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 4.116507530212402,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.03415544703602791,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 2.5738282203674316,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.5933796167373657,
+ "eval_runtime": 60.4011,
+ "eval_samples_per_second": 40.43,
+ "eval_steps_per_second": 0.646,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013961605584644,
+ "grad_norm": 0.052276045083999634,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 4.0301923751831055,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027923211169284,
+ "grad_norm": 0.06514342874288559,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 4.031280517578125,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041884816753928,
+ "grad_norm": 0.05948134884238243,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 4.060680389404297,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05584642233857,
+ "grad_norm": 0.05251640826463699,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 4.0218400955200195,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069808027923212,
+ "grad_norm": 0.05426032841205597,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 4.026187896728516,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083769633507853,
+ "grad_norm": 0.052583567798137665,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 4.016767501831055,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097731239092496,
+ "grad_norm": 0.04966135695576668,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 4.043508529663086,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111692844677137,
+ "grad_norm": 0.053838398307561874,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 4.054832458496094,
+ "step": 1520
+ },
+ {
+ "epoch": 21.12565445026178,
+ "grad_norm": 0.05304184556007385,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 4.03280782699585,
+ "step": 1521
+ },
+ {
+ "epoch": 21.13961605584642,
+ "grad_norm": 0.05252664163708687,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 4.039841175079346,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153577661431065,
+ "grad_norm": 0.054593686014413834,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 4.048308372497559,
+ "step": 1523
+ },
+ {
+ "epoch": 21.167539267015705,
+ "grad_norm": 0.050722211599349976,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 4.051860332489014,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18150087260035,
+ "grad_norm": 0.05248095095157623,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 4.056846618652344,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195462478184993,
+ "grad_norm": 0.05166107788681984,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 4.045178413391113,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209424083769633,
+ "grad_norm": 0.04987792298197746,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 4.014827728271484,
+ "step": 1527
+ },
+ {
+ "epoch": 21.223385689354277,
+ "grad_norm": 0.05316167324781418,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 4.033143997192383,
+ "step": 1528
+ },
+ {
+ "epoch": 21.237347294938917,
+ "grad_norm": 0.05228257551789284,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 4.033842086791992,
+ "step": 1529
+ },
+ {
+ "epoch": 21.25130890052356,
+ "grad_norm": 0.049349382519721985,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 4.025982856750488,
+ "step": 1530
+ },
+ {
+ "epoch": 21.2652705061082,
+ "grad_norm": 0.055104926228523254,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 4.059151649475098,
+ "step": 1531
+ },
+ {
+ "epoch": 21.279232111692846,
+ "grad_norm": 0.052629340440034866,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 4.032320976257324,
+ "step": 1532
+ },
+ {
+ "epoch": 21.293193717277486,
+ "grad_norm": 0.05396415665745735,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 4.0425567626953125,
+ "step": 1533
+ },
+ {
+ "epoch": 21.30715532286213,
+ "grad_norm": 0.054417286068201065,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 4.03759241104126,
+ "step": 1534
+ },
+ {
+ "epoch": 21.32111692844677,
+ "grad_norm": 0.061248697340488434,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 4.0441083908081055,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335078534031414,
+ "grad_norm": 0.05934970825910568,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 4.050718307495117,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349040139616054,
+ "grad_norm": 0.052805181592702866,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 4.069460868835449,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363001745200698,
+ "grad_norm": 0.052820611745119095,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 4.063657283782959,
+ "step": 1538
+ },
+ {
+ "epoch": 21.376963350785342,
+ "grad_norm": 0.05043052136898041,
+ "learning_rate": 0.00038509205478744,
+ "loss": 4.029552459716797,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390924956369982,
+ "grad_norm": 0.04828709363937378,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 4.05656623840332,
+ "step": 1540
+ },
+ {
+ "epoch": 21.404886561954626,
+ "grad_norm": 0.049760788679122925,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 4.066338539123535,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418848167539267,
+ "grad_norm": 0.05037159472703934,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 4.06976842880249,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43280977312391,
+ "grad_norm": 0.05276317894458771,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 4.062975883483887,
+ "step": 1543
+ },
+ {
+ "epoch": 21.44677137870855,
+ "grad_norm": 0.05571872368454933,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 4.065622329711914,
+ "step": 1544
+ },
+ {
+ "epoch": 21.460732984293195,
+ "grad_norm": 0.05438469722867012,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 4.074963569641113,
+ "step": 1545
+ },
+ {
+ "epoch": 21.474694589877835,
+ "grad_norm": 0.053415946662425995,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 4.080055236816406,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48865619546248,
+ "grad_norm": 0.05216296389698982,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 4.091355800628662,
+ "step": 1547
+ },
+ {
+ "epoch": 21.50261780104712,
+ "grad_norm": 0.051212459802627563,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 4.058528900146484,
+ "step": 1548
+ },
+ {
+ "epoch": 21.516579406631763,
+ "grad_norm": 0.05033959075808525,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 4.064615726470947,
+ "step": 1549
+ },
+ {
+ "epoch": 21.530541012216403,
+ "grad_norm": 0.04665188118815422,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 4.059698104858398,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544502617801047,
+ "grad_norm": 0.050221361219882965,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 4.068614959716797,
+ "step": 1551
+ },
+ {
+ "epoch": 21.55846422338569,
+ "grad_norm": 0.05089551955461502,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 4.0833892822265625,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57242582897033,
+ "grad_norm": 0.051966603845357895,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 4.083354949951172,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586387434554975,
+ "grad_norm": 0.04981927201151848,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 4.083340644836426,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600349040139616,
+ "grad_norm": 0.05141765624284744,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 4.066590309143066,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61431064572426,
+ "grad_norm": 0.048985838890075684,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 4.088237762451172,
+ "step": 1556
+ },
+ {
+ "epoch": 21.6282722513089,
+ "grad_norm": 0.05013054236769676,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 4.081377983093262,
+ "step": 1557
+ },
+ {
+ "epoch": 21.642233856893544,
+ "grad_norm": 0.05170975625514984,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 4.054303169250488,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656195462478184,
+ "grad_norm": 0.048257142305374146,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 4.103672504425049,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670157068062828,
+ "grad_norm": 0.049942679703235626,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 4.029951572418213,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68411867364747,
+ "grad_norm": 0.05083315819501877,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 4.048798084259033,
+ "step": 1561
+ },
+ {
+ "epoch": 21.698080279232112,
+ "grad_norm": 0.04984128847718239,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 4.079655647277832,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712041884816752,
+ "grad_norm": 0.04985307902097702,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 4.071877956390381,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726003490401396,
+ "grad_norm": 0.051137227565050125,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 4.076539993286133,
+ "step": 1564
+ },
+ {
+ "epoch": 21.739965095986037,
+ "grad_norm": 0.04968905821442604,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 4.084498882293701,
+ "step": 1565
+ },
+ {
+ "epoch": 21.75392670157068,
+ "grad_norm": 0.049119509756565094,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 4.066773414611816,
+ "step": 1566
+ },
+ {
+ "epoch": 21.767888307155324,
+ "grad_norm": 0.05107425898313522,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 4.065309047698975,
+ "step": 1567
+ },
+ {
+ "epoch": 21.781849912739965,
+ "grad_norm": 0.049012504518032074,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 4.062110900878906,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79581151832461,
+ "grad_norm": 0.04972011595964432,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 4.0630903244018555,
+ "step": 1569
+ },
+ {
+ "epoch": 21.80977312390925,
+ "grad_norm": 0.04822700098156929,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 4.106127738952637,
+ "step": 1570
+ },
+ {
+ "epoch": 21.823734729493893,
+ "grad_norm": 0.04808267205953598,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 4.085200309753418,
+ "step": 1571
+ },
+ {
+ "epoch": 21.837696335078533,
+ "grad_norm": 0.048644762486219406,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 4.0856170654296875,
+ "step": 1572
+ },
+ {
+ "epoch": 21.851657940663177,
+ "grad_norm": 0.05232283100485802,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 4.126387119293213,
+ "step": 1573
+ },
+ {
+ "epoch": 21.865619546247817,
+ "grad_norm": 0.05095643922686577,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 4.108254432678223,
+ "step": 1574
+ },
+ {
+ "epoch": 21.87958115183246,
+ "grad_norm": 0.045972101390361786,
+ "learning_rate": 0.000375827577564042,
+ "loss": 4.058515548706055,
+ "step": 1575
+ },
+ {
+ "epoch": 21.8935427574171,
+ "grad_norm": 0.04974910616874695,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 4.089197158813477,
+ "step": 1576
+ },
+ {
+ "epoch": 21.907504363001745,
+ "grad_norm": 0.04890159144997597,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 4.1078410148620605,
+ "step": 1577
+ },
+ {
+ "epoch": 21.921465968586386,
+ "grad_norm": 0.04738525673747063,
+ "learning_rate": 0.000375051971546236,
+ "loss": 4.1219282150268555,
+ "step": 1578
+ },
+ {
+ "epoch": 21.93542757417103,
+ "grad_norm": 0.04713096842169762,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 4.10295295715332,
+ "step": 1579
+ },
+ {
+ "epoch": 21.949389179755673,
+ "grad_norm": 0.04834369197487831,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 4.102747917175293,
+ "step": 1580
+ },
+ {
+ "epoch": 21.963350785340314,
+ "grad_norm": 0.047073688358068466,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 4.08024787902832,
+ "step": 1581
+ },
+ {
+ "epoch": 21.977312390924958,
+ "grad_norm": 0.04568164795637131,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 4.078930854797363,
+ "step": 1582
+ },
+ {
+ "epoch": 21.991273996509598,
+ "grad_norm": 0.04822581633925438,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 4.089601516723633,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.03507270663976669,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 2.5586814880371094,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.595272958278656,
+ "eval_runtime": 60.7797,
+ "eval_samples_per_second": 40.178,
+ "eval_steps_per_second": 0.642,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013961605584644,
+ "grad_norm": 0.05232533439993858,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 4.011311054229736,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027923211169284,
+ "grad_norm": 0.058515798300504684,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 4.0049638748168945,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041884816753928,
+ "grad_norm": 0.055082619190216064,
+ "learning_rate": 0.000372722041257003,
+ "loss": 3.9904184341430664,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05584642233857,
+ "grad_norm": 0.05213981494307518,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 4.003454685211182,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069808027923212,
+ "grad_norm": 0.051619648933410645,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 3.9900388717651367,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083769633507853,
+ "grad_norm": 0.055212799459695816,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 3.973494052886963,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097731239092496,
+ "grad_norm": 0.05262092128396034,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 4.00933837890625,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111692844677137,
+ "grad_norm": 0.05012505501508713,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 4.01861047744751,
+ "step": 1592
+ },
+ {
+ "epoch": 22.12565445026178,
+ "grad_norm": 0.049657098948955536,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 4.011143684387207,
+ "step": 1593
+ },
+ {
+ "epoch": 22.13961605584642,
+ "grad_norm": 0.05187768116593361,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 3.9922969341278076,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153577661431065,
+ "grad_norm": 0.05053112655878067,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 4.007021903991699,
+ "step": 1595
+ },
+ {
+ "epoch": 22.167539267015705,
+ "grad_norm": 0.048542529344558716,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 3.9966416358947754,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18150087260035,
+ "grad_norm": 0.050972841680049896,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 3.9803225994110107,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195462478184993,
+ "grad_norm": 0.05001165345311165,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 4.0140509605407715,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209424083769633,
+ "grad_norm": 0.05100536718964577,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 3.9906058311462402,
+ "step": 1599
+ },
+ {
+ "epoch": 22.223385689354277,
+ "grad_norm": 0.052977193146944046,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 4.012383460998535,
+ "step": 1600
+ },
+ {
+ "epoch": 22.237347294938917,
+ "grad_norm": 0.051219623535871506,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 3.982931613922119,
+ "step": 1601
+ },
+ {
+ "epoch": 22.25130890052356,
+ "grad_norm": 0.05275029316544533,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 4.032669544219971,
+ "step": 1602
+ },
+ {
+ "epoch": 22.2652705061082,
+ "grad_norm": 0.04781945049762726,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 3.9933102130889893,
+ "step": 1603
+ },
+ {
+ "epoch": 22.279232111692846,
+ "grad_norm": 0.054087597876787186,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 4.0075201988220215,
+ "step": 1604
+ },
+ {
+ "epoch": 22.293193717277486,
+ "grad_norm": 0.05485915020108223,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 4.004299640655518,
+ "step": 1605
+ },
+ {
+ "epoch": 22.30715532286213,
+ "grad_norm": 0.05148882791399956,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 4.044161796569824,
+ "step": 1606
+ },
+ {
+ "epoch": 22.32111692844677,
+ "grad_norm": 0.051912397146224976,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 3.9777069091796875,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335078534031414,
+ "grad_norm": 0.052630942314863205,
+ "learning_rate": 0.000367268105959126,
+ "loss": 4.029580116271973,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349040139616054,
+ "grad_norm": 0.0512809231877327,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 4.022092342376709,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363001745200698,
+ "grad_norm": 0.05094102770090103,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 4.018102169036865,
+ "step": 1610
+ },
+ {
+ "epoch": 22.376963350785342,
+ "grad_norm": 0.04972759634256363,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 4.027868270874023,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390924956369982,
+ "grad_norm": 0.05072534456849098,
+ "learning_rate": 0.000366226612810024,
+ "loss": 3.995988368988037,
+ "step": 1612
+ },
+ {
+ "epoch": 22.404886561954626,
+ "grad_norm": 0.050623781979084015,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 4.019711494445801,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418848167539267,
+ "grad_norm": 0.05035007372498512,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 4.007514953613281,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43280977312391,
+ "grad_norm": 0.054426416754722595,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 4.0167131423950195,
+ "step": 1615
+ },
+ {
+ "epoch": 22.44677137870855,
+ "grad_norm": 0.05365690588951111,
+ "learning_rate": 0.000365184304613104,
+ "loss": 4.0313920974731445,
+ "step": 1616
+ },
+ {
+ "epoch": 22.460732984293195,
+ "grad_norm": 0.0509161539375782,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 4.022063255310059,
+ "step": 1617
+ },
+ {
+ "epoch": 22.474694589877835,
+ "grad_norm": 0.05309660732746124,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 4.017317771911621,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48865619546248,
+ "grad_norm": 0.0488961897790432,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 4.028288841247559,
+ "step": 1619
+ },
+ {
+ "epoch": 22.50261780104712,
+ "grad_norm": 0.04982095956802368,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 4.007115840911865,
+ "step": 1620
+ },
+ {
+ "epoch": 22.516579406631763,
+ "grad_norm": 0.05136589705944061,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 4.028022766113281,
+ "step": 1621
+ },
+ {
+ "epoch": 22.530541012216403,
+ "grad_norm": 0.04865279048681259,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 4.015928268432617,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544502617801047,
+ "grad_norm": 0.05131891369819641,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 4.041997909545898,
+ "step": 1623
+ },
+ {
+ "epoch": 22.55846422338569,
+ "grad_norm": 0.05063068866729736,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 4.042117118835449,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57242582897033,
+ "grad_norm": 0.049348752945661545,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 4.012609481811523,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586387434554975,
+ "grad_norm": 0.05123874917626381,
+ "learning_rate": 0.000362575056487632,
+ "loss": 4.03358793258667,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600349040139616,
+ "grad_norm": 0.04921206086874008,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 4.039857864379883,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61431064572426,
+ "grad_norm": 0.05011922866106033,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 4.015835762023926,
+ "step": 1628
+ },
+ {
+ "epoch": 22.6282722513089,
+ "grad_norm": 0.049950532615184784,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 4.068299293518066,
+ "step": 1629
+ },
+ {
+ "epoch": 22.642233856893544,
+ "grad_norm": 0.04962172359228134,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 4.03205680847168,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656195462478184,
+ "grad_norm": 0.05059773847460747,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 4.065027236938477,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670157068062828,
+ "grad_norm": 0.04986768960952759,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 4.01540470123291,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68411867364747,
+ "grad_norm": 0.050884976983070374,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 4.060809135437012,
+ "step": 1633
+ },
+ {
+ "epoch": 22.698080279232112,
+ "grad_norm": 0.052478861063718796,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 4.088409423828125,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712041884816752,
+ "grad_norm": 0.05118529871106148,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 4.071645736694336,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726003490401396,
+ "grad_norm": 0.05059075728058815,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 4.028438568115234,
+ "step": 1636
+ },
+ {
+ "epoch": 22.739965095986037,
+ "grad_norm": 0.049222562462091446,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 4.02166223526001,
+ "step": 1637
+ },
+ {
+ "epoch": 22.75392670157068,
+ "grad_norm": 0.049673549830913544,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 4.0511860847473145,
+ "step": 1638
+ },
+ {
+ "epoch": 22.767888307155324,
+ "grad_norm": 0.04954935610294342,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 4.036063194274902,
+ "step": 1639
+ },
+ {
+ "epoch": 22.781849912739965,
+ "grad_norm": 0.05110877379775047,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 4.081101417541504,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79581151832461,
+ "grad_norm": 0.049009695649147034,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 4.042314529418945,
+ "step": 1641
+ },
+ {
+ "epoch": 22.80977312390925,
+ "grad_norm": 0.0496281161904335,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 4.028315544128418,
+ "step": 1642
+ },
+ {
+ "epoch": 22.823734729493893,
+ "grad_norm": 0.04931195452809334,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 4.018457412719727,
+ "step": 1643
+ },
+ {
+ "epoch": 22.837696335078533,
+ "grad_norm": 0.05006880313158035,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 4.0709991455078125,
+ "step": 1644
+ },
+ {
+ "epoch": 22.851657940663177,
+ "grad_norm": 0.04956946149468422,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 4.065585136413574,
+ "step": 1645
+ },
+ {
+ "epoch": 22.865619546247817,
+ "grad_norm": 0.048182398080825806,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 4.0522589683532715,
+ "step": 1646
+ },
+ {
+ "epoch": 22.87958115183246,
+ "grad_norm": 0.04964807257056236,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 4.06989860534668,
+ "step": 1647
+ },
+ {
+ "epoch": 22.8935427574171,
+ "grad_norm": 0.049299128353595734,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 4.082858085632324,
+ "step": 1648
+ },
+ {
+ "epoch": 22.907504363001745,
+ "grad_norm": 0.04893320053815842,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 4.047144889831543,
+ "step": 1649
+ },
+ {
+ "epoch": 22.921465968586386,
+ "grad_norm": 0.04772866889834404,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 4.043060302734375,
+ "step": 1650
+ },
+ {
+ "epoch": 22.93542757417103,
+ "grad_norm": 0.049417730420827866,
+ "learning_rate": 0.000356031397755195,
+ "loss": 4.031186580657959,
+ "step": 1651
+ },
+ {
+ "epoch": 22.949389179755673,
+ "grad_norm": 0.049944665282964706,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 4.074260711669922,
+ "step": 1652
+ },
+ {
+ "epoch": 22.963350785340314,
+ "grad_norm": 0.04856271669268608,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 4.054103851318359,
+ "step": 1653
+ },
+ {
+ "epoch": 22.977312390924958,
+ "grad_norm": 0.0470513179898262,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 4.041568756103516,
+ "step": 1654
+ },
+ {
+ "epoch": 22.991273996509598,
+ "grad_norm": 0.049018729478120804,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 4.066161632537842,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.034226156771183014,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 2.5147223472595215,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.5970026850700378,
+ "eval_runtime": 60.0787,
+ "eval_samples_per_second": 40.647,
+ "eval_steps_per_second": 0.649,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013961605584644,
+ "grad_norm": 0.05091085657477379,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 3.95741868019104,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027923211169284,
+ "grad_norm": 0.05874285474419594,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 3.966327667236328,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041884816753928,
+ "grad_norm": 0.05394266918301582,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.937232732772827,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05584642233857,
+ "grad_norm": 0.05120540037751198,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 3.9637370109558105,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069808027923212,
+ "grad_norm": 0.05310065299272537,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 3.998631477355957,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083769633507853,
+ "grad_norm": 0.05127466097474098,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.949101448059082,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097731239092496,
+ "grad_norm": 0.05201594531536102,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 3.986267566680908,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111692844677137,
+ "grad_norm": 0.051927369087934494,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 3.9604103565216064,
+ "step": 1664
+ },
+ {
+ "epoch": 23.12565445026178,
+ "grad_norm": 0.05125141143798828,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.9339547157287598,
+ "step": 1665
+ },
+ {
+ "epoch": 23.13961605584642,
+ "grad_norm": 0.05241088196635246,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 3.971261978149414,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153577661431065,
+ "grad_norm": 0.050563592463731766,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.9450082778930664,
+ "step": 1667
+ },
+ {
+ "epoch": 23.167539267015705,
+ "grad_norm": 0.05204763635993004,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 3.9696707725524902,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18150087260035,
+ "grad_norm": 0.05518610402941704,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 3.9816057682037354,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195462478184993,
+ "grad_norm": 0.05336211621761322,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.932252883911133,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209424083769633,
+ "grad_norm": 0.052204642444849014,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 3.9818131923675537,
+ "step": 1671
+ },
+ {
+ "epoch": 23.223385689354277,
+ "grad_norm": 0.05299545079469681,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.9488978385925293,
+ "step": 1672
+ },
+ {
+ "epoch": 23.237347294938917,
+ "grad_norm": 0.05143733322620392,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 4.00959587097168,
+ "step": 1673
+ },
+ {
+ "epoch": 23.25130890052356,
+ "grad_norm": 0.05186324194073677,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.9529542922973633,
+ "step": 1674
+ },
+ {
+ "epoch": 23.2652705061082,
+ "grad_norm": 0.05334410071372986,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 3.96693754196167,
+ "step": 1675
+ },
+ {
+ "epoch": 23.279232111692846,
+ "grad_norm": 0.05345796048641205,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 3.9674174785614014,
+ "step": 1676
+ },
+ {
+ "epoch": 23.293193717277486,
+ "grad_norm": 0.050988901406526566,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 3.9698638916015625,
+ "step": 1677
+ },
+ {
+ "epoch": 23.30715532286213,
+ "grad_norm": 0.056467678397893906,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 3.9645955562591553,
+ "step": 1678
+ },
+ {
+ "epoch": 23.32111692844677,
+ "grad_norm": 0.05554647371172905,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 4.0036091804504395,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335078534031414,
+ "grad_norm": 0.05607818067073822,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 3.973773717880249,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349040139616054,
+ "grad_norm": 0.05734477937221527,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 3.987692356109619,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363001745200698,
+ "grad_norm": 0.052175361663103104,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 3.9761803150177,
+ "step": 1682
+ },
+ {
+ "epoch": 23.376963350785342,
+ "grad_norm": 0.053368814289569855,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 3.965020179748535,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390924956369982,
+ "grad_norm": 0.05505364388227463,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 3.987494468688965,
+ "step": 1684
+ },
+ {
+ "epoch": 23.404886561954626,
+ "grad_norm": 0.0524514764547348,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 3.9624319076538086,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418848167539267,
+ "grad_norm": 0.05681568384170532,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 4.008373737335205,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43280977312391,
+ "grad_norm": 0.05344909429550171,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 3.9840354919433594,
+ "step": 1687
+ },
+ {
+ "epoch": 23.44677137870855,
+ "grad_norm": 0.052657466381788254,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 3.992262363433838,
+ "step": 1688
+ },
+ {
+ "epoch": 23.460732984293195,
+ "grad_norm": 0.05526747554540634,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 3.97529935836792,
+ "step": 1689
+ },
+ {
+ "epoch": 23.474694589877835,
+ "grad_norm": 0.05492551624774933,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 4.004708766937256,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48865619546248,
+ "grad_norm": 0.053742099553346634,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 3.994016647338867,
+ "step": 1691
+ },
+ {
+ "epoch": 23.50261780104712,
+ "grad_norm": 0.05363722890615463,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 3.9942378997802734,
+ "step": 1692
+ },
+ {
+ "epoch": 23.516579406631763,
+ "grad_norm": 0.05167972669005394,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 3.998253345489502,
+ "step": 1693
+ },
+ {
+ "epoch": 23.530541012216403,
+ "grad_norm": 0.0518590584397316,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 3.9850645065307617,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544502617801047,
+ "grad_norm": 0.05521136894822121,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 4.02669620513916,
+ "step": 1695
+ },
+ {
+ "epoch": 23.55846422338569,
+ "grad_norm": 0.054986029863357544,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 4.022308826446533,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57242582897033,
+ "grad_norm": 0.05405082181096077,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 4.021487236022949,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586387434554975,
+ "grad_norm": 0.051257841289043427,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 3.9724135398864746,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600349040139616,
+ "grad_norm": 0.05109119415283203,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 3.978914499282837,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61431064572426,
+ "grad_norm": 0.05303420498967171,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 4.027509689331055,
+ "step": 1700
+ },
+ {
+ "epoch": 23.6282722513089,
+ "grad_norm": 0.055543459951877594,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 3.9750852584838867,
+ "step": 1701
+ },
+ {
+ "epoch": 23.642233856893544,
+ "grad_norm": 0.05068540200591087,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 4.029600143432617,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656195462478184,
+ "grad_norm": 0.04991193488240242,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 3.9901509284973145,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670157068062828,
+ "grad_norm": 0.05403713881969452,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 4.017679214477539,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68411867364747,
+ "grad_norm": 0.05203867331147194,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 4.008406639099121,
+ "step": 1705
+ },
+ {
+ "epoch": 23.698080279232112,
+ "grad_norm": 0.05106597766280174,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 4.009983062744141,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712041884816752,
+ "grad_norm": 0.04976251348853111,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 4.007230281829834,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726003490401396,
+ "grad_norm": 0.05199650675058365,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 3.9902141094207764,
+ "step": 1708
+ },
+ {
+ "epoch": 23.739965095986037,
+ "grad_norm": 0.05285118520259857,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 3.9966492652893066,
+ "step": 1709
+ },
+ {
+ "epoch": 23.75392670157068,
+ "grad_norm": 0.051937371492385864,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 3.962397336959839,
+ "step": 1710
+ },
+ {
+ "epoch": 23.767888307155324,
+ "grad_norm": 0.05179062858223915,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 4.027141571044922,
+ "step": 1711
+ },
+ {
+ "epoch": 23.781849912739965,
+ "grad_norm": 0.049667056649923325,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 4.024991512298584,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79581151832461,
+ "grad_norm": 0.05204658955335617,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 4.002647876739502,
+ "step": 1713
+ },
+ {
+ "epoch": 23.80977312390925,
+ "grad_norm": 0.05069981515407562,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 4.035422325134277,
+ "step": 1714
+ },
+ {
+ "epoch": 23.823734729493893,
+ "grad_norm": 0.05044785887002945,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 4.042465686798096,
+ "step": 1715
+ },
+ {
+ "epoch": 23.837696335078533,
+ "grad_norm": 0.053949885070323944,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 4.002717971801758,
+ "step": 1716
+ },
+ {
+ "epoch": 23.851657940663177,
+ "grad_norm": 0.05139258876442909,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 4.026609897613525,
+ "step": 1717
+ },
+ {
+ "epoch": 23.865619546247817,
+ "grad_norm": 0.05316372960805893,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 4.019758224487305,
+ "step": 1718
+ },
+ {
+ "epoch": 23.87958115183246,
+ "grad_norm": 0.050791382789611816,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 3.9859633445739746,
+ "step": 1719
+ },
+ {
+ "epoch": 23.8935427574171,
+ "grad_norm": 0.05124847590923309,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 4.038055419921875,
+ "step": 1720
+ },
+ {
+ "epoch": 23.907504363001745,
+ "grad_norm": 0.0495096817612648,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 4.02764368057251,
+ "step": 1721
+ },
+ {
+ "epoch": 23.921465968586386,
+ "grad_norm": 0.05205567181110382,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 4.003867149353027,
+ "step": 1722
+ },
+ {
+ "epoch": 23.93542757417103,
+ "grad_norm": 0.05233347788453102,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 4.024157524108887,
+ "step": 1723
+ },
+ {
+ "epoch": 23.949389179755673,
+ "grad_norm": 0.05126349255442619,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 4.05349063873291,
+ "step": 1724
+ },
+ {
+ "epoch": 23.963350785340314,
+ "grad_norm": 0.050006967037916183,
+ "learning_rate": 0.000336518346307424,
+ "loss": 4.024704933166504,
+ "step": 1725
+ },
+ {
+ "epoch": 23.977312390924958,
+ "grad_norm": 0.052343279123306274,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 4.008549213409424,
+ "step": 1726
+ },
+ {
+ "epoch": 23.991273996509598,
+ "grad_norm": 0.05126652494072914,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 4.002162933349609,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.03560106083750725,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 2.511049270629883,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.5985628962516785,
+ "eval_runtime": 59.0712,
+ "eval_samples_per_second": 41.34,
+ "eval_steps_per_second": 0.66,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013961605584644,
+ "grad_norm": 0.05205906555056572,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.936516284942627,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027923211169284,
+ "grad_norm": 0.056368257850408554,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.95200777053833,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041884816753928,
+ "grad_norm": 0.05075085908174515,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.940810441970825,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05584642233857,
+ "grad_norm": 0.05119713768362999,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.90130615234375,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069808027923212,
+ "grad_norm": 0.05254847928881645,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.9491095542907715,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083769633507853,
+ "grad_norm": 0.053025662899017334,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.9371917247772217,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097731239092496,
+ "grad_norm": 0.05402665585279465,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 3.9651317596435547,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111692844677137,
+ "grad_norm": 0.0519571453332901,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.9189062118530273,
+ "step": 1736
+ },
+ {
+ "epoch": 24.12565445026178,
+ "grad_norm": 0.05436019226908684,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.948184013366699,
+ "step": 1737
+ },
+ {
+ "epoch": 24.13961605584642,
+ "grad_norm": 0.056572459638118744,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.9076738357543945,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153577661431065,
+ "grad_norm": 0.05553183704614639,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.933659076690674,
+ "step": 1739
+ },
+ {
+ "epoch": 24.167539267015705,
+ "grad_norm": 0.05554374307394028,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.941751480102539,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18150087260035,
+ "grad_norm": 0.05451614409685135,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.928593635559082,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195462478184993,
+ "grad_norm": 0.055458780378103256,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 3.961820602416992,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209424083769633,
+ "grad_norm": 0.054980456829071045,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 3.9736528396606445,
+ "step": 1743
+ },
+ {
+ "epoch": 24.223385689354277,
+ "grad_norm": 0.05538604408502579,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.935093641281128,
+ "step": 1744
+ },
+ {
+ "epoch": 24.237347294938917,
+ "grad_norm": 0.05580678582191467,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.9240689277648926,
+ "step": 1745
+ },
+ {
+ "epoch": 24.25130890052356,
+ "grad_norm": 0.052294690161943436,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.937718152999878,
+ "step": 1746
+ },
+ {
+ "epoch": 24.2652705061082,
+ "grad_norm": 0.052018795162439346,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.942049503326416,
+ "step": 1747
+ },
+ {
+ "epoch": 24.279232111692846,
+ "grad_norm": 0.0550091527402401,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.9215564727783203,
+ "step": 1748
+ },
+ {
+ "epoch": 24.293193717277486,
+ "grad_norm": 0.05447716638445854,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.9154953956604004,
+ "step": 1749
+ },
+ {
+ "epoch": 24.30715532286213,
+ "grad_norm": 0.05300029739737511,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 3.9642622470855713,
+ "step": 1750
+ },
+ {
+ "epoch": 24.32111692844677,
+ "grad_norm": 0.05312013998627663,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.938897132873535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335078534031414,
+ "grad_norm": 0.053737837821245193,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.94972562789917,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349040139616054,
+ "grad_norm": 0.05087469890713692,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.9301164150238037,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363001745200698,
+ "grad_norm": 0.05179893225431442,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.9393205642700195,
+ "step": 1754
+ },
+ {
+ "epoch": 24.376963350785342,
+ "grad_norm": 0.05277438461780548,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 3.960968494415283,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390924956369982,
+ "grad_norm": 0.05090474337339401,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 3.9516637325286865,
+ "step": 1756
+ },
+ {
+ "epoch": 24.404886561954626,
+ "grad_norm": 0.052786506712436676,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.9332518577575684,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418848167539267,
+ "grad_norm": 0.052095942199230194,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.9108662605285645,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43280977312391,
+ "grad_norm": 0.05310450494289398,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 3.9746041297912598,
+ "step": 1759
+ },
+ {
+ "epoch": 24.44677137870855,
+ "grad_norm": 0.05078566074371338,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.93316650390625,
+ "step": 1760
+ },
+ {
+ "epoch": 24.460732984293195,
+ "grad_norm": 0.0519949309527874,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.9489870071411133,
+ "step": 1761
+ },
+ {
+ "epoch": 24.474694589877835,
+ "grad_norm": 0.05239349976181984,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 3.955118417739868,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48865619546248,
+ "grad_norm": 0.05193360522389412,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 3.9707555770874023,
+ "step": 1763
+ },
+ {
+ "epoch": 24.50261780104712,
+ "grad_norm": 0.05397362262010574,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 3.973053455352783,
+ "step": 1764
+ },
+ {
+ "epoch": 24.516579406631763,
+ "grad_norm": 0.05379117652773857,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 3.957289695739746,
+ "step": 1765
+ },
+ {
+ "epoch": 24.530541012216403,
+ "grad_norm": 0.0526256300508976,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 3.9595577716827393,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544502617801047,
+ "grad_norm": 0.05377693101763725,
+ "learning_rate": 0.000325372061241796,
+ "loss": 3.97849178314209,
+ "step": 1767
+ },
+ {
+ "epoch": 24.55846422338569,
+ "grad_norm": 0.05182809755206108,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 3.9714043140411377,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57242582897033,
+ "grad_norm": 0.05364793911576271,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 3.9582200050354004,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586387434554975,
+ "grad_norm": 0.051558300852775574,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 3.9747936725616455,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600349040139616,
+ "grad_norm": 0.053487639874219894,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 3.9556500911712646,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61431064572426,
+ "grad_norm": 0.05146970972418785,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.94858455657959,
+ "step": 1772
+ },
+ {
+ "epoch": 24.6282722513089,
+ "grad_norm": 0.05099836364388466,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 3.9716954231262207,
+ "step": 1773
+ },
+ {
+ "epoch": 24.642233856893544,
+ "grad_norm": 0.054634734988212585,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 3.974881887435913,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656195462478184,
+ "grad_norm": 0.054601747542619705,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 3.9746241569519043,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670157068062828,
+ "grad_norm": 0.052188221365213394,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 3.957106113433838,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68411867364747,
+ "grad_norm": 0.055673446506261826,
+ "learning_rate": 0.000322712903929477,
+ "loss": 3.9803967475891113,
+ "step": 1777
+ },
+ {
+ "epoch": 24.698080279232112,
+ "grad_norm": 0.05397328734397888,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 3.967238426208496,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712041884816752,
+ "grad_norm": 0.05384642630815506,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 3.949392795562744,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726003490401396,
+ "grad_norm": 0.05452995374798775,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.9432482719421387,
+ "step": 1780
+ },
+ {
+ "epoch": 24.739965095986037,
+ "grad_norm": 0.052430473268032074,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 3.9944653511047363,
+ "step": 1781
+ },
+ {
+ "epoch": 24.75392670157068,
+ "grad_norm": 0.05514240264892578,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 3.97713565826416,
+ "step": 1782
+ },
+ {
+ "epoch": 24.767888307155324,
+ "grad_norm": 0.05189919099211693,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 3.983150005340576,
+ "step": 1783
+ },
+ {
+ "epoch": 24.781849912739965,
+ "grad_norm": 0.05590776726603508,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 3.9548301696777344,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79581151832461,
+ "grad_norm": 0.05564190074801445,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 3.9725170135498047,
+ "step": 1785
+ },
+ {
+ "epoch": 24.80977312390925,
+ "grad_norm": 0.05121326446533203,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 3.9722347259521484,
+ "step": 1786
+ },
+ {
+ "epoch": 24.823734729493893,
+ "grad_norm": 0.05444202199578285,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 3.9634029865264893,
+ "step": 1787
+ },
+ {
+ "epoch": 24.837696335078533,
+ "grad_norm": 0.05242442339658737,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 3.973270893096924,
+ "step": 1788
+ },
+ {
+ "epoch": 24.851657940663177,
+ "grad_norm": 0.05497463420033455,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 3.9851887226104736,
+ "step": 1789
+ },
+ {
+ "epoch": 24.865619546247817,
+ "grad_norm": 0.05482025071978569,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 4.001110076904297,
+ "step": 1790
+ },
+ {
+ "epoch": 24.87958115183246,
+ "grad_norm": 0.054101862013339996,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 3.9782891273498535,
+ "step": 1791
+ },
+ {
+ "epoch": 24.8935427574171,
+ "grad_norm": 0.05538944527506828,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 3.973541498184204,
+ "step": 1792
+ },
+ {
+ "epoch": 24.907504363001745,
+ "grad_norm": 0.0502893440425396,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 3.9875426292419434,
+ "step": 1793
+ },
+ {
+ "epoch": 24.921465968586386,
+ "grad_norm": 0.05576152354478836,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.947035551071167,
+ "step": 1794
+ },
+ {
+ "epoch": 24.93542757417103,
+ "grad_norm": 0.05356278270483017,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 3.9683666229248047,
+ "step": 1795
+ },
+ {
+ "epoch": 24.949389179755673,
+ "grad_norm": 0.05377693101763725,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 3.9973440170288086,
+ "step": 1796
+ },
+ {
+ "epoch": 24.963350785340314,
+ "grad_norm": 0.05334382876753807,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 3.9843404293060303,
+ "step": 1797
+ },
+ {
+ "epoch": 24.977312390924958,
+ "grad_norm": 0.05514024198055267,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 3.9920568466186523,
+ "step": 1798
+ },
+ {
+ "epoch": 24.991273996509598,
+ "grad_norm": 0.05631054937839508,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 4.005859375,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.03653077408671379,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 2.503864049911499,
+ "step": 1800
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 7.680735615909888e+17,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-1800/training_args.bin b/runs/l2r90-wd0033-ssl/checkpoint-1800/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..313c41733ff55284c06034a59bb885b30e40ffc8
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-1800/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7ddd0bdddba68d3e19ab33477000f3fc684fe4ee8e3580eeb8c02646e5fc27ac
+size 4856
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2160/chat_template.jinja b/runs/l2r90-wd0033-ssl/checkpoint-2160/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2160/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2160/config.json b/runs/l2r90-wd0033-ssl/checkpoint-2160/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2160/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2160/generation_config.json b/runs/l2r90-wd0033-ssl/checkpoint-2160/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2160/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2160/model.safetensors b/runs/l2r90-wd0033-ssl/checkpoint-2160/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..7cd0333054c53313f8f2ef23205d4e52329a4557
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2160/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a33f6385687c8f3eb38d32372970ac1d88e2aa66ee1bd9cf69ed6d6eac826460
+size 583356232
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2160/optimizer.pt b/runs/l2r90-wd0033-ssl/checkpoint-2160/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..bd17a7801308a99a3cc3def64ba28e3634e6583a
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2160/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:21eb8eeaafa61acae85bffbe8247bb9118bde52837c56ab7366cb51efa656d1f
+size 1166827898
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2160/rng_state_0.pth b/runs/l2r90-wd0033-ssl/checkpoint-2160/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..f7b0f80bddd9d1575a028e4ed08ac5261f9a78cc
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2160/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:36272cc63a644540024f08ffb529f51f3af1b1c7c616e1e8762def8137bd7cc5
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2160/rng_state_1.pth b/runs/l2r90-wd0033-ssl/checkpoint-2160/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..7b62eb37975baa72a3c8fce7a6ede595f7ae575b
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2160/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f87b906c8d2b07e01c0df3c22ac835949a28469a6ca239354ae46dbeab434b1f
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2160/scheduler.pt b/runs/l2r90-wd0033-ssl/checkpoint-2160/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..35aefe7762df83336d8bfddab210c09c85e6b3aa
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2160/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a7d6a0a5c94f665a32f55d7b034b4fc829f9f44c468d851ff5b24b22d8388e63
+size 1064
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2160/tokenizer.json b/runs/l2r90-wd0033-ssl/checkpoint-2160/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2160/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2160/tokenizer_config.json b/runs/l2r90-wd0033-ssl/checkpoint-2160/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2160/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2160/trainer_state.json b/runs/l2r90-wd0033-ssl/checkpoint-2160/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..3fd62602e142ac609e037d83676a4c131ed34e30
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2160/trainer_state.json
@@ -0,0 +1,15378 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 30.0,
+ "eval_steps": 500,
+ "global_step": 2160,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013961605584642234,
+ "grad_norm": 0.26733073592185974,
+ "learning_rate": 0.0,
+ "loss": 7.990724086761475,
+ "step": 1
+ },
+ {
+ "epoch": 0.027923211169284468,
+ "grad_norm": 0.26712340116500854,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 7.990110874176025,
+ "step": 2
+ },
+ {
+ "epoch": 0.041884816753926704,
+ "grad_norm": 0.2646932601928711,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 7.956935405731201,
+ "step": 3
+ },
+ {
+ "epoch": 0.055846422338568937,
+ "grad_norm": 0.24684545397758484,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 7.890539646148682,
+ "step": 4
+ },
+ {
+ "epoch": 0.06980802792321117,
+ "grad_norm": 0.21016643941402435,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 7.8146138191223145,
+ "step": 5
+ },
+ {
+ "epoch": 0.08376963350785341,
+ "grad_norm": 0.17992742359638214,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 7.724307060241699,
+ "step": 6
+ },
+ {
+ "epoch": 0.09773123909249563,
+ "grad_norm": 0.14356893301010132,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 7.645662307739258,
+ "step": 7
+ },
+ {
+ "epoch": 0.11169284467713787,
+ "grad_norm": 0.10830964893102646,
+ "learning_rate": 4.2e-05,
+ "loss": 7.582019329071045,
+ "step": 8
+ },
+ {
+ "epoch": 0.1256544502617801,
+ "grad_norm": 0.08265161514282227,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 7.5230793952941895,
+ "step": 9
+ },
+ {
+ "epoch": 0.13961605584642234,
+ "grad_norm": 0.06423710286617279,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 7.473812103271484,
+ "step": 10
+ },
+ {
+ "epoch": 0.15357766143106458,
+ "grad_norm": 0.05132246017456055,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 7.440786838531494,
+ "step": 11
+ },
+ {
+ "epoch": 0.16753926701570682,
+ "grad_norm": 0.040572404861450195,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 7.41290283203125,
+ "step": 12
+ },
+ {
+ "epoch": 0.18150087260034903,
+ "grad_norm": 0.03138705715537071,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 7.39749002456665,
+ "step": 13
+ },
+ {
+ "epoch": 0.19546247818499127,
+ "grad_norm": 0.02553008496761322,
+ "learning_rate": 7.8e-05,
+ "loss": 7.387092590332031,
+ "step": 14
+ },
+ {
+ "epoch": 0.2094240837696335,
+ "grad_norm": 0.022178977727890015,
+ "learning_rate": 8.4e-05,
+ "loss": 7.378993511199951,
+ "step": 15
+ },
+ {
+ "epoch": 0.22338568935427575,
+ "grad_norm": 0.019915293902158737,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 7.37220573425293,
+ "step": 16
+ },
+ {
+ "epoch": 0.23734729493891799,
+ "grad_norm": 0.017390629276633263,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 7.366759300231934,
+ "step": 17
+ },
+ {
+ "epoch": 0.2513089005235602,
+ "grad_norm": 0.015472771599888802,
+ "learning_rate": 0.000102,
+ "loss": 7.3652663230896,
+ "step": 18
+ },
+ {
+ "epoch": 0.26527050610820246,
+ "grad_norm": 0.013599214144051075,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 7.359714984893799,
+ "step": 19
+ },
+ {
+ "epoch": 0.2792321116928447,
+ "grad_norm": 0.012442861683666706,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 7.353302001953125,
+ "step": 20
+ },
+ {
+ "epoch": 0.2931937172774869,
+ "grad_norm": 0.010471965186297894,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 7.357417583465576,
+ "step": 21
+ },
+ {
+ "epoch": 0.30715532286212915,
+ "grad_norm": 0.009647224098443985,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 7.351763725280762,
+ "step": 22
+ },
+ {
+ "epoch": 0.32111692844677137,
+ "grad_norm": 0.008228210732340813,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 7.345434188842773,
+ "step": 23
+ },
+ {
+ "epoch": 0.33507853403141363,
+ "grad_norm": 0.007454514969140291,
+ "learning_rate": 0.000138,
+ "loss": 7.342520236968994,
+ "step": 24
+ },
+ {
+ "epoch": 0.34904013961605584,
+ "grad_norm": 0.0069856527261435986,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 7.338870048522949,
+ "step": 25
+ },
+ {
+ "epoch": 0.36300174520069806,
+ "grad_norm": 0.006091665010899305,
+ "learning_rate": 0.00015,
+ "loss": 7.328121662139893,
+ "step": 26
+ },
+ {
+ "epoch": 0.3769633507853403,
+ "grad_norm": 0.005525320768356323,
+ "learning_rate": 0.000156,
+ "loss": 7.335029602050781,
+ "step": 27
+ },
+ {
+ "epoch": 0.39092495636998253,
+ "grad_norm": 0.0061133443377912045,
+ "learning_rate": 0.000162,
+ "loss": 7.319417953491211,
+ "step": 28
+ },
+ {
+ "epoch": 0.4048865619546248,
+ "grad_norm": 0.006040909793227911,
+ "learning_rate": 0.000168,
+ "loss": 7.326376438140869,
+ "step": 29
+ },
+ {
+ "epoch": 0.418848167539267,
+ "grad_norm": 0.006003216840326786,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 7.319552421569824,
+ "step": 30
+ },
+ {
+ "epoch": 0.4328097731239092,
+ "grad_norm": 0.007096003741025925,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 7.323486328125,
+ "step": 31
+ },
+ {
+ "epoch": 0.4467713787085515,
+ "grad_norm": 0.007725459989160299,
+ "learning_rate": 0.000186,
+ "loss": 7.3190717697143555,
+ "step": 32
+ },
+ {
+ "epoch": 0.4607329842931937,
+ "grad_norm": 0.007338670082390308,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 7.317154884338379,
+ "step": 33
+ },
+ {
+ "epoch": 0.47469458987783597,
+ "grad_norm": 0.007709544617682695,
+ "learning_rate": 0.000198,
+ "loss": 7.311826705932617,
+ "step": 34
+ },
+ {
+ "epoch": 0.4886561954624782,
+ "grad_norm": 0.009249784983694553,
+ "learning_rate": 0.000204,
+ "loss": 7.322498798370361,
+ "step": 35
+ },
+ {
+ "epoch": 0.5026178010471204,
+ "grad_norm": 0.006950767710804939,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 7.314571380615234,
+ "step": 36
+ },
+ {
+ "epoch": 0.5165794066317626,
+ "grad_norm": 0.007583985570818186,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 7.309540748596191,
+ "step": 37
+ },
+ {
+ "epoch": 0.5305410122164049,
+ "grad_norm": 0.014731820672750473,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 7.307004928588867,
+ "step": 38
+ },
+ {
+ "epoch": 0.5445026178010471,
+ "grad_norm": 0.016276473179459572,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 7.299642562866211,
+ "step": 39
+ },
+ {
+ "epoch": 0.5584642233856894,
+ "grad_norm": 0.020018689334392548,
+ "learning_rate": 0.000234,
+ "loss": 7.298466682434082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5724258289703316,
+ "grad_norm": 0.014941536821424961,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 7.289968967437744,
+ "step": 41
+ },
+ {
+ "epoch": 0.5863874345549738,
+ "grad_norm": 0.018299812451004982,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 7.276564598083496,
+ "step": 42
+ },
+ {
+ "epoch": 0.6003490401396161,
+ "grad_norm": 0.01615075021982193,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.2575554847717285,
+ "step": 43
+ },
+ {
+ "epoch": 0.6143106457242583,
+ "grad_norm": 0.013063831254839897,
+ "learning_rate": 0.000258,
+ "loss": 7.253885269165039,
+ "step": 44
+ },
+ {
+ "epoch": 0.6282722513089005,
+ "grad_norm": 0.015211676247417927,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.245779037475586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6422338568935427,
+ "grad_norm": 0.02093500830233097,
+ "learning_rate": 0.00027,
+ "loss": 7.225130081176758,
+ "step": 46
+ },
+ {
+ "epoch": 0.6561954624781849,
+ "grad_norm": 0.01808053068816662,
+ "learning_rate": 0.000276,
+ "loss": 7.209654331207275,
+ "step": 47
+ },
+ {
+ "epoch": 0.6701570680628273,
+ "grad_norm": 0.02209644205868244,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.197124481201172,
+ "step": 48
+ },
+ {
+ "epoch": 0.6841186736474695,
+ "grad_norm": 0.023524878546595573,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.1841936111450195,
+ "step": 49
+ },
+ {
+ "epoch": 0.6980802792321117,
+ "grad_norm": 0.031148087233304977,
+ "learning_rate": 0.000294,
+ "loss": 7.188040733337402,
+ "step": 50
+ },
+ {
+ "epoch": 0.7120418848167539,
+ "grad_norm": 0.02931929938495159,
+ "learning_rate": 0.0003,
+ "loss": 7.17371940612793,
+ "step": 51
+ },
+ {
+ "epoch": 0.7260034904013961,
+ "grad_norm": 0.01611483097076416,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.1714324951171875,
+ "step": 52
+ },
+ {
+ "epoch": 0.7399650959860384,
+ "grad_norm": 0.02619907818734646,
+ "learning_rate": 0.000312,
+ "loss": 7.158261299133301,
+ "step": 53
+ },
+ {
+ "epoch": 0.7539267015706806,
+ "grad_norm": 0.049462560564279556,
+ "learning_rate": 0.000318,
+ "loss": 7.167076110839844,
+ "step": 54
+ },
+ {
+ "epoch": 0.7678883071553229,
+ "grad_norm": 0.09835070371627808,
+ "learning_rate": 0.000324,
+ "loss": 7.154292106628418,
+ "step": 55
+ },
+ {
+ "epoch": 0.7818499127399651,
+ "grad_norm": 0.11406510323286057,
+ "learning_rate": 0.00033,
+ "loss": 7.162016868591309,
+ "step": 56
+ },
+ {
+ "epoch": 0.7958115183246073,
+ "grad_norm": 0.027026303112506866,
+ "learning_rate": 0.000336,
+ "loss": 7.141416549682617,
+ "step": 57
+ },
+ {
+ "epoch": 0.8097731239092496,
+ "grad_norm": 0.06690579652786255,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.13959264755249,
+ "step": 58
+ },
+ {
+ "epoch": 0.8237347294938918,
+ "grad_norm": 0.020919431000947952,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.126513481140137,
+ "step": 59
+ },
+ {
+ "epoch": 0.837696335078534,
+ "grad_norm": 0.05618130415678024,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.124345779418945,
+ "step": 60
+ },
+ {
+ "epoch": 0.8516579406631762,
+ "grad_norm": 0.03402302414178848,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.100615501403809,
+ "step": 61
+ },
+ {
+ "epoch": 0.8656195462478184,
+ "grad_norm": 0.03646906837821007,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.116507530212402,
+ "step": 62
+ },
+ {
+ "epoch": 0.8795811518324608,
+ "grad_norm": 0.03076397068798542,
+ "learning_rate": 0.000372,
+ "loss": 7.097659111022949,
+ "step": 63
+ },
+ {
+ "epoch": 0.893542757417103,
+ "grad_norm": 0.03329646959900856,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.097290992736816,
+ "step": 64
+ },
+ {
+ "epoch": 0.9075043630017452,
+ "grad_norm": 0.01998630352318287,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.092268943786621,
+ "step": 65
+ },
+ {
+ "epoch": 0.9214659685863874,
+ "grad_norm": 0.03278940171003342,
+ "learning_rate": 0.00039,
+ "loss": 7.088292121887207,
+ "step": 66
+ },
+ {
+ "epoch": 0.9354275741710296,
+ "grad_norm": 0.024620421230793,
+ "learning_rate": 0.000396,
+ "loss": 7.078650951385498,
+ "step": 67
+ },
+ {
+ "epoch": 0.9493891797556719,
+ "grad_norm": 0.03750383481383324,
+ "learning_rate": 0.000402,
+ "loss": 7.052778244018555,
+ "step": 68
+ },
+ {
+ "epoch": 0.9633507853403142,
+ "grad_norm": 0.07558075338602066,
+ "learning_rate": 0.000408,
+ "loss": 7.045991897583008,
+ "step": 69
+ },
+ {
+ "epoch": 0.9773123909249564,
+ "grad_norm": 0.16021482646465302,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.023152828216553,
+ "step": 70
+ },
+ {
+ "epoch": 0.9912739965095986,
+ "grad_norm": 0.24981893599033356,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.067540645599365,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.10255520790815353,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 4.425640106201172,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8861758708953857,
+ "eval_runtime": 58.2699,
+ "eval_samples_per_second": 41.908,
+ "eval_steps_per_second": 0.669,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139616055846423,
+ "grad_norm": 0.13374797999858856,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.073373317718506,
+ "step": 73
+ },
+ {
+ "epoch": 1.0279232111692844,
+ "grad_norm": 0.05079387128353119,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.0243306159973145,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418848167539267,
+ "grad_norm": 0.10056068748235703,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.014476299285889,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558464223385688,
+ "grad_norm": 0.06995758414268494,
+ "learning_rate": 0.00045,
+ "loss": 7.011931419372559,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698080279232112,
+ "grad_norm": 0.0483892522752285,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 6.9822797775268555,
+ "step": 77
+ },
+ {
+ "epoch": 1.0837696335078535,
+ "grad_norm": 0.08408170193433762,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 6.964348793029785,
+ "step": 78
+ },
+ {
+ "epoch": 1.0977312390924956,
+ "grad_norm": 0.08027710020542145,
+ "learning_rate": 0.000468,
+ "loss": 6.96852445602417,
+ "step": 79
+ },
+ {
+ "epoch": 1.111692844677138,
+ "grad_norm": 0.05647159367799759,
+ "learning_rate": 0.000474,
+ "loss": 6.944484710693359,
+ "step": 80
+ },
+ {
+ "epoch": 1.12565445026178,
+ "grad_norm": 0.07464330643415451,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 6.9246745109558105,
+ "step": 81
+ },
+ {
+ "epoch": 1.1396160558464223,
+ "grad_norm": 0.04451674595475197,
+ "learning_rate": 0.000486,
+ "loss": 6.895403861999512,
+ "step": 82
+ },
+ {
+ "epoch": 1.1535776614310647,
+ "grad_norm": 0.05820256471633911,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 6.887338638305664,
+ "step": 83
+ },
+ {
+ "epoch": 1.1675392670157068,
+ "grad_norm": 0.02925296127796173,
+ "learning_rate": 0.000498,
+ "loss": 6.857797622680664,
+ "step": 84
+ },
+ {
+ "epoch": 1.181500872600349,
+ "grad_norm": 0.05096987634897232,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 6.854822158813477,
+ "step": 85
+ },
+ {
+ "epoch": 1.1954624781849912,
+ "grad_norm": 0.03483197093009949,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 6.840914726257324,
+ "step": 86
+ },
+ {
+ "epoch": 1.2094240837696335,
+ "grad_norm": 0.03323264792561531,
+ "learning_rate": 0.000516,
+ "loss": 6.834497451782227,
+ "step": 87
+ },
+ {
+ "epoch": 1.2233856893542758,
+ "grad_norm": 0.03290868178009987,
+ "learning_rate": 0.000522,
+ "loss": 6.820746898651123,
+ "step": 88
+ },
+ {
+ "epoch": 1.237347294938918,
+ "grad_norm": 0.025398844853043556,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 6.795746803283691,
+ "step": 89
+ },
+ {
+ "epoch": 1.2513089005235603,
+ "grad_norm": 0.030700596049427986,
+ "learning_rate": 0.000534,
+ "loss": 6.7867889404296875,
+ "step": 90
+ },
+ {
+ "epoch": 1.2652705061082026,
+ "grad_norm": 0.019589420408010483,
+ "learning_rate": 0.00054,
+ "loss": 6.77443790435791,
+ "step": 91
+ },
+ {
+ "epoch": 1.2792321116928447,
+ "grad_norm": 0.03313920646905899,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 6.762271881103516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2931937172774868,
+ "grad_norm": 0.025389468297362328,
+ "learning_rate": 0.000552,
+ "loss": 6.750258445739746,
+ "step": 93
+ },
+ {
+ "epoch": 1.307155322862129,
+ "grad_norm": 0.040296655148267746,
+ "learning_rate": 0.000558,
+ "loss": 6.729808807373047,
+ "step": 94
+ },
+ {
+ "epoch": 1.3211169284467714,
+ "grad_norm": 0.04735933244228363,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 6.737982749938965,
+ "step": 95
+ },
+ {
+ "epoch": 1.3350785340314135,
+ "grad_norm": 0.05529680848121643,
+ "learning_rate": 0.00057,
+ "loss": 6.735714912414551,
+ "step": 96
+ },
+ {
+ "epoch": 1.3490401396160558,
+ "grad_norm": 0.07148554176092148,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 6.708859920501709,
+ "step": 97
+ },
+ {
+ "epoch": 1.3630017452006982,
+ "grad_norm": 0.059708479791879654,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 6.708860397338867,
+ "step": 98
+ },
+ {
+ "epoch": 1.3769633507853403,
+ "grad_norm": 0.027111195027828217,
+ "learning_rate": 0.000588,
+ "loss": 6.686685085296631,
+ "step": 99
+ },
+ {
+ "epoch": 1.3909249563699826,
+ "grad_norm": 0.0500967763364315,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 6.676325798034668,
+ "step": 100
+ },
+ {
+ "epoch": 1.404886561954625,
+ "grad_norm": 0.044565796852111816,
+ "learning_rate": 0.0006,
+ "loss": 6.674530029296875,
+ "step": 101
+ },
+ {
+ "epoch": 1.418848167539267,
+ "grad_norm": 0.03608180209994316,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 6.653741359710693,
+ "step": 102
+ },
+ {
+ "epoch": 1.4328097731239091,
+ "grad_norm": 0.030394982546567917,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 6.643152236938477,
+ "step": 103
+ },
+ {
+ "epoch": 1.4467713787085514,
+ "grad_norm": 0.040185656398534775,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 6.628686904907227,
+ "step": 104
+ },
+ {
+ "epoch": 1.4607329842931938,
+ "grad_norm": 0.021049322560429573,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 6.610864639282227,
+ "step": 105
+ },
+ {
+ "epoch": 1.4746945898778359,
+ "grad_norm": 0.04164277762174606,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 6.602880954742432,
+ "step": 106
+ },
+ {
+ "epoch": 1.4886561954624782,
+ "grad_norm": 0.0389864556491375,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 6.599456787109375,
+ "step": 107
+ },
+ {
+ "epoch": 1.5026178010471205,
+ "grad_norm": 0.05805973336100578,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 6.588700294494629,
+ "step": 108
+ },
+ {
+ "epoch": 1.5165794066317626,
+ "grad_norm": 0.08825036883354187,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 6.587496757507324,
+ "step": 109
+ },
+ {
+ "epoch": 1.530541012216405,
+ "grad_norm": 0.12521523237228394,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 6.626501083374023,
+ "step": 110
+ },
+ {
+ "epoch": 1.5445026178010473,
+ "grad_norm": 0.07429645210504532,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 6.590241432189941,
+ "step": 111
+ },
+ {
+ "epoch": 1.5584642233856894,
+ "grad_norm": 0.06463009119033813,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 6.578714370727539,
+ "step": 112
+ },
+ {
+ "epoch": 1.5724258289703315,
+ "grad_norm": 0.07508203387260437,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 6.561270713806152,
+ "step": 113
+ },
+ {
+ "epoch": 1.5863874345549738,
+ "grad_norm": 0.04125197231769562,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 6.558821678161621,
+ "step": 114
+ },
+ {
+ "epoch": 1.600349040139616,
+ "grad_norm": 0.05457386001944542,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.541611671447754,
+ "step": 115
+ },
+ {
+ "epoch": 1.6143106457242582,
+ "grad_norm": 0.03684385493397713,
+ "learning_rate": 0.000599973080557108,
+ "loss": 6.547738075256348,
+ "step": 116
+ },
+ {
+ "epoch": 1.6282722513089005,
+ "grad_norm": 0.05888746306300163,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.529491424560547,
+ "step": 117
+ },
+ {
+ "epoch": 1.6422338568935428,
+ "grad_norm": 0.056668370962142944,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.519444942474365,
+ "step": 118
+ },
+ {
+ "epoch": 1.656195462478185,
+ "grad_norm": 0.09123896062374115,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.522984981536865,
+ "step": 119
+ },
+ {
+ "epoch": 1.6701570680628273,
+ "grad_norm": 0.12380023300647736,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.515375137329102,
+ "step": 120
+ },
+ {
+ "epoch": 1.6841186736474696,
+ "grad_norm": 0.06953322887420654,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.500491142272949,
+ "step": 121
+ },
+ {
+ "epoch": 1.6980802792321117,
+ "grad_norm": 0.08613869547843933,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.5097198486328125,
+ "step": 122
+ },
+ {
+ "epoch": 1.7120418848167538,
+ "grad_norm": 0.08296383172273636,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.487521171569824,
+ "step": 123
+ },
+ {
+ "epoch": 1.7260034904013961,
+ "grad_norm": 0.05744783207774162,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.485134124755859,
+ "step": 124
+ },
+ {
+ "epoch": 1.7399650959860384,
+ "grad_norm": 0.04108566418290138,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.466182708740234,
+ "step": 125
+ },
+ {
+ "epoch": 1.7539267015706805,
+ "grad_norm": 0.05374931916594505,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.479011535644531,
+ "step": 126
+ },
+ {
+ "epoch": 1.7678883071553229,
+ "grad_norm": 0.03867914527654648,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.465902328491211,
+ "step": 127
+ },
+ {
+ "epoch": 1.7818499127399652,
+ "grad_norm": 0.05477374792098999,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.453082084655762,
+ "step": 128
+ },
+ {
+ "epoch": 1.7958115183246073,
+ "grad_norm": 0.06508298963308334,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.446089744567871,
+ "step": 129
+ },
+ {
+ "epoch": 1.8097731239092496,
+ "grad_norm": 0.06388624757528305,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.451748847961426,
+ "step": 130
+ },
+ {
+ "epoch": 1.823734729493892,
+ "grad_norm": 0.04223188757896423,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.4492387771606445,
+ "step": 131
+ },
+ {
+ "epoch": 1.837696335078534,
+ "grad_norm": 0.04528658092021942,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.409390926361084,
+ "step": 132
+ },
+ {
+ "epoch": 1.8516579406631761,
+ "grad_norm": 0.04962967336177826,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.415774345397949,
+ "step": 133
+ },
+ {
+ "epoch": 1.8656195462478184,
+ "grad_norm": 0.031944390386343,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.411528587341309,
+ "step": 134
+ },
+ {
+ "epoch": 1.8795811518324608,
+ "grad_norm": 0.028411149978637695,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.392329216003418,
+ "step": 135
+ },
+ {
+ "epoch": 1.8935427574171029,
+ "grad_norm": 0.03516148403286934,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.392238616943359,
+ "step": 136
+ },
+ {
+ "epoch": 1.9075043630017452,
+ "grad_norm": 0.027800120413303375,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.39179801940918,
+ "step": 137
+ },
+ {
+ "epoch": 1.9214659685863875,
+ "grad_norm": 0.02362987957894802,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.378960609436035,
+ "step": 138
+ },
+ {
+ "epoch": 1.9354275741710296,
+ "grad_norm": 0.033999308943748474,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.36802339553833,
+ "step": 139
+ },
+ {
+ "epoch": 1.949389179755672,
+ "grad_norm": 0.03290879726409912,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.366308689117432,
+ "step": 140
+ },
+ {
+ "epoch": 1.9633507853403143,
+ "grad_norm": 0.0286843404173851,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.350399971008301,
+ "step": 141
+ },
+ {
+ "epoch": 1.9773123909249564,
+ "grad_norm": 0.027145877480506897,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.33477783203125,
+ "step": 142
+ },
+ {
+ "epoch": 1.9912739965095985,
+ "grad_norm": 0.03530878946185112,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.322859764099121,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.028133150190114975,
+ "learning_rate": 0.000599778806120077,
+ "loss": 3.955288887023926,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7927101254463196,
+ "eval_runtime": 60.8372,
+ "eval_samples_per_second": 40.14,
+ "eval_steps_per_second": 0.641,
+ "step": 144
+ },
+ {
+ "epoch": 2.013961605584642,
+ "grad_norm": 0.03054557554423809,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.311975479125977,
+ "step": 145
+ },
+ {
+ "epoch": 2.0279232111692846,
+ "grad_norm": 0.05406459793448448,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.3030853271484375,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418848167539267,
+ "grad_norm": 0.10911019891500473,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.329056739807129,
+ "step": 147
+ },
+ {
+ "epoch": 2.055846422338569,
+ "grad_norm": 0.13948315382003784,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.362421035766602,
+ "step": 148
+ },
+ {
+ "epoch": 2.0698080279232114,
+ "grad_norm": 0.13176991045475006,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.3888139724731445,
+ "step": 149
+ },
+ {
+ "epoch": 2.0837696335078535,
+ "grad_norm": 0.09615885466337204,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.35154914855957,
+ "step": 150
+ },
+ {
+ "epoch": 2.0977312390924956,
+ "grad_norm": 0.06462561339139938,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.335790634155273,
+ "step": 151
+ },
+ {
+ "epoch": 2.1116928446771377,
+ "grad_norm": 0.05186621472239494,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.324396133422852,
+ "step": 152
+ },
+ {
+ "epoch": 2.1256544502617802,
+ "grad_norm": 0.06095981225371361,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.318549156188965,
+ "step": 153
+ },
+ {
+ "epoch": 2.1396160558464223,
+ "grad_norm": 0.036121442914009094,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.300607681274414,
+ "step": 154
+ },
+ {
+ "epoch": 2.1535776614310644,
+ "grad_norm": 0.055716633796691895,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.307400703430176,
+ "step": 155
+ },
+ {
+ "epoch": 2.167539267015707,
+ "grad_norm": 0.037626396864652634,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.287364959716797,
+ "step": 156
+ },
+ {
+ "epoch": 2.181500872600349,
+ "grad_norm": 0.04329098388552666,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.28244686126709,
+ "step": 157
+ },
+ {
+ "epoch": 2.195462478184991,
+ "grad_norm": 0.031868450343608856,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.2602410316467285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2094240837696333,
+ "grad_norm": 0.03647344186902046,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.263988971710205,
+ "step": 159
+ },
+ {
+ "epoch": 2.223385689354276,
+ "grad_norm": 0.03141765296459198,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.263205051422119,
+ "step": 160
+ },
+ {
+ "epoch": 2.237347294938918,
+ "grad_norm": 0.028253229334950447,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.253107070922852,
+ "step": 161
+ },
+ {
+ "epoch": 2.25130890052356,
+ "grad_norm": 0.03608971834182739,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.2571702003479,
+ "step": 162
+ },
+ {
+ "epoch": 2.2652705061082026,
+ "grad_norm": 0.02644292451441288,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.241732597351074,
+ "step": 163
+ },
+ {
+ "epoch": 2.2792321116928447,
+ "grad_norm": 0.028604455292224884,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.230289459228516,
+ "step": 164
+ },
+ {
+ "epoch": 2.2931937172774868,
+ "grad_norm": 0.023704614490270615,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.226101398468018,
+ "step": 165
+ },
+ {
+ "epoch": 2.3071553228621293,
+ "grad_norm": 0.020691027864813805,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.2096405029296875,
+ "step": 166
+ },
+ {
+ "epoch": 2.3211169284467714,
+ "grad_norm": 0.025598427280783653,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.22075080871582,
+ "step": 167
+ },
+ {
+ "epoch": 2.3350785340314135,
+ "grad_norm": 0.030546002089977264,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.216886520385742,
+ "step": 168
+ },
+ {
+ "epoch": 2.349040139616056,
+ "grad_norm": 0.026066245511174202,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.2022504806518555,
+ "step": 169
+ },
+ {
+ "epoch": 2.363001745200698,
+ "grad_norm": 0.023802239447832108,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.189509391784668,
+ "step": 170
+ },
+ {
+ "epoch": 2.3769633507853403,
+ "grad_norm": 0.029805002734065056,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.186277389526367,
+ "step": 171
+ },
+ {
+ "epoch": 2.3909249563699824,
+ "grad_norm": 0.043327830731868744,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.164876937866211,
+ "step": 172
+ },
+ {
+ "epoch": 2.404886561954625,
+ "grad_norm": 0.058787211775779724,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.179936408996582,
+ "step": 173
+ },
+ {
+ "epoch": 2.418848167539267,
+ "grad_norm": 0.0787397176027298,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.1830244064331055,
+ "step": 174
+ },
+ {
+ "epoch": 2.432809773123909,
+ "grad_norm": 0.0899854376912117,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.190343856811523,
+ "step": 175
+ },
+ {
+ "epoch": 2.4467713787085517,
+ "grad_norm": 0.06520707905292511,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.185057163238525,
+ "step": 176
+ },
+ {
+ "epoch": 2.4607329842931938,
+ "grad_norm": 0.1002504974603653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.171279430389404,
+ "step": 177
+ },
+ {
+ "epoch": 2.474694589877836,
+ "grad_norm": 0.09591004997491837,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.174382209777832,
+ "step": 178
+ },
+ {
+ "epoch": 2.488656195462478,
+ "grad_norm": 0.12127121537923813,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.185717582702637,
+ "step": 179
+ },
+ {
+ "epoch": 2.5026178010471205,
+ "grad_norm": 0.09640590846538544,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.196681976318359,
+ "step": 180
+ },
+ {
+ "epoch": 2.5165794066317626,
+ "grad_norm": 0.048341453075408936,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.158774375915527,
+ "step": 181
+ },
+ {
+ "epoch": 2.530541012216405,
+ "grad_norm": 0.08206334710121155,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.17445707321167,
+ "step": 182
+ },
+ {
+ "epoch": 2.5445026178010473,
+ "grad_norm": 0.038934968411922455,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.170527458190918,
+ "step": 183
+ },
+ {
+ "epoch": 2.5584642233856894,
+ "grad_norm": 0.060987457633018494,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.1501994132995605,
+ "step": 184
+ },
+ {
+ "epoch": 2.5724258289703315,
+ "grad_norm": 0.033269546926021576,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.144186019897461,
+ "step": 185
+ },
+ {
+ "epoch": 2.5863874345549736,
+ "grad_norm": 0.046137236058712006,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.1236677169799805,
+ "step": 186
+ },
+ {
+ "epoch": 2.600349040139616,
+ "grad_norm": 0.038265109062194824,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.121979236602783,
+ "step": 187
+ },
+ {
+ "epoch": 2.614310645724258,
+ "grad_norm": 0.039128001779317856,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.124156951904297,
+ "step": 188
+ },
+ {
+ "epoch": 2.6282722513089007,
+ "grad_norm": 0.025303540751338005,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.108036994934082,
+ "step": 189
+ },
+ {
+ "epoch": 2.642233856893543,
+ "grad_norm": 0.04441308230161667,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.1211161613464355,
+ "step": 190
+ },
+ {
+ "epoch": 2.656195462478185,
+ "grad_norm": 0.03083532303571701,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.094561576843262,
+ "step": 191
+ },
+ {
+ "epoch": 2.670157068062827,
+ "grad_norm": 0.02951296605169773,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.091608047485352,
+ "step": 192
+ },
+ {
+ "epoch": 2.6841186736474696,
+ "grad_norm": 0.02385987527668476,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.087080001831055,
+ "step": 193
+ },
+ {
+ "epoch": 2.6980802792321117,
+ "grad_norm": 0.03315780311822891,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.088132381439209,
+ "step": 194
+ },
+ {
+ "epoch": 2.712041884816754,
+ "grad_norm": 0.02410835586488247,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.081460952758789,
+ "step": 195
+ },
+ {
+ "epoch": 2.7260034904013963,
+ "grad_norm": 0.033069487661123276,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.0696539878845215,
+ "step": 196
+ },
+ {
+ "epoch": 2.7399650959860384,
+ "grad_norm": 0.031288471072912216,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.047612190246582,
+ "step": 197
+ },
+ {
+ "epoch": 2.7539267015706805,
+ "grad_norm": 0.03399677947163582,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.019026756286621,
+ "step": 198
+ },
+ {
+ "epoch": 2.7678883071553226,
+ "grad_norm": 0.031948331743478775,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.050133228302002,
+ "step": 199
+ },
+ {
+ "epoch": 2.781849912739965,
+ "grad_norm": 0.049206435680389404,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.045332431793213,
+ "step": 200
+ },
+ {
+ "epoch": 2.7958115183246073,
+ "grad_norm": 0.07528849691152573,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.059202194213867,
+ "step": 201
+ },
+ {
+ "epoch": 2.80977312390925,
+ "grad_norm": 0.11815470457077026,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.077486038208008,
+ "step": 202
+ },
+ {
+ "epoch": 2.823734729493892,
+ "grad_norm": 0.0867348462343216,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.045670032501221,
+ "step": 203
+ },
+ {
+ "epoch": 2.837696335078534,
+ "grad_norm": 0.05716705322265625,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.034373760223389,
+ "step": 204
+ },
+ {
+ "epoch": 2.851657940663176,
+ "grad_norm": 0.05944284051656723,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.049911975860596,
+ "step": 205
+ },
+ {
+ "epoch": 2.8656195462478182,
+ "grad_norm": 0.03220470994710922,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.039861679077148,
+ "step": 206
+ },
+ {
+ "epoch": 2.8795811518324608,
+ "grad_norm": 0.057081177830696106,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.010714530944824,
+ "step": 207
+ },
+ {
+ "epoch": 2.893542757417103,
+ "grad_norm": 0.04037516191601753,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.018877983093262,
+ "step": 208
+ },
+ {
+ "epoch": 2.9075043630017454,
+ "grad_norm": 0.036492880433797836,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.009493827819824,
+ "step": 209
+ },
+ {
+ "epoch": 2.9214659685863875,
+ "grad_norm": 0.04785023257136345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.0140862464904785,
+ "step": 210
+ },
+ {
+ "epoch": 2.9354275741710296,
+ "grad_norm": 0.07721036672592163,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.030359268188477,
+ "step": 211
+ },
+ {
+ "epoch": 2.9493891797556717,
+ "grad_norm": 0.14220190048217773,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.045989990234375,
+ "step": 212
+ },
+ {
+ "epoch": 2.9633507853403143,
+ "grad_norm": 0.10265158116817474,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.043968200683594,
+ "step": 213
+ },
+ {
+ "epoch": 2.9773123909249564,
+ "grad_norm": 0.051398951560258865,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.022278785705566,
+ "step": 214
+ },
+ {
+ "epoch": 2.9912739965095985,
+ "grad_norm": 0.06044618785381317,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.017296314239502,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.03284750506281853,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 3.7544798851013184,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7531244158744812,
+ "eval_runtime": 60.3986,
+ "eval_samples_per_second": 40.431,
+ "eval_steps_per_second": 0.646,
+ "step": 216
+ },
+ {
+ "epoch": 3.013961605584642,
+ "grad_norm": 0.03895840048789978,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 5.9985857009887695,
+ "step": 217
+ },
+ {
+ "epoch": 3.0279232111692846,
+ "grad_norm": 0.04554992541670799,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 5.984710693359375,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418848167539267,
+ "grad_norm": 0.03359116613864899,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 5.984078407287598,
+ "step": 219
+ },
+ {
+ "epoch": 3.055846422338569,
+ "grad_norm": 0.035171881318092346,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 5.979424476623535,
+ "step": 220
+ },
+ {
+ "epoch": 3.0698080279232114,
+ "grad_norm": 0.032774072140455246,
+ "learning_rate": 0.000598278794687186,
+ "loss": 5.94467830657959,
+ "step": 221
+ },
+ {
+ "epoch": 3.0837696335078535,
+ "grad_norm": 0.032842908054590225,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 5.95986270904541,
+ "step": 222
+ },
+ {
+ "epoch": 3.0977312390924956,
+ "grad_norm": 0.02870488539338112,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 5.951186656951904,
+ "step": 223
+ },
+ {
+ "epoch": 3.1116928446771377,
+ "grad_norm": 0.030777374282479286,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 5.942493438720703,
+ "step": 224
+ },
+ {
+ "epoch": 3.1256544502617802,
+ "grad_norm": 0.026439018547534943,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 5.938275337219238,
+ "step": 225
+ },
+ {
+ "epoch": 3.1396160558464223,
+ "grad_norm": 0.028689472004771233,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 5.931982040405273,
+ "step": 226
+ },
+ {
+ "epoch": 3.1535776614310644,
+ "grad_norm": 0.028100362047553062,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 5.919692039489746,
+ "step": 227
+ },
+ {
+ "epoch": 3.167539267015707,
+ "grad_norm": 0.02407177910208702,
+ "learning_rate": 0.00059807235435179,
+ "loss": 5.909029006958008,
+ "step": 228
+ },
+ {
+ "epoch": 3.181500872600349,
+ "grad_norm": 0.0237748846411705,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 5.9131598472595215,
+ "step": 229
+ },
+ {
+ "epoch": 3.195462478184991,
+ "grad_norm": 0.022418556734919548,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 5.909396171569824,
+ "step": 230
+ },
+ {
+ "epoch": 3.2094240837696333,
+ "grad_norm": 0.024656571447849274,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 5.885243892669678,
+ "step": 231
+ },
+ {
+ "epoch": 3.223385689354276,
+ "grad_norm": 0.02136118710041046,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 5.898958206176758,
+ "step": 232
+ },
+ {
+ "epoch": 3.237347294938918,
+ "grad_norm": 0.025290116667747498,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 5.886986255645752,
+ "step": 233
+ },
+ {
+ "epoch": 3.25130890052356,
+ "grad_norm": 0.02656404674053192,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 5.897647857666016,
+ "step": 234
+ },
+ {
+ "epoch": 3.2652705061082026,
+ "grad_norm": 0.019188793376088142,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 5.875659942626953,
+ "step": 235
+ },
+ {
+ "epoch": 3.2792321116928447,
+ "grad_norm": 0.02308926172554493,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 5.859380722045898,
+ "step": 236
+ },
+ {
+ "epoch": 3.2931937172774868,
+ "grad_norm": 0.029775725677609444,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 5.873961448669434,
+ "step": 237
+ },
+ {
+ "epoch": 3.3071553228621293,
+ "grad_norm": 0.03875259310007095,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 5.863171100616455,
+ "step": 238
+ },
+ {
+ "epoch": 3.3211169284467714,
+ "grad_norm": 0.052446842193603516,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 5.886547088623047,
+ "step": 239
+ },
+ {
+ "epoch": 3.3350785340314135,
+ "grad_norm": 0.05583837628364563,
+ "learning_rate": 0.000597691359743243,
+ "loss": 5.8849287033081055,
+ "step": 240
+ },
+ {
+ "epoch": 3.349040139616056,
+ "grad_norm": 0.034729067236185074,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 5.863419532775879,
+ "step": 241
+ },
+ {
+ "epoch": 3.363001745200698,
+ "grad_norm": 0.03590013459324837,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 5.852169990539551,
+ "step": 242
+ },
+ {
+ "epoch": 3.3769633507853403,
+ "grad_norm": 0.035041507333517075,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 5.836690902709961,
+ "step": 243
+ },
+ {
+ "epoch": 3.3909249563699824,
+ "grad_norm": 0.02857939340174198,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 5.836485862731934,
+ "step": 244
+ },
+ {
+ "epoch": 3.404886561954625,
+ "grad_norm": 0.03176642581820488,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 5.860719203948975,
+ "step": 245
+ },
+ {
+ "epoch": 3.418848167539267,
+ "grad_norm": 0.026105748489499092,
+ "learning_rate": 0.000597488038559129,
+ "loss": 5.8354949951171875,
+ "step": 246
+ },
+ {
+ "epoch": 3.432809773123909,
+ "grad_norm": 0.029094981029629707,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 5.846247673034668,
+ "step": 247
+ },
+ {
+ "epoch": 3.4467713787085517,
+ "grad_norm": 0.02388785220682621,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 5.837607383728027,
+ "step": 248
+ },
+ {
+ "epoch": 3.4607329842931938,
+ "grad_norm": 0.03094472736120224,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 5.832090377807617,
+ "step": 249
+ },
+ {
+ "epoch": 3.474694589877836,
+ "grad_norm": 0.03227658197283745,
+ "learning_rate": 0.000597347745466109,
+ "loss": 5.83722448348999,
+ "step": 250
+ },
+ {
+ "epoch": 3.488656195462478,
+ "grad_norm": 0.042813923209905624,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 5.851722717285156,
+ "step": 251
+ },
+ {
+ "epoch": 3.5026178010471205,
+ "grad_norm": 0.05825525149703026,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.795719146728516,
+ "step": 252
+ },
+ {
+ "epoch": 3.5165794066317626,
+ "grad_norm": 0.0867447778582573,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.813750267028809,
+ "step": 253
+ },
+ {
+ "epoch": 3.530541012216405,
+ "grad_norm": 0.09192710369825363,
+ "learning_rate": 0.000597203657966319,
+ "loss": 5.823488235473633,
+ "step": 254
+ },
+ {
+ "epoch": 3.5445026178010473,
+ "grad_norm": 0.06069248169660568,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 5.814342021942139,
+ "step": 255
+ },
+ {
+ "epoch": 3.5584642233856894,
+ "grad_norm": 0.06259952485561371,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 5.834272384643555,
+ "step": 256
+ },
+ {
+ "epoch": 3.5724258289703315,
+ "grad_norm": 0.0513639897108078,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.807951927185059,
+ "step": 257
+ },
+ {
+ "epoch": 3.5863874345549736,
+ "grad_norm": 0.03327798470854759,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.790238857269287,
+ "step": 258
+ },
+ {
+ "epoch": 3.600349040139616,
+ "grad_norm": 0.03963128477334976,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.7956695556640625,
+ "step": 259
+ },
+ {
+ "epoch": 3.614310645724258,
+ "grad_norm": 0.031153453513979912,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 5.7874531745910645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6282722513089007,
+ "grad_norm": 0.031419433653354645,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 5.779650688171387,
+ "step": 261
+ },
+ {
+ "epoch": 3.642233856893543,
+ "grad_norm": 0.0288684144616127,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.776577949523926,
+ "step": 262
+ },
+ {
+ "epoch": 3.656195462478185,
+ "grad_norm": 0.028315918520092964,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.778571128845215,
+ "step": 263
+ },
+ {
+ "epoch": 3.670157068062827,
+ "grad_norm": 0.024977795779705048,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 5.776470184326172,
+ "step": 264
+ },
+ {
+ "epoch": 3.6841186736474696,
+ "grad_norm": 0.026822632178664207,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.755108833312988,
+ "step": 265
+ },
+ {
+ "epoch": 3.6980802792321117,
+ "grad_norm": 0.02738630585372448,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.74041748046875,
+ "step": 266
+ },
+ {
+ "epoch": 3.712041884816754,
+ "grad_norm": 0.02361280657351017,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.764027118682861,
+ "step": 267
+ },
+ {
+ "epoch": 3.7260034904013963,
+ "grad_norm": 0.029926009476184845,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.780390739440918,
+ "step": 268
+ },
+ {
+ "epoch": 3.7399650959860384,
+ "grad_norm": 0.03519507870078087,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.758105278015137,
+ "step": 269
+ },
+ {
+ "epoch": 3.7539267015706805,
+ "grad_norm": 0.030114633962512016,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.757230758666992,
+ "step": 270
+ },
+ {
+ "epoch": 3.7678883071553226,
+ "grad_norm": 0.02467748522758484,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.7615461349487305,
+ "step": 271
+ },
+ {
+ "epoch": 3.781849912739965,
+ "grad_norm": 0.03425183147192001,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.749448776245117,
+ "step": 272
+ },
+ {
+ "epoch": 3.7958115183246073,
+ "grad_norm": 0.046307582408189774,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.7356977462768555,
+ "step": 273
+ },
+ {
+ "epoch": 3.80977312390925,
+ "grad_norm": 0.054102856665849686,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.741869926452637,
+ "step": 274
+ },
+ {
+ "epoch": 3.823734729493892,
+ "grad_norm": 0.07026689499616623,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.721103668212891,
+ "step": 275
+ },
+ {
+ "epoch": 3.837696335078534,
+ "grad_norm": 0.0733039602637291,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.731414794921875,
+ "step": 276
+ },
+ {
+ "epoch": 3.851657940663176,
+ "grad_norm": 0.05342920497059822,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.734922885894775,
+ "step": 277
+ },
+ {
+ "epoch": 3.8656195462478182,
+ "grad_norm": 0.0504680834710598,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.732746124267578,
+ "step": 278
+ },
+ {
+ "epoch": 3.8795811518324608,
+ "grad_norm": 0.038991592824459076,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.719473838806152,
+ "step": 279
+ },
+ {
+ "epoch": 3.893542757417103,
+ "grad_norm": 0.04434404894709587,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.709481239318848,
+ "step": 280
+ },
+ {
+ "epoch": 3.9075043630017454,
+ "grad_norm": 0.047285184264183044,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.693531036376953,
+ "step": 281
+ },
+ {
+ "epoch": 3.9214659685863875,
+ "grad_norm": 0.045081138610839844,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.728959083557129,
+ "step": 282
+ },
+ {
+ "epoch": 3.9354275741710296,
+ "grad_norm": 0.025133663788437843,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.719696998596191,
+ "step": 283
+ },
+ {
+ "epoch": 3.9493891797556717,
+ "grad_norm": 0.03353596478700638,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.700235366821289,
+ "step": 284
+ },
+ {
+ "epoch": 3.9633507853403143,
+ "grad_norm": 0.035162411630153656,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.684836387634277,
+ "step": 285
+ },
+ {
+ "epoch": 3.9773123909249564,
+ "grad_norm": 0.04025450348854065,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.692097187042236,
+ "step": 286
+ },
+ {
+ "epoch": 3.9912739965095985,
+ "grad_norm": 0.04995397478342056,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.6766862869262695,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.031619708985090256,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 3.5519747734069824,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.714687705039978,
+ "eval_runtime": 59.637,
+ "eval_samples_per_second": 40.948,
+ "eval_steps_per_second": 0.654,
+ "step": 288
+ },
+ {
+ "epoch": 4.013961605584642,
+ "grad_norm": 0.036328598856925964,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.666294097900391,
+ "step": 289
+ },
+ {
+ "epoch": 4.027923211169284,
+ "grad_norm": 0.07326336950063705,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.6879167556762695,
+ "step": 290
+ },
+ {
+ "epoch": 4.041884816753926,
+ "grad_norm": 0.07947526127099991,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.674883842468262,
+ "step": 291
+ },
+ {
+ "epoch": 4.055846422338569,
+ "grad_norm": 0.04599885269999504,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.688086986541748,
+ "step": 292
+ },
+ {
+ "epoch": 4.069808027923211,
+ "grad_norm": 0.0465150885283947,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.649483680725098,
+ "step": 293
+ },
+ {
+ "epoch": 4.0837696335078535,
+ "grad_norm": 0.038542453199625015,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.653055667877197,
+ "step": 294
+ },
+ {
+ "epoch": 4.097731239092496,
+ "grad_norm": 0.03659164905548096,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.65955924987793,
+ "step": 295
+ },
+ {
+ "epoch": 4.111692844677138,
+ "grad_norm": 0.03824038803577423,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.650853157043457,
+ "step": 296
+ },
+ {
+ "epoch": 4.12565445026178,
+ "grad_norm": 0.03591204807162285,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.643499374389648,
+ "step": 297
+ },
+ {
+ "epoch": 4.139616055846423,
+ "grad_norm": 0.03269682824611664,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.657600402832031,
+ "step": 298
+ },
+ {
+ "epoch": 4.153577661431065,
+ "grad_norm": 0.03343537449836731,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.6352081298828125,
+ "step": 299
+ },
+ {
+ "epoch": 4.167539267015707,
+ "grad_norm": 0.034283801913261414,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.626757621765137,
+ "step": 300
+ },
+ {
+ "epoch": 4.181500872600349,
+ "grad_norm": 0.034614723175764084,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.6316680908203125,
+ "step": 301
+ },
+ {
+ "epoch": 4.195462478184991,
+ "grad_norm": 0.0340128131210804,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.648160934448242,
+ "step": 302
+ },
+ {
+ "epoch": 4.209424083769633,
+ "grad_norm": 0.03068280965089798,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.619738578796387,
+ "step": 303
+ },
+ {
+ "epoch": 4.223385689354275,
+ "grad_norm": 0.030269058421254158,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.609238624572754,
+ "step": 304
+ },
+ {
+ "epoch": 4.237347294938918,
+ "grad_norm": 0.03612440824508667,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.602194786071777,
+ "step": 305
+ },
+ {
+ "epoch": 4.2513089005235605,
+ "grad_norm": 0.044292811304330826,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.61159610748291,
+ "step": 306
+ },
+ {
+ "epoch": 4.265270506108203,
+ "grad_norm": 0.04449291527271271,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.6274237632751465,
+ "step": 307
+ },
+ {
+ "epoch": 4.279232111692845,
+ "grad_norm": 0.05014493316411972,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.589034080505371,
+ "step": 308
+ },
+ {
+ "epoch": 4.293193717277487,
+ "grad_norm": 0.05286206305027008,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.600770950317383,
+ "step": 309
+ },
+ {
+ "epoch": 4.307155322862129,
+ "grad_norm": 0.048459846526384354,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.5796098709106445,
+ "step": 310
+ },
+ {
+ "epoch": 4.321116928446771,
+ "grad_norm": 0.05329663306474686,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.590407371520996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335078534031414,
+ "grad_norm": 0.052595119923353195,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.583693504333496,
+ "step": 312
+ },
+ {
+ "epoch": 4.349040139616056,
+ "grad_norm": 0.05164601653814316,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.603281021118164,
+ "step": 313
+ },
+ {
+ "epoch": 4.363001745200698,
+ "grad_norm": 0.04657137766480446,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.591344833374023,
+ "step": 314
+ },
+ {
+ "epoch": 4.37696335078534,
+ "grad_norm": 0.03804599866271019,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.580578804016113,
+ "step": 315
+ },
+ {
+ "epoch": 4.390924956369982,
+ "grad_norm": 0.03835320845246315,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.597836971282959,
+ "step": 316
+ },
+ {
+ "epoch": 4.4048865619546245,
+ "grad_norm": 0.034059133380651474,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.567502021789551,
+ "step": 317
+ },
+ {
+ "epoch": 4.418848167539267,
+ "grad_norm": 0.03350386768579483,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.552022933959961,
+ "step": 318
+ },
+ {
+ "epoch": 4.43280977312391,
+ "grad_norm": 0.032574303448200226,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.559177875518799,
+ "step": 319
+ },
+ {
+ "epoch": 4.446771378708552,
+ "grad_norm": 0.030893854796886444,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.549288272857666,
+ "step": 320
+ },
+ {
+ "epoch": 4.460732984293194,
+ "grad_norm": 0.03255036100745201,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.5400390625,
+ "step": 321
+ },
+ {
+ "epoch": 4.474694589877836,
+ "grad_norm": 0.03009258769452572,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.5328474044799805,
+ "step": 322
+ },
+ {
+ "epoch": 4.488656195462478,
+ "grad_norm": 0.03178320452570915,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.5594482421875,
+ "step": 323
+ },
+ {
+ "epoch": 4.50261780104712,
+ "grad_norm": 0.03360465541481972,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.5402727127075195,
+ "step": 324
+ },
+ {
+ "epoch": 4.516579406631763,
+ "grad_norm": 0.04275049269199371,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.534474849700928,
+ "step": 325
+ },
+ {
+ "epoch": 4.530541012216405,
+ "grad_norm": 0.04140641540288925,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.5324859619140625,
+ "step": 326
+ },
+ {
+ "epoch": 4.544502617801047,
+ "grad_norm": 0.04769204929471016,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.4919352531433105,
+ "step": 327
+ },
+ {
+ "epoch": 4.558464223385689,
+ "grad_norm": 0.04402285814285278,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.520858287811279,
+ "step": 328
+ },
+ {
+ "epoch": 4.5724258289703315,
+ "grad_norm": 0.038526229560375214,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.523778438568115,
+ "step": 329
+ },
+ {
+ "epoch": 4.5863874345549736,
+ "grad_norm": 0.05422377213835716,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.517973899841309,
+ "step": 330
+ },
+ {
+ "epoch": 4.600349040139616,
+ "grad_norm": 0.07236721366643906,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.548083782196045,
+ "step": 331
+ },
+ {
+ "epoch": 4.614310645724259,
+ "grad_norm": 0.06861061602830887,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.535577774047852,
+ "step": 332
+ },
+ {
+ "epoch": 4.628272251308901,
+ "grad_norm": 0.07148340344429016,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.54155158996582,
+ "step": 333
+ },
+ {
+ "epoch": 4.642233856893543,
+ "grad_norm": 0.07838302850723267,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.552852630615234,
+ "step": 334
+ },
+ {
+ "epoch": 4.656195462478185,
+ "grad_norm": 0.10946758091449738,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.555066108703613,
+ "step": 335
+ },
+ {
+ "epoch": 4.670157068062827,
+ "grad_norm": 0.10143530368804932,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.584932804107666,
+ "step": 336
+ },
+ {
+ "epoch": 4.684118673647469,
+ "grad_norm": 0.08926606178283691,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.570462226867676,
+ "step": 337
+ },
+ {
+ "epoch": 4.698080279232112,
+ "grad_norm": 0.07060826569795609,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.5716753005981445,
+ "step": 338
+ },
+ {
+ "epoch": 4.712041884816754,
+ "grad_norm": 0.0418456569314003,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.556496620178223,
+ "step": 339
+ },
+ {
+ "epoch": 4.726003490401396,
+ "grad_norm": 0.05952411890029907,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.567801475524902,
+ "step": 340
+ },
+ {
+ "epoch": 4.739965095986038,
+ "grad_norm": 0.042999885976314545,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.537403106689453,
+ "step": 341
+ },
+ {
+ "epoch": 4.7539267015706805,
+ "grad_norm": 0.04426449537277222,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.552387237548828,
+ "step": 342
+ },
+ {
+ "epoch": 4.767888307155323,
+ "grad_norm": 0.040835145860910416,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.53051233291626,
+ "step": 343
+ },
+ {
+ "epoch": 4.781849912739965,
+ "grad_norm": 0.03819291293621063,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.531174659729004,
+ "step": 344
+ },
+ {
+ "epoch": 4.795811518324607,
+ "grad_norm": 0.030088599771261215,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.499328136444092,
+ "step": 345
+ },
+ {
+ "epoch": 4.80977312390925,
+ "grad_norm": 0.029984362423419952,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.504540920257568,
+ "step": 346
+ },
+ {
+ "epoch": 4.823734729493892,
+ "grad_norm": 0.027402915060520172,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.50899076461792,
+ "step": 347
+ },
+ {
+ "epoch": 4.837696335078534,
+ "grad_norm": 0.025096213445067406,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.480479717254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.851657940663176,
+ "grad_norm": 0.026828132569789886,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.501729965209961,
+ "step": 349
+ },
+ {
+ "epoch": 4.865619546247818,
+ "grad_norm": 0.023715168237686157,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.466860771179199,
+ "step": 350
+ },
+ {
+ "epoch": 4.879581151832461,
+ "grad_norm": 0.024503640830516815,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.491827964782715,
+ "step": 351
+ },
+ {
+ "epoch": 4.893542757417103,
+ "grad_norm": 0.020891105756163597,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.478384971618652,
+ "step": 352
+ },
+ {
+ "epoch": 4.907504363001745,
+ "grad_norm": 0.021373765543103218,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.4612135887146,
+ "step": 353
+ },
+ {
+ "epoch": 4.9214659685863875,
+ "grad_norm": 0.021773340180516243,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.4547319412231445,
+ "step": 354
+ },
+ {
+ "epoch": 4.93542757417103,
+ "grad_norm": 0.019637586548924446,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.431741714477539,
+ "step": 355
+ },
+ {
+ "epoch": 4.949389179755672,
+ "grad_norm": 0.023108696565032005,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.448612213134766,
+ "step": 356
+ },
+ {
+ "epoch": 4.963350785340314,
+ "grad_norm": 0.021935103461146355,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.447112083435059,
+ "step": 357
+ },
+ {
+ "epoch": 4.977312390924956,
+ "grad_norm": 0.023260753601789474,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.436091423034668,
+ "step": 358
+ },
+ {
+ "epoch": 4.991273996509599,
+ "grad_norm": 0.021355057135224342,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.408421993255615,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.014992325566709042,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 3.4004223346710205,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6834986805915833,
+ "eval_runtime": 60.4148,
+ "eval_samples_per_second": 40.421,
+ "eval_steps_per_second": 0.646,
+ "step": 360
+ },
+ {
+ "epoch": 5.013961605584642,
+ "grad_norm": 0.02471218630671501,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.396238803863525,
+ "step": 361
+ },
+ {
+ "epoch": 5.027923211169284,
+ "grad_norm": 0.027672411873936653,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.4048919677734375,
+ "step": 362
+ },
+ {
+ "epoch": 5.041884816753926,
+ "grad_norm": 0.031560689210891724,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.399665832519531,
+ "step": 363
+ },
+ {
+ "epoch": 5.055846422338569,
+ "grad_norm": 0.03459061309695244,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.396360397338867,
+ "step": 364
+ },
+ {
+ "epoch": 5.069808027923211,
+ "grad_norm": 0.04040960222482681,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.405309677124023,
+ "step": 365
+ },
+ {
+ "epoch": 5.0837696335078535,
+ "grad_norm": 0.05444670096039772,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.408894062042236,
+ "step": 366
+ },
+ {
+ "epoch": 5.097731239092496,
+ "grad_norm": 0.05816745012998581,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.403566360473633,
+ "step": 367
+ },
+ {
+ "epoch": 5.111692844677138,
+ "grad_norm": 0.06330069154500961,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.395961761474609,
+ "step": 368
+ },
+ {
+ "epoch": 5.12565445026178,
+ "grad_norm": 0.07458917796611786,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.400968551635742,
+ "step": 369
+ },
+ {
+ "epoch": 5.139616055846423,
+ "grad_norm": 0.06231356039643288,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.412624359130859,
+ "step": 370
+ },
+ {
+ "epoch": 5.153577661431065,
+ "grad_norm": 0.05212346464395523,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.394794464111328,
+ "step": 371
+ },
+ {
+ "epoch": 5.167539267015707,
+ "grad_norm": 0.050003375858068466,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.380125045776367,
+ "step": 372
+ },
+ {
+ "epoch": 5.181500872600349,
+ "grad_norm": 0.048393815755844116,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.390806198120117,
+ "step": 373
+ },
+ {
+ "epoch": 5.195462478184991,
+ "grad_norm": 0.04200794920325279,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.387467384338379,
+ "step": 374
+ },
+ {
+ "epoch": 5.209424083769633,
+ "grad_norm": 0.04931448772549629,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.37552547454834,
+ "step": 375
+ },
+ {
+ "epoch": 5.223385689354275,
+ "grad_norm": 0.03851857781410217,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.373126983642578,
+ "step": 376
+ },
+ {
+ "epoch": 5.237347294938918,
+ "grad_norm": 0.03458717465400696,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.346826553344727,
+ "step": 377
+ },
+ {
+ "epoch": 5.2513089005235605,
+ "grad_norm": 0.031107677146792412,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.353341102600098,
+ "step": 378
+ },
+ {
+ "epoch": 5.265270506108203,
+ "grad_norm": 0.03077058307826519,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.358076095581055,
+ "step": 379
+ },
+ {
+ "epoch": 5.279232111692845,
+ "grad_norm": 0.028170401230454445,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.354019641876221,
+ "step": 380
+ },
+ {
+ "epoch": 5.293193717277487,
+ "grad_norm": 0.02871990203857422,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.368258476257324,
+ "step": 381
+ },
+ {
+ "epoch": 5.307155322862129,
+ "grad_norm": 0.03046584501862526,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.342505931854248,
+ "step": 382
+ },
+ {
+ "epoch": 5.321116928446771,
+ "grad_norm": 0.02781563811004162,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.339411735534668,
+ "step": 383
+ },
+ {
+ "epoch": 5.335078534031414,
+ "grad_norm": 0.028825288638472557,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.341948509216309,
+ "step": 384
+ },
+ {
+ "epoch": 5.349040139616056,
+ "grad_norm": 0.026684079319238663,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.340881824493408,
+ "step": 385
+ },
+ {
+ "epoch": 5.363001745200698,
+ "grad_norm": 0.02429167740046978,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.342733383178711,
+ "step": 386
+ },
+ {
+ "epoch": 5.37696335078534,
+ "grad_norm": 0.02855239436030388,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.313878536224365,
+ "step": 387
+ },
+ {
+ "epoch": 5.390924956369982,
+ "grad_norm": 0.029651548713445663,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.312665939331055,
+ "step": 388
+ },
+ {
+ "epoch": 5.4048865619546245,
+ "grad_norm": 0.027191132307052612,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3086838722229,
+ "step": 389
+ },
+ {
+ "epoch": 5.418848167539267,
+ "grad_norm": 0.026719525456428528,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.305359840393066,
+ "step": 390
+ },
+ {
+ "epoch": 5.43280977312391,
+ "grad_norm": 0.026763100177049637,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.3164381980896,
+ "step": 391
+ },
+ {
+ "epoch": 5.446771378708552,
+ "grad_norm": 0.03125109151005745,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.3083953857421875,
+ "step": 392
+ },
+ {
+ "epoch": 5.460732984293194,
+ "grad_norm": 0.0394827201962471,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.275996208190918,
+ "step": 393
+ },
+ {
+ "epoch": 5.474694589877836,
+ "grad_norm": 0.040579475462436676,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.292995929718018,
+ "step": 394
+ },
+ {
+ "epoch": 5.488656195462478,
+ "grad_norm": 0.037223003804683685,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.298526763916016,
+ "step": 395
+ },
+ {
+ "epoch": 5.50261780104712,
+ "grad_norm": 0.04207765311002731,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.287985801696777,
+ "step": 396
+ },
+ {
+ "epoch": 5.516579406631763,
+ "grad_norm": 0.048299964517354965,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.244608402252197,
+ "step": 397
+ },
+ {
+ "epoch": 5.530541012216405,
+ "grad_norm": 0.04650307446718216,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.285453796386719,
+ "step": 398
+ },
+ {
+ "epoch": 5.544502617801047,
+ "grad_norm": 0.045339711010456085,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.312187671661377,
+ "step": 399
+ },
+ {
+ "epoch": 5.558464223385689,
+ "grad_norm": 0.041067708283662796,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.292692184448242,
+ "step": 400
+ },
+ {
+ "epoch": 5.5724258289703315,
+ "grad_norm": 0.05455026403069496,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.253364562988281,
+ "step": 401
+ },
+ {
+ "epoch": 5.5863874345549736,
+ "grad_norm": 0.06660915911197662,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.262646675109863,
+ "step": 402
+ },
+ {
+ "epoch": 5.600349040139616,
+ "grad_norm": 0.06634793430566788,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.274322986602783,
+ "step": 403
+ },
+ {
+ "epoch": 5.614310645724259,
+ "grad_norm": 0.08996330946683884,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.287951469421387,
+ "step": 404
+ },
+ {
+ "epoch": 5.628272251308901,
+ "grad_norm": 0.08526470512151718,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.3007097244262695,
+ "step": 405
+ },
+ {
+ "epoch": 5.642233856893543,
+ "grad_norm": 0.08766836673021317,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.332538604736328,
+ "step": 406
+ },
+ {
+ "epoch": 5.656195462478185,
+ "grad_norm": 0.07055048644542694,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.296169281005859,
+ "step": 407
+ },
+ {
+ "epoch": 5.670157068062827,
+ "grad_norm": 0.07520575076341629,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.309683322906494,
+ "step": 408
+ },
+ {
+ "epoch": 5.684118673647469,
+ "grad_norm": 0.06115756183862686,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.311301231384277,
+ "step": 409
+ },
+ {
+ "epoch": 5.698080279232112,
+ "grad_norm": 0.05862593650817871,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.28066349029541,
+ "step": 410
+ },
+ {
+ "epoch": 5.712041884816754,
+ "grad_norm": 0.06169601157307625,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.286840438842773,
+ "step": 411
+ },
+ {
+ "epoch": 5.726003490401396,
+ "grad_norm": 0.05270719155669212,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.302594184875488,
+ "step": 412
+ },
+ {
+ "epoch": 5.739965095986038,
+ "grad_norm": 0.03912634402513504,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.262563705444336,
+ "step": 413
+ },
+ {
+ "epoch": 5.7539267015706805,
+ "grad_norm": 0.03215811401605606,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.269310474395752,
+ "step": 414
+ },
+ {
+ "epoch": 5.767888307155323,
+ "grad_norm": 0.03236516937613487,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.268766403198242,
+ "step": 415
+ },
+ {
+ "epoch": 5.781849912739965,
+ "grad_norm": 0.027429452165961266,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.283801555633545,
+ "step": 416
+ },
+ {
+ "epoch": 5.795811518324607,
+ "grad_norm": 0.02990538813173771,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.22291374206543,
+ "step": 417
+ },
+ {
+ "epoch": 5.80977312390925,
+ "grad_norm": 0.030410146340727806,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.2281599044799805,
+ "step": 418
+ },
+ {
+ "epoch": 5.823734729493892,
+ "grad_norm": 0.030068187043070793,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.250294208526611,
+ "step": 419
+ },
+ {
+ "epoch": 5.837696335078534,
+ "grad_norm": 0.02800513058900833,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.226508617401123,
+ "step": 420
+ },
+ {
+ "epoch": 5.851657940663176,
+ "grad_norm": 0.023823332041502,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.247838020324707,
+ "step": 421
+ },
+ {
+ "epoch": 5.865619546247818,
+ "grad_norm": 0.024974249303340912,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.2392473220825195,
+ "step": 422
+ },
+ {
+ "epoch": 5.879581151832461,
+ "grad_norm": 0.023320546373724937,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.209543704986572,
+ "step": 423
+ },
+ {
+ "epoch": 5.893542757417103,
+ "grad_norm": 0.024201491847634315,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.219725608825684,
+ "step": 424
+ },
+ {
+ "epoch": 5.907504363001745,
+ "grad_norm": 0.021649666130542755,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.203141212463379,
+ "step": 425
+ },
+ {
+ "epoch": 5.9214659685863875,
+ "grad_norm": 0.0248002577573061,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.179058074951172,
+ "step": 426
+ },
+ {
+ "epoch": 5.93542757417103,
+ "grad_norm": 0.028549587354063988,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.202126502990723,
+ "step": 427
+ },
+ {
+ "epoch": 5.949389179755672,
+ "grad_norm": 0.0331057533621788,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.208107948303223,
+ "step": 428
+ },
+ {
+ "epoch": 5.963350785340314,
+ "grad_norm": 0.031843364238739014,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.184683799743652,
+ "step": 429
+ },
+ {
+ "epoch": 5.977312390924956,
+ "grad_norm": 0.024927377700805664,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.180486679077148,
+ "step": 430
+ },
+ {
+ "epoch": 5.991273996509599,
+ "grad_norm": 0.026685267686843872,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.186746597290039,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.018545351922512054,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 3.217043876647949,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.653471827507019,
+ "eval_runtime": 61.1586,
+ "eval_samples_per_second": 39.929,
+ "eval_steps_per_second": 0.638,
+ "step": 432
+ },
+ {
+ "epoch": 6.013961605584642,
+ "grad_norm": 0.025291118770837784,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.156534194946289,
+ "step": 433
+ },
+ {
+ "epoch": 6.027923211169284,
+ "grad_norm": 0.03774752840399742,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.144041538238525,
+ "step": 434
+ },
+ {
+ "epoch": 6.041884816753926,
+ "grad_norm": 0.04244692996144295,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.158862113952637,
+ "step": 435
+ },
+ {
+ "epoch": 6.055846422338569,
+ "grad_norm": 0.039782389998435974,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.166165351867676,
+ "step": 436
+ },
+ {
+ "epoch": 6.069808027923211,
+ "grad_norm": 0.05489896610379219,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.155766487121582,
+ "step": 437
+ },
+ {
+ "epoch": 6.0837696335078535,
+ "grad_norm": 0.059844326227903366,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.153733253479004,
+ "step": 438
+ },
+ {
+ "epoch": 6.097731239092496,
+ "grad_norm": 0.05539938807487488,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.130112648010254,
+ "step": 439
+ },
+ {
+ "epoch": 6.111692844677138,
+ "grad_norm": 0.05283183231949806,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.136910915374756,
+ "step": 440
+ },
+ {
+ "epoch": 6.12565445026178,
+ "grad_norm": 0.054407261312007904,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.165216445922852,
+ "step": 441
+ },
+ {
+ "epoch": 6.139616055846423,
+ "grad_norm": 0.0509788915514946,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.140888214111328,
+ "step": 442
+ },
+ {
+ "epoch": 6.153577661431065,
+ "grad_norm": 0.05018748342990875,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.13115930557251,
+ "step": 443
+ },
+ {
+ "epoch": 6.167539267015707,
+ "grad_norm": 0.047012437134981155,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.146507263183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181500872600349,
+ "grad_norm": 0.04943335801362991,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.135608673095703,
+ "step": 445
+ },
+ {
+ "epoch": 6.195462478184991,
+ "grad_norm": 0.05304260924458504,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.136394023895264,
+ "step": 446
+ },
+ {
+ "epoch": 6.209424083769633,
+ "grad_norm": 0.04554047808051109,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.124661922454834,
+ "step": 447
+ },
+ {
+ "epoch": 6.223385689354275,
+ "grad_norm": 0.040672432631254196,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.129773139953613,
+ "step": 448
+ },
+ {
+ "epoch": 6.237347294938918,
+ "grad_norm": 0.03442879393696785,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.128028392791748,
+ "step": 449
+ },
+ {
+ "epoch": 6.2513089005235605,
+ "grad_norm": 0.03927934914827347,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.092485427856445,
+ "step": 450
+ },
+ {
+ "epoch": 6.265270506108203,
+ "grad_norm": 0.029380977153778076,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.077669143676758,
+ "step": 451
+ },
+ {
+ "epoch": 6.279232111692845,
+ "grad_norm": 0.03331821784377098,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.138857841491699,
+ "step": 452
+ },
+ {
+ "epoch": 6.293193717277487,
+ "grad_norm": 0.02939118817448616,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.082937240600586,
+ "step": 453
+ },
+ {
+ "epoch": 6.307155322862129,
+ "grad_norm": 0.02730412967503071,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.093183517456055,
+ "step": 454
+ },
+ {
+ "epoch": 6.321116928446771,
+ "grad_norm": 0.02964780665934086,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.1028151512146,
+ "step": 455
+ },
+ {
+ "epoch": 6.335078534031414,
+ "grad_norm": 0.026350129395723343,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.066362380981445,
+ "step": 456
+ },
+ {
+ "epoch": 6.349040139616056,
+ "grad_norm": 0.027010295540094376,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.070037841796875,
+ "step": 457
+ },
+ {
+ "epoch": 6.363001745200698,
+ "grad_norm": 0.02906128577888012,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.107708930969238,
+ "step": 458
+ },
+ {
+ "epoch": 6.37696335078534,
+ "grad_norm": 0.028619443997740746,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.073182582855225,
+ "step": 459
+ },
+ {
+ "epoch": 6.390924956369982,
+ "grad_norm": 0.029981834813952446,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.076602935791016,
+ "step": 460
+ },
+ {
+ "epoch": 6.4048865619546245,
+ "grad_norm": 0.03255423903465271,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.063301086425781,
+ "step": 461
+ },
+ {
+ "epoch": 6.418848167539267,
+ "grad_norm": 0.03863866627216339,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.041812896728516,
+ "step": 462
+ },
+ {
+ "epoch": 6.43280977312391,
+ "grad_norm": 0.03860306739807129,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.06298828125,
+ "step": 463
+ },
+ {
+ "epoch": 6.446771378708552,
+ "grad_norm": 0.03378435596823692,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.0434370040893555,
+ "step": 464
+ },
+ {
+ "epoch": 6.460732984293194,
+ "grad_norm": 0.03276759013533592,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.036187171936035,
+ "step": 465
+ },
+ {
+ "epoch": 6.474694589877836,
+ "grad_norm": 0.04130513221025467,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.039098739624023,
+ "step": 466
+ },
+ {
+ "epoch": 6.488656195462478,
+ "grad_norm": 0.05389940366148949,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.062798976898193,
+ "step": 467
+ },
+ {
+ "epoch": 6.50261780104712,
+ "grad_norm": 0.05881704390048981,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.067028522491455,
+ "step": 468
+ },
+ {
+ "epoch": 6.516579406631763,
+ "grad_norm": 0.061897799372673035,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.056947231292725,
+ "step": 469
+ },
+ {
+ "epoch": 6.530541012216405,
+ "grad_norm": 0.06549323350191116,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.089262008666992,
+ "step": 470
+ },
+ {
+ "epoch": 6.544502617801047,
+ "grad_norm": 0.048082590103149414,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.055979251861572,
+ "step": 471
+ },
+ {
+ "epoch": 6.558464223385689,
+ "grad_norm": 0.03768027946352959,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.0660600662231445,
+ "step": 472
+ },
+ {
+ "epoch": 6.5724258289703315,
+ "grad_norm": 0.043922603130340576,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 5.036745071411133,
+ "step": 473
+ },
+ {
+ "epoch": 6.5863874345549736,
+ "grad_norm": 0.044848836958408356,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.073650360107422,
+ "step": 474
+ },
+ {
+ "epoch": 6.600349040139616,
+ "grad_norm": 0.05841026455163956,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.053045272827148,
+ "step": 475
+ },
+ {
+ "epoch": 6.614310645724259,
+ "grad_norm": 0.0650017261505127,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 5.052206039428711,
+ "step": 476
+ },
+ {
+ "epoch": 6.628272251308901,
+ "grad_norm": 0.048979002982378006,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 5.054277420043945,
+ "step": 477
+ },
+ {
+ "epoch": 6.642233856893543,
+ "grad_norm": 0.05085799843072891,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 5.063844680786133,
+ "step": 478
+ },
+ {
+ "epoch": 6.656195462478185,
+ "grad_norm": 0.04336017370223999,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 5.045474052429199,
+ "step": 479
+ },
+ {
+ "epoch": 6.670157068062827,
+ "grad_norm": 0.04113380238413811,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 5.031338214874268,
+ "step": 480
+ },
+ {
+ "epoch": 6.684118673647469,
+ "grad_norm": 0.03199538215994835,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 5.030025959014893,
+ "step": 481
+ },
+ {
+ "epoch": 6.698080279232112,
+ "grad_norm": 0.03043113648891449,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 5.033163070678711,
+ "step": 482
+ },
+ {
+ "epoch": 6.712041884816754,
+ "grad_norm": 0.031694378703832626,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 5.004739761352539,
+ "step": 483
+ },
+ {
+ "epoch": 6.726003490401396,
+ "grad_norm": 0.030243348330259323,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 5.012566089630127,
+ "step": 484
+ },
+ {
+ "epoch": 6.739965095986038,
+ "grad_norm": 0.02937416359782219,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 5.039803504943848,
+ "step": 485
+ },
+ {
+ "epoch": 6.7539267015706805,
+ "grad_norm": 0.029616443440318108,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 5.026117324829102,
+ "step": 486
+ },
+ {
+ "epoch": 6.767888307155323,
+ "grad_norm": 0.029567131772637367,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 5.006569862365723,
+ "step": 487
+ },
+ {
+ "epoch": 6.781849912739965,
+ "grad_norm": 0.029443370178341866,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.9941487312316895,
+ "step": 488
+ },
+ {
+ "epoch": 6.795811518324607,
+ "grad_norm": 0.03434352949261665,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.976431846618652,
+ "step": 489
+ },
+ {
+ "epoch": 6.80977312390925,
+ "grad_norm": 0.03845234215259552,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.993168354034424,
+ "step": 490
+ },
+ {
+ "epoch": 6.823734729493892,
+ "grad_norm": 0.05058208107948303,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.030162334442139,
+ "step": 491
+ },
+ {
+ "epoch": 6.837696335078534,
+ "grad_norm": 0.05961710214614868,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 5.007053375244141,
+ "step": 492
+ },
+ {
+ "epoch": 6.851657940663176,
+ "grad_norm": 0.056259628385305405,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 5.020356178283691,
+ "step": 493
+ },
+ {
+ "epoch": 6.865619546247818,
+ "grad_norm": 0.048628658056259155,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 5.037652969360352,
+ "step": 494
+ },
+ {
+ "epoch": 6.879581151832461,
+ "grad_norm": 0.040836453437805176,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.990390777587891,
+ "step": 495
+ },
+ {
+ "epoch": 6.893542757417103,
+ "grad_norm": 0.04112066701054573,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 5.017822742462158,
+ "step": 496
+ },
+ {
+ "epoch": 6.907504363001745,
+ "grad_norm": 0.03956829756498337,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.993971824645996,
+ "step": 497
+ },
+ {
+ "epoch": 6.9214659685863875,
+ "grad_norm": 0.03887917473912239,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.982889175415039,
+ "step": 498
+ },
+ {
+ "epoch": 6.93542757417103,
+ "grad_norm": 0.031012587249279022,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.961717128753662,
+ "step": 499
+ },
+ {
+ "epoch": 6.949389179755672,
+ "grad_norm": 0.03443489596247673,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.9904046058654785,
+ "step": 500
+ },
+ {
+ "epoch": 6.963350785340314,
+ "grad_norm": 0.035175010561943054,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.983619213104248,
+ "step": 501
+ },
+ {
+ "epoch": 6.977312390924956,
+ "grad_norm": 0.031724050641059875,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 5.000522613525391,
+ "step": 502
+ },
+ {
+ "epoch": 6.991273996509599,
+ "grad_norm": 0.032037995755672455,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.947064399719238,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.01883271522819996,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 3.112166166305542,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6296830177307129,
+ "eval_runtime": 59.5495,
+ "eval_samples_per_second": 41.008,
+ "eval_steps_per_second": 0.655,
+ "step": 504
+ },
+ {
+ "epoch": 7.013961605584642,
+ "grad_norm": 0.031091563403606415,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.919820785522461,
+ "step": 505
+ },
+ {
+ "epoch": 7.027923211169284,
+ "grad_norm": 0.04074421525001526,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.945765972137451,
+ "step": 506
+ },
+ {
+ "epoch": 7.041884816753926,
+ "grad_norm": 0.04225052148103714,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.935208797454834,
+ "step": 507
+ },
+ {
+ "epoch": 7.055846422338569,
+ "grad_norm": 0.0399051196873188,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.951314926147461,
+ "step": 508
+ },
+ {
+ "epoch": 7.069808027923211,
+ "grad_norm": 0.04785890504717827,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.92843770980835,
+ "step": 509
+ },
+ {
+ "epoch": 7.0837696335078535,
+ "grad_norm": 0.04922301694750786,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.901020050048828,
+ "step": 510
+ },
+ {
+ "epoch": 7.097731239092496,
+ "grad_norm": 0.05307295173406601,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.961151599884033,
+ "step": 511
+ },
+ {
+ "epoch": 7.111692844677138,
+ "grad_norm": 0.07003362476825714,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.9427876472473145,
+ "step": 512
+ },
+ {
+ "epoch": 7.12565445026178,
+ "grad_norm": 0.06498776376247406,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.947838306427002,
+ "step": 513
+ },
+ {
+ "epoch": 7.139616055846423,
+ "grad_norm": 0.058723628520965576,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.938859939575195,
+ "step": 514
+ },
+ {
+ "epoch": 7.153577661431065,
+ "grad_norm": 0.06590746343135834,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.962221145629883,
+ "step": 515
+ },
+ {
+ "epoch": 7.167539267015707,
+ "grad_norm": 0.052714601159095764,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.931011199951172,
+ "step": 516
+ },
+ {
+ "epoch": 7.181500872600349,
+ "grad_norm": 0.04296804964542389,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.93586540222168,
+ "step": 517
+ },
+ {
+ "epoch": 7.195462478184991,
+ "grad_norm": 0.03699173778295517,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.910763263702393,
+ "step": 518
+ },
+ {
+ "epoch": 7.209424083769633,
+ "grad_norm": 0.0317913293838501,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.9256439208984375,
+ "step": 519
+ },
+ {
+ "epoch": 7.223385689354275,
+ "grad_norm": 0.030313925817608833,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.938329696655273,
+ "step": 520
+ },
+ {
+ "epoch": 7.237347294938918,
+ "grad_norm": 0.02879229187965393,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.924966812133789,
+ "step": 521
+ },
+ {
+ "epoch": 7.2513089005235605,
+ "grad_norm": 0.02689451351761818,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.9079742431640625,
+ "step": 522
+ },
+ {
+ "epoch": 7.265270506108203,
+ "grad_norm": 0.028403930366039276,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.913238525390625,
+ "step": 523
+ },
+ {
+ "epoch": 7.279232111692845,
+ "grad_norm": 0.025912854820489883,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.896944046020508,
+ "step": 524
+ },
+ {
+ "epoch": 7.293193717277487,
+ "grad_norm": 0.024899760261178017,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.884920120239258,
+ "step": 525
+ },
+ {
+ "epoch": 7.307155322862129,
+ "grad_norm": 0.026912573724985123,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.9181060791015625,
+ "step": 526
+ },
+ {
+ "epoch": 7.321116928446771,
+ "grad_norm": 0.029539721086621284,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.914867401123047,
+ "step": 527
+ },
+ {
+ "epoch": 7.335078534031414,
+ "grad_norm": 0.028527872636914253,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.907217979431152,
+ "step": 528
+ },
+ {
+ "epoch": 7.349040139616056,
+ "grad_norm": 0.025879016146063805,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.905783176422119,
+ "step": 529
+ },
+ {
+ "epoch": 7.363001745200698,
+ "grad_norm": 0.026395199820399284,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.880505561828613,
+ "step": 530
+ },
+ {
+ "epoch": 7.37696335078534,
+ "grad_norm": 0.025072351098060608,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.870449066162109,
+ "step": 531
+ },
+ {
+ "epoch": 7.390924956369982,
+ "grad_norm": 0.02751019410789013,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.888952255249023,
+ "step": 532
+ },
+ {
+ "epoch": 7.4048865619546245,
+ "grad_norm": 0.025810658931732178,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.894552230834961,
+ "step": 533
+ },
+ {
+ "epoch": 7.418848167539267,
+ "grad_norm": 0.02399870567023754,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.8768229484558105,
+ "step": 534
+ },
+ {
+ "epoch": 7.43280977312391,
+ "grad_norm": 0.024639368057250977,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.9030914306640625,
+ "step": 535
+ },
+ {
+ "epoch": 7.446771378708552,
+ "grad_norm": 0.02543291077017784,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.86467170715332,
+ "step": 536
+ },
+ {
+ "epoch": 7.460732984293194,
+ "grad_norm": 0.02702544629573822,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.888283729553223,
+ "step": 537
+ },
+ {
+ "epoch": 7.474694589877836,
+ "grad_norm": 0.028904743492603302,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.866081237792969,
+ "step": 538
+ },
+ {
+ "epoch": 7.488656195462478,
+ "grad_norm": 0.03055848367512226,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.890739440917969,
+ "step": 539
+ },
+ {
+ "epoch": 7.50261780104712,
+ "grad_norm": 0.03410084918141365,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.862918853759766,
+ "step": 540
+ },
+ {
+ "epoch": 7.516579406631763,
+ "grad_norm": 0.03738495334982872,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.864728927612305,
+ "step": 541
+ },
+ {
+ "epoch": 7.530541012216405,
+ "grad_norm": 0.037646882236003876,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.865185737609863,
+ "step": 542
+ },
+ {
+ "epoch": 7.544502617801047,
+ "grad_norm": 0.03418618440628052,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.866114139556885,
+ "step": 543
+ },
+ {
+ "epoch": 7.558464223385689,
+ "grad_norm": 0.036924004554748535,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.883186340332031,
+ "step": 544
+ },
+ {
+ "epoch": 7.5724258289703315,
+ "grad_norm": 0.03590598329901695,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.863725662231445,
+ "step": 545
+ },
+ {
+ "epoch": 7.5863874345549736,
+ "grad_norm": 0.037363454699516296,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.841030120849609,
+ "step": 546
+ },
+ {
+ "epoch": 7.600349040139616,
+ "grad_norm": 0.046537864953279495,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.873837471008301,
+ "step": 547
+ },
+ {
+ "epoch": 7.614310645724259,
+ "grad_norm": 0.057226769626140594,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.865679740905762,
+ "step": 548
+ },
+ {
+ "epoch": 7.628272251308901,
+ "grad_norm": 0.06175961345434189,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.888645648956299,
+ "step": 549
+ },
+ {
+ "epoch": 7.642233856893543,
+ "grad_norm": 0.05063694715499878,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.877828598022461,
+ "step": 550
+ },
+ {
+ "epoch": 7.656195462478185,
+ "grad_norm": 0.03950628638267517,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.88637638092041,
+ "step": 551
+ },
+ {
+ "epoch": 7.670157068062827,
+ "grad_norm": 0.04183337092399597,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.8597917556762695,
+ "step": 552
+ },
+ {
+ "epoch": 7.684118673647469,
+ "grad_norm": 0.04642999917268753,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.880380153656006,
+ "step": 553
+ },
+ {
+ "epoch": 7.698080279232112,
+ "grad_norm": 0.05333106964826584,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.853932857513428,
+ "step": 554
+ },
+ {
+ "epoch": 7.712041884816754,
+ "grad_norm": 0.04764975979924202,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.870235443115234,
+ "step": 555
+ },
+ {
+ "epoch": 7.726003490401396,
+ "grad_norm": 0.043072231113910675,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.822697639465332,
+ "step": 556
+ },
+ {
+ "epoch": 7.739965095986038,
+ "grad_norm": 0.03913857042789459,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.849332809448242,
+ "step": 557
+ },
+ {
+ "epoch": 7.7539267015706805,
+ "grad_norm": 0.034836407750844955,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.842280387878418,
+ "step": 558
+ },
+ {
+ "epoch": 7.767888307155323,
+ "grad_norm": 0.033344294875860214,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.837811470031738,
+ "step": 559
+ },
+ {
+ "epoch": 7.781849912739965,
+ "grad_norm": 0.03342701122164726,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.876040458679199,
+ "step": 560
+ },
+ {
+ "epoch": 7.795811518324607,
+ "grad_norm": 0.034638747572898865,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.830340385437012,
+ "step": 561
+ },
+ {
+ "epoch": 7.80977312390925,
+ "grad_norm": 0.03358135372400284,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.828607559204102,
+ "step": 562
+ },
+ {
+ "epoch": 7.823734729493892,
+ "grad_norm": 0.036001041531562805,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.827905654907227,
+ "step": 563
+ },
+ {
+ "epoch": 7.837696335078534,
+ "grad_norm": 0.038079630583524704,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.866955280303955,
+ "step": 564
+ },
+ {
+ "epoch": 7.851657940663176,
+ "grad_norm": 0.03728470206260681,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.828039169311523,
+ "step": 565
+ },
+ {
+ "epoch": 7.865619546247818,
+ "grad_norm": 0.03920375928282738,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.841631889343262,
+ "step": 566
+ },
+ {
+ "epoch": 7.879581151832461,
+ "grad_norm": 0.032353926450014114,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.861110210418701,
+ "step": 567
+ },
+ {
+ "epoch": 7.893542757417103,
+ "grad_norm": 0.03241422027349472,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.818441390991211,
+ "step": 568
+ },
+ {
+ "epoch": 7.907504363001745,
+ "grad_norm": 0.03879080340266228,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.823749542236328,
+ "step": 569
+ },
+ {
+ "epoch": 7.9214659685863875,
+ "grad_norm": 0.03486498072743416,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.830697059631348,
+ "step": 570
+ },
+ {
+ "epoch": 7.93542757417103,
+ "grad_norm": 0.03279713913798332,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.815736770629883,
+ "step": 571
+ },
+ {
+ "epoch": 7.949389179755672,
+ "grad_norm": 0.03313925117254257,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.8364033699035645,
+ "step": 572
+ },
+ {
+ "epoch": 7.963350785340314,
+ "grad_norm": 0.03405062481760979,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.816010475158691,
+ "step": 573
+ },
+ {
+ "epoch": 7.977312390924956,
+ "grad_norm": 0.03259236365556717,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.813413619995117,
+ "step": 574
+ },
+ {
+ "epoch": 7.991273996509599,
+ "grad_norm": 0.030408985912799835,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.833840370178223,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.022680889815092087,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 2.99565052986145,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6126086115837097,
+ "eval_runtime": 60.2662,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 576
+ },
+ {
+ "epoch": 8.013961605584642,
+ "grad_norm": 0.029732126742601395,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.7816314697265625,
+ "step": 577
+ },
+ {
+ "epoch": 8.027923211169284,
+ "grad_norm": 0.04357453063130379,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.7575225830078125,
+ "step": 578
+ },
+ {
+ "epoch": 8.041884816753926,
+ "grad_norm": 0.051903221756219864,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.802519798278809,
+ "step": 579
+ },
+ {
+ "epoch": 8.055846422338568,
+ "grad_norm": 0.0539076030254364,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.799004554748535,
+ "step": 580
+ },
+ {
+ "epoch": 8.06980802792321,
+ "grad_norm": 0.043642807751894,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.7896528244018555,
+ "step": 581
+ },
+ {
+ "epoch": 8.083769633507853,
+ "grad_norm": 0.0469554178416729,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.765430927276611,
+ "step": 582
+ },
+ {
+ "epoch": 8.097731239092496,
+ "grad_norm": 0.042060062289237976,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.784132957458496,
+ "step": 583
+ },
+ {
+ "epoch": 8.111692844677139,
+ "grad_norm": 0.035101912915706635,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.767234802246094,
+ "step": 584
+ },
+ {
+ "epoch": 8.12565445026178,
+ "grad_norm": 0.042576879262924194,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.762052536010742,
+ "step": 585
+ },
+ {
+ "epoch": 8.139616055846423,
+ "grad_norm": 0.04231419041752815,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.752150535583496,
+ "step": 586
+ },
+ {
+ "epoch": 8.153577661431065,
+ "grad_norm": 0.046912599354982376,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.771881103515625,
+ "step": 587
+ },
+ {
+ "epoch": 8.167539267015707,
+ "grad_norm": 0.051271144300699234,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.773284912109375,
+ "step": 588
+ },
+ {
+ "epoch": 8.181500872600349,
+ "grad_norm": 0.04655958339571953,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.769917964935303,
+ "step": 589
+ },
+ {
+ "epoch": 8.195462478184991,
+ "grad_norm": 0.04463687539100647,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.77938175201416,
+ "step": 590
+ },
+ {
+ "epoch": 8.209424083769633,
+ "grad_norm": 0.047959886491298676,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.785962104797363,
+ "step": 591
+ },
+ {
+ "epoch": 8.223385689354275,
+ "grad_norm": 0.04414986073970795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.786281585693359,
+ "step": 592
+ },
+ {
+ "epoch": 8.237347294938917,
+ "grad_norm": 0.044150348752737045,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.777070045471191,
+ "step": 593
+ },
+ {
+ "epoch": 8.25130890052356,
+ "grad_norm": 0.037002332508563995,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.77547550201416,
+ "step": 594
+ },
+ {
+ "epoch": 8.265270506108202,
+ "grad_norm": 0.03417367860674858,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.782101154327393,
+ "step": 595
+ },
+ {
+ "epoch": 8.279232111692846,
+ "grad_norm": 0.03069210797548294,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.745595455169678,
+ "step": 596
+ },
+ {
+ "epoch": 8.293193717277488,
+ "grad_norm": 0.03228718787431717,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.7663655281066895,
+ "step": 597
+ },
+ {
+ "epoch": 8.30715532286213,
+ "grad_norm": 0.028603695333003998,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.767274856567383,
+ "step": 598
+ },
+ {
+ "epoch": 8.321116928446772,
+ "grad_norm": 0.03098377026617527,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.7853684425354,
+ "step": 599
+ },
+ {
+ "epoch": 8.335078534031414,
+ "grad_norm": 0.029242439195513725,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.738029956817627,
+ "step": 600
+ },
+ {
+ "epoch": 8.349040139616056,
+ "grad_norm": 0.02751081995666027,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.76801872253418,
+ "step": 601
+ },
+ {
+ "epoch": 8.363001745200698,
+ "grad_norm": 0.02838125079870224,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.743884086608887,
+ "step": 602
+ },
+ {
+ "epoch": 8.37696335078534,
+ "grad_norm": 0.029203062877058983,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.746668815612793,
+ "step": 603
+ },
+ {
+ "epoch": 8.390924956369982,
+ "grad_norm": 0.02679944597184658,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.760764122009277,
+ "step": 604
+ },
+ {
+ "epoch": 8.404886561954624,
+ "grad_norm": 0.028955848887562752,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.7353668212890625,
+ "step": 605
+ },
+ {
+ "epoch": 8.418848167539267,
+ "grad_norm": 0.02880851738154888,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.766837120056152,
+ "step": 606
+ },
+ {
+ "epoch": 8.432809773123909,
+ "grad_norm": 0.02495497092604637,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.78959321975708,
+ "step": 607
+ },
+ {
+ "epoch": 8.44677137870855,
+ "grad_norm": 0.026064753532409668,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.736433029174805,
+ "step": 608
+ },
+ {
+ "epoch": 8.460732984293193,
+ "grad_norm": 0.027566388249397278,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.782382965087891,
+ "step": 609
+ },
+ {
+ "epoch": 8.474694589877837,
+ "grad_norm": 0.02967359498143196,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.735286712646484,
+ "step": 610
+ },
+ {
+ "epoch": 8.488656195462479,
+ "grad_norm": 0.03079676441848278,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.725196838378906,
+ "step": 611
+ },
+ {
+ "epoch": 8.502617801047121,
+ "grad_norm": 0.032531846314668655,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.719333648681641,
+ "step": 612
+ },
+ {
+ "epoch": 8.516579406631763,
+ "grad_norm": 0.03120352514088154,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.723845958709717,
+ "step": 613
+ },
+ {
+ "epoch": 8.530541012216405,
+ "grad_norm": 0.03328424692153931,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.7669782638549805,
+ "step": 614
+ },
+ {
+ "epoch": 8.544502617801047,
+ "grad_norm": 0.035601649433374405,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.714925765991211,
+ "step": 615
+ },
+ {
+ "epoch": 8.55846422338569,
+ "grad_norm": 0.03675796464085579,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.715317726135254,
+ "step": 616
+ },
+ {
+ "epoch": 8.572425828970331,
+ "grad_norm": 0.03701988235116005,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.724606990814209,
+ "step": 617
+ },
+ {
+ "epoch": 8.586387434554974,
+ "grad_norm": 0.036719705909490585,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.746044635772705,
+ "step": 618
+ },
+ {
+ "epoch": 8.600349040139616,
+ "grad_norm": 0.034756604582071304,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.74885368347168,
+ "step": 619
+ },
+ {
+ "epoch": 8.614310645724258,
+ "grad_norm": 0.03237922117114067,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.721858501434326,
+ "step": 620
+ },
+ {
+ "epoch": 8.6282722513089,
+ "grad_norm": 0.028420142829418182,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.733814239501953,
+ "step": 621
+ },
+ {
+ "epoch": 8.642233856893542,
+ "grad_norm": 0.026203038170933723,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.726088523864746,
+ "step": 622
+ },
+ {
+ "epoch": 8.656195462478184,
+ "grad_norm": 0.02627231739461422,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.727700233459473,
+ "step": 623
+ },
+ {
+ "epoch": 8.670157068062828,
+ "grad_norm": 0.02920261025428772,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.731454849243164,
+ "step": 624
+ },
+ {
+ "epoch": 8.68411867364747,
+ "grad_norm": 0.02792907878756523,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.734154224395752,
+ "step": 625
+ },
+ {
+ "epoch": 8.698080279232112,
+ "grad_norm": 0.026770325377583504,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.715337753295898,
+ "step": 626
+ },
+ {
+ "epoch": 8.712041884816754,
+ "grad_norm": 0.027346540242433548,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.716512680053711,
+ "step": 627
+ },
+ {
+ "epoch": 8.726003490401396,
+ "grad_norm": 0.031965773552656174,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.7072038650512695,
+ "step": 628
+ },
+ {
+ "epoch": 8.739965095986038,
+ "grad_norm": 0.03708025813102722,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.749785900115967,
+ "step": 629
+ },
+ {
+ "epoch": 8.75392670157068,
+ "grad_norm": 0.04846490919589996,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.735586166381836,
+ "step": 630
+ },
+ {
+ "epoch": 8.767888307155323,
+ "grad_norm": 0.053352758288383484,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.720524787902832,
+ "step": 631
+ },
+ {
+ "epoch": 8.781849912739965,
+ "grad_norm": 0.040559202432632446,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.719901084899902,
+ "step": 632
+ },
+ {
+ "epoch": 8.795811518324607,
+ "grad_norm": 0.03333032503724098,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.708139419555664,
+ "step": 633
+ },
+ {
+ "epoch": 8.809773123909249,
+ "grad_norm": 0.030532941222190857,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.728184223175049,
+ "step": 634
+ },
+ {
+ "epoch": 8.823734729493891,
+ "grad_norm": 0.03337649628520012,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.726020336151123,
+ "step": 635
+ },
+ {
+ "epoch": 8.837696335078533,
+ "grad_norm": 0.03352045267820358,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.7196855545043945,
+ "step": 636
+ },
+ {
+ "epoch": 8.851657940663177,
+ "grad_norm": 0.03662872314453125,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.732823371887207,
+ "step": 637
+ },
+ {
+ "epoch": 8.86561954624782,
+ "grad_norm": 0.037085745483636856,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.728443145751953,
+ "step": 638
+ },
+ {
+ "epoch": 8.879581151832461,
+ "grad_norm": 0.036635689437389374,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.713988780975342,
+ "step": 639
+ },
+ {
+ "epoch": 8.893542757417103,
+ "grad_norm": 0.03719659522175789,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.715950965881348,
+ "step": 640
+ },
+ {
+ "epoch": 8.907504363001745,
+ "grad_norm": 0.03719159588217735,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.7124924659729,
+ "step": 641
+ },
+ {
+ "epoch": 8.921465968586388,
+ "grad_norm": 0.04275721684098244,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.708404064178467,
+ "step": 642
+ },
+ {
+ "epoch": 8.93542757417103,
+ "grad_norm": 0.04204598814249039,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.715372085571289,
+ "step": 643
+ },
+ {
+ "epoch": 8.949389179755672,
+ "grad_norm": 0.03938259929418564,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.729745864868164,
+ "step": 644
+ },
+ {
+ "epoch": 8.963350785340314,
+ "grad_norm": 0.036226850003004074,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.733749866485596,
+ "step": 645
+ },
+ {
+ "epoch": 8.977312390924956,
+ "grad_norm": 0.03542874753475189,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.706825256347656,
+ "step": 646
+ },
+ {
+ "epoch": 8.991273996509598,
+ "grad_norm": 0.029055381193757057,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.718567848205566,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.01990380510687828,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 2.943183183670044,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.601699709892273,
+ "eval_runtime": 60.0221,
+ "eval_samples_per_second": 40.685,
+ "eval_steps_per_second": 0.65,
+ "step": 648
+ },
+ {
+ "epoch": 9.013961605584642,
+ "grad_norm": 0.030904127284884453,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.660327911376953,
+ "step": 649
+ },
+ {
+ "epoch": 9.027923211169284,
+ "grad_norm": 0.0441313236951828,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.701565742492676,
+ "step": 650
+ },
+ {
+ "epoch": 9.041884816753926,
+ "grad_norm": 0.047547727823257446,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.666440486907959,
+ "step": 651
+ },
+ {
+ "epoch": 9.055846422338568,
+ "grad_norm": 0.04549463838338852,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.660548210144043,
+ "step": 652
+ },
+ {
+ "epoch": 9.06980802792321,
+ "grad_norm": 0.050346389412879944,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.640229225158691,
+ "step": 653
+ },
+ {
+ "epoch": 9.083769633507853,
+ "grad_norm": 0.058505453169345856,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.673881530761719,
+ "step": 654
+ },
+ {
+ "epoch": 9.097731239092496,
+ "grad_norm": 0.051989227533340454,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.693943977355957,
+ "step": 655
+ },
+ {
+ "epoch": 9.111692844677139,
+ "grad_norm": 0.052564945071935654,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.644315242767334,
+ "step": 656
+ },
+ {
+ "epoch": 9.12565445026178,
+ "grad_norm": 0.05647841840982437,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.681768417358398,
+ "step": 657
+ },
+ {
+ "epoch": 9.139616055846423,
+ "grad_norm": 0.058640867471694946,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.676457405090332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153577661431065,
+ "grad_norm": 0.05386025458574295,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.659588813781738,
+ "step": 659
+ },
+ {
+ "epoch": 9.167539267015707,
+ "grad_norm": 0.048699021339416504,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.680031776428223,
+ "step": 660
+ },
+ {
+ "epoch": 9.181500872600349,
+ "grad_norm": 0.04374176263809204,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.672671318054199,
+ "step": 661
+ },
+ {
+ "epoch": 9.195462478184991,
+ "grad_norm": 0.042337436228990555,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.673431873321533,
+ "step": 662
+ },
+ {
+ "epoch": 9.209424083769633,
+ "grad_norm": 0.04564661160111427,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.632595062255859,
+ "step": 663
+ },
+ {
+ "epoch": 9.223385689354275,
+ "grad_norm": 0.04404835030436516,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.678891181945801,
+ "step": 664
+ },
+ {
+ "epoch": 9.237347294938917,
+ "grad_norm": 0.040642790496349335,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.633418560028076,
+ "step": 665
+ },
+ {
+ "epoch": 9.25130890052356,
+ "grad_norm": 0.036492060869932175,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.670109748840332,
+ "step": 666
+ },
+ {
+ "epoch": 9.265270506108202,
+ "grad_norm": 0.03735879808664322,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.608570098876953,
+ "step": 667
+ },
+ {
+ "epoch": 9.279232111692846,
+ "grad_norm": 0.034563641995191574,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.632505416870117,
+ "step": 668
+ },
+ {
+ "epoch": 9.293193717277488,
+ "grad_norm": 0.03522484749555588,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.6475629806518555,
+ "step": 669
+ },
+ {
+ "epoch": 9.30715532286213,
+ "grad_norm": 0.03216472268104553,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.627873420715332,
+ "step": 670
+ },
+ {
+ "epoch": 9.321116928446772,
+ "grad_norm": 0.029155759140849113,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.652706146240234,
+ "step": 671
+ },
+ {
+ "epoch": 9.335078534031414,
+ "grad_norm": 0.031060541048645973,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.644545078277588,
+ "step": 672
+ },
+ {
+ "epoch": 9.349040139616056,
+ "grad_norm": 0.034979186952114105,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.642645359039307,
+ "step": 673
+ },
+ {
+ "epoch": 9.363001745200698,
+ "grad_norm": 0.035258810967206955,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.647383689880371,
+ "step": 674
+ },
+ {
+ "epoch": 9.37696335078534,
+ "grad_norm": 0.031552575528621674,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.667779445648193,
+ "step": 675
+ },
+ {
+ "epoch": 9.390924956369982,
+ "grad_norm": 0.03193951025605202,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.63625431060791,
+ "step": 676
+ },
+ {
+ "epoch": 9.404886561954624,
+ "grad_norm": 0.029811711981892586,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.658000946044922,
+ "step": 677
+ },
+ {
+ "epoch": 9.418848167539267,
+ "grad_norm": 0.028464138507843018,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.6545515060424805,
+ "step": 678
+ },
+ {
+ "epoch": 9.432809773123909,
+ "grad_norm": 0.0315735749900341,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.641153335571289,
+ "step": 679
+ },
+ {
+ "epoch": 9.44677137870855,
+ "grad_norm": 0.033412106335163116,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.626931667327881,
+ "step": 680
+ },
+ {
+ "epoch": 9.460732984293193,
+ "grad_norm": 0.03757626190781593,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.671254634857178,
+ "step": 681
+ },
+ {
+ "epoch": 9.474694589877837,
+ "grad_norm": 0.03910793736577034,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.651954174041748,
+ "step": 682
+ },
+ {
+ "epoch": 9.488656195462479,
+ "grad_norm": 0.04015549272298813,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.613324165344238,
+ "step": 683
+ },
+ {
+ "epoch": 9.502617801047121,
+ "grad_norm": 0.040689144283533096,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.639243125915527,
+ "step": 684
+ },
+ {
+ "epoch": 9.516579406631763,
+ "grad_norm": 0.04075361415743828,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.649250030517578,
+ "step": 685
+ },
+ {
+ "epoch": 9.530541012216405,
+ "grad_norm": 0.03914607688784599,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.66444730758667,
+ "step": 686
+ },
+ {
+ "epoch": 9.544502617801047,
+ "grad_norm": 0.033253733068704605,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.647335052490234,
+ "step": 687
+ },
+ {
+ "epoch": 9.55846422338569,
+ "grad_norm": 0.03179008886218071,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.6727800369262695,
+ "step": 688
+ },
+ {
+ "epoch": 9.572425828970331,
+ "grad_norm": 0.03477025404572487,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.655778884887695,
+ "step": 689
+ },
+ {
+ "epoch": 9.586387434554974,
+ "grad_norm": 0.030094800516963005,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.680851936340332,
+ "step": 690
+ },
+ {
+ "epoch": 9.600349040139616,
+ "grad_norm": 0.030153103172779083,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.654638290405273,
+ "step": 691
+ },
+ {
+ "epoch": 9.614310645724258,
+ "grad_norm": 0.029009278863668442,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.625916481018066,
+ "step": 692
+ },
+ {
+ "epoch": 9.6282722513089,
+ "grad_norm": 0.027248440310359,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.656370162963867,
+ "step": 693
+ },
+ {
+ "epoch": 9.642233856893542,
+ "grad_norm": 0.027727462351322174,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.631473541259766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656195462478184,
+ "grad_norm": 0.026710307225584984,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.636524200439453,
+ "step": 695
+ },
+ {
+ "epoch": 9.670157068062828,
+ "grad_norm": 0.02726181596517563,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.629701614379883,
+ "step": 696
+ },
+ {
+ "epoch": 9.68411867364747,
+ "grad_norm": 0.027146652340888977,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.640240669250488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698080279232112,
+ "grad_norm": 0.027165887877345085,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.64462947845459,
+ "step": 698
+ },
+ {
+ "epoch": 9.712041884816754,
+ "grad_norm": 0.03025982715189457,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.6271820068359375,
+ "step": 699
+ },
+ {
+ "epoch": 9.726003490401396,
+ "grad_norm": 0.02837025746703148,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.629697799682617,
+ "step": 700
+ },
+ {
+ "epoch": 9.739965095986038,
+ "grad_norm": 0.026777995750308037,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.646146774291992,
+ "step": 701
+ },
+ {
+ "epoch": 9.75392670157068,
+ "grad_norm": 0.024626802653074265,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.641671657562256,
+ "step": 702
+ },
+ {
+ "epoch": 9.767888307155323,
+ "grad_norm": 0.025329895317554474,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.626345157623291,
+ "step": 703
+ },
+ {
+ "epoch": 9.781849912739965,
+ "grad_norm": 0.027804844081401825,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.621457099914551,
+ "step": 704
+ },
+ {
+ "epoch": 9.795811518324607,
+ "grad_norm": 0.028763437643647194,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.6124114990234375,
+ "step": 705
+ },
+ {
+ "epoch": 9.809773123909249,
+ "grad_norm": 0.030769916251301765,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.604399681091309,
+ "step": 706
+ },
+ {
+ "epoch": 9.823734729493891,
+ "grad_norm": 0.03060949593782425,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.640804290771484,
+ "step": 707
+ },
+ {
+ "epoch": 9.837696335078533,
+ "grad_norm": 0.029832778498530388,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.593376159667969,
+ "step": 708
+ },
+ {
+ "epoch": 9.851657940663177,
+ "grad_norm": 0.02952776849269867,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.619533538818359,
+ "step": 709
+ },
+ {
+ "epoch": 9.86561954624782,
+ "grad_norm": 0.02894212119281292,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.640516757965088,
+ "step": 710
+ },
+ {
+ "epoch": 9.879581151832461,
+ "grad_norm": 0.026435624808073044,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.6342997550964355,
+ "step": 711
+ },
+ {
+ "epoch": 9.893542757417103,
+ "grad_norm": 0.02591666206717491,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.630884170532227,
+ "step": 712
+ },
+ {
+ "epoch": 9.907504363001745,
+ "grad_norm": 0.02702309936285019,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.632085800170898,
+ "step": 713
+ },
+ {
+ "epoch": 9.921465968586388,
+ "grad_norm": 0.026766657829284668,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.6231231689453125,
+ "step": 714
+ },
+ {
+ "epoch": 9.93542757417103,
+ "grad_norm": 0.026666559278964996,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.601619243621826,
+ "step": 715
+ },
+ {
+ "epoch": 9.949389179755672,
+ "grad_norm": 0.026801064610481262,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.641724109649658,
+ "step": 716
+ },
+ {
+ "epoch": 9.963350785340314,
+ "grad_norm": 0.03152798116207123,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.642511367797852,
+ "step": 717
+ },
+ {
+ "epoch": 9.977312390924956,
+ "grad_norm": 0.035406943410634995,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.635779857635498,
+ "step": 718
+ },
+ {
+ "epoch": 9.991273996509598,
+ "grad_norm": 0.037504684180021286,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.629463195800781,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02263174019753933,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 2.876767635345459,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.5942904949188232,
+ "eval_runtime": 60.2086,
+ "eval_samples_per_second": 40.559,
+ "eval_steps_per_second": 0.648,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03707852587103844,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.5765533447265625,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.05454680323600769,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.576363563537598,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.04832114651799202,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.586915969848633,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.050469402223825455,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.5265092849731445,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0622565895318985,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.592833518981934,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0640949010848999,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.5855560302734375,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.06502196192741394,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.591236114501953,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.06167149916291237,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.56747579574585,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.06747374683618546,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.60452938079834,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.07210271060466766,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.561588764190674,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.0733448714017868,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.610766410827637,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.0731528028845787,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.589570045471191,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.0822535902261734,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.631077289581299,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.06921929866075516,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.628293991088867,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.07443586736917496,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.627321243286133,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.07153813540935516,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.641068935394287,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.069196417927742,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.615875720977783,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06483697891235352,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.6165313720703125,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.05482150986790657,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.599890232086182,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.05364019051194191,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.61989164352417,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.053216829895973206,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.564305305480957,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05740506947040558,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.582683563232422,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.05587657541036606,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.61264705657959,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.050252825021743774,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.648050308227539,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.050955042243003845,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.615367889404297,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04903746768832207,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.617378234863281,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04252410680055618,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.609130859375,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.03801053389906883,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.625149250030518,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.03422372043132782,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.606106281280518,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.030507879331707954,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.6125993728637695,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.0340375155210495,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.5690202713012695,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.030507374554872513,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.6042985916137695,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027383938431739807,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.585570812225342,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.028792303055524826,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.566786766052246,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02866457775235176,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.566376686096191,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.028657952323555946,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.581750869750977,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.028227943927049637,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.576788902282715,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.024757765233516693,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.59373664855957,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.027442635968327522,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.570978164672852,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.02590034157037735,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.571286201477051,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.025696570053696632,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.542209625244141,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.025478003546595573,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.5900421142578125,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.02563696913421154,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.544609546661377,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.025326019152998924,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.5808515548706055,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.024502627551555634,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.560579776763916,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.02687518671154976,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.567154884338379,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.025280235335230827,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.560011863708496,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.026266315951943398,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.5758442878723145,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.026058891788125038,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.583309173583984,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.02620471641421318,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.547737121582031,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02626362442970276,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.556955337524414,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.024503158405423164,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.580735206604004,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02580532431602478,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.564000129699707,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.024437161162495613,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.545594215393066,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.026664750650525093,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.56415319442749,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.02499818615615368,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.550689697265625,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.027514580637216568,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.5548248291015625,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.026799403131008148,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.548295021057129,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.02685476280748844,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.5660600662231445,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.026104001328349113,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.556082248687744,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.027479661628603935,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.5580244064331055,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.02638247236609459,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.550073623657227,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.025853842496871948,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.578088760375977,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.02636682614684105,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.58807373046875,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.02684398740530014,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.561017036437988,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.028538964688777924,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.548762321472168,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.028327809646725655,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.548969745635986,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.025131452828645706,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.564047813415527,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.026498543098568916,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.553997039794922,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.030081963166594505,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.5787787437438965,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.03201213851571083,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.566424369812012,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.020518779754638672,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8421616554260254,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.5899484157562256,
+ "eval_runtime": 60.2671,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03692612797021866,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.492715835571289,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.050258126109838486,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.480323791503906,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.04179152473807335,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.484371185302734,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.04160204157233238,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.4905524253845215,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.03974860534071922,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.503803253173828,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.037031810730695724,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.494629859924316,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.038762833923101425,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.495111465454102,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.0412563718855381,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.490113735198975,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.041371624916791916,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.490643501281738,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04151557758450508,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.49881649017334,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04037577286362648,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.492178916931152,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.03676290065050125,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.477460861206055,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.038055796176195145,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.509150981903076,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.041152238845825195,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.47133731842041,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.03728092834353447,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.508072376251221,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.035338856279850006,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.505619049072266,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.037813492119312286,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.509489059448242,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.03789111226797104,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.508646488189697,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.03880828619003296,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.52567720413208,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.03611127287149429,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.498889446258545,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.036903638392686844,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.498601913452148,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03620576485991478,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.490840911865234,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03555903211236,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.506166458129883,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.03441924974322319,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.498001575469971,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.0334339402616024,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.491082191467285,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.031418606638908386,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.527408599853516,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03073921985924244,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.514107704162598,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.029589885845780373,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.48323917388916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.02974501997232437,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.51338529586792,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.029247552156448364,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.504205226898193,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.03132032975554466,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.505674362182617,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03088037669658661,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.477198600769043,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.031054604798555374,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.498721122741699,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03293661028146744,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.527826309204102,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.03181406110525131,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.495999336242676,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.029380083084106445,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.530351638793945,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.029296347871422768,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.492367744445801,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.02958036959171295,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.493037700653076,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03078395500779152,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.47504186630249,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03280661255121231,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.518409729003906,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.03323954716324806,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.460753917694092,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.032254140824079514,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.474285125732422,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.03018898330628872,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.5162353515625,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.02802245132625103,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.538684844970703,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.02821732498705387,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.4850287437438965,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.029189350083470345,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.50770378112793,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.02851196378469467,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.499988555908203,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.02741030789911747,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.499852180480957,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.028708625584840775,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.465211868286133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.028519680723547935,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.502464294433594,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.027846841141581535,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.503248691558838,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030456233769655228,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.490391731262207,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029952052980661392,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.50056266784668,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.02925778552889824,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.49215030670166,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.029679695144295692,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.523841857910156,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.031178541481494904,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.499871730804443,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.031264498829841614,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.50546932220459,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03482109308242798,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.494480133056641,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.03341105580329895,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.532899856567383,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028623664751648903,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.5016279220581055,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028600146993994713,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.511589050292969,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.029127901419997215,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.502058982849121,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.027898237109184265,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.493182182312012,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.028683282434940338,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.519793510437012,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.02887023612856865,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.457289695739746,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.02811296097934246,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.49398946762085,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.028485111892223358,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.478344440460205,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02914245054125786,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.487666130065918,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.028030497953295708,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.518488883972168,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.027973609045147896,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.515271186828613,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.027458522468805313,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.490856647491455,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.01909293793141842,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.822120189666748,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5866912603378296,
+ "eval_runtime": 61.3185,
+ "eval_samples_per_second": 39.825,
+ "eval_steps_per_second": 0.636,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03683992475271225,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.438363075256348,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.05042923241853714,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.443560600280762,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.055914171040058136,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.422282695770264,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.057007886469364166,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.442644119262695,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.05999710038304329,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.44780969619751,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06472347676753998,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.485074996948242,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.06923902779817581,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.457515716552734,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.0710320770740509,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.465476989746094,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.06942574679851532,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.426226615905762,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.06867868453264236,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.438092231750488,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.0688796266913414,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.446664810180664,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.0690719485282898,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.456482887268066,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.06068912148475647,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.4730730056762695,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.05411486327648163,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.4596452713012695,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.056623928248882294,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.418197154998779,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.05634097009897232,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.482892036437988,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.05056057870388031,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.457589149475098,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.04460986703634262,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.451439380645752,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.045022401958703995,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.452855110168457,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.04054102674126625,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.417323112487793,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.03682882338762283,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.465713977813721,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.03528819605708122,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.4502339363098145,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.03568802773952484,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.422399520874023,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.0363088883459568,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.447678089141846,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.03460463136434555,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.437664985656738,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.034751344472169876,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.489151954650879,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.033658601343631744,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.4471635818481445,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.03433239087462425,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.455332279205322,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.032074104994535446,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.435157775878906,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.032168325036764145,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.435118675231934,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.03007678873836994,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.42781925201416,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03094375506043434,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.433105945587158,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.032186444848775864,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.383000373840332,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.030610304325819016,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.473257064819336,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.030257435515522957,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.453041076660156,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.0301323514431715,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.396578788757324,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.030934512615203857,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.445223808288574,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.029907869175076485,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.442287445068359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03163841366767883,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.434650421142578,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.029627414420247078,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.454374313354492,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.031026875600218773,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.413784027099609,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03099491260945797,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.458438873291016,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.02981731668114662,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.438560962677002,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.02927570603787899,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.44493293762207,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.029779717326164246,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.445121765136719,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.02909068390727043,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.4736127853393555,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.029217200353741646,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.434604644775391,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.029872911050915718,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.454591751098633,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.028742222115397453,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.468222141265869,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.030313383787870407,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.45759391784668,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.03014662116765976,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.446331977844238,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029481127858161926,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.478209495544434,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.028485266491770744,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.444445610046387,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.02857700176537037,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.430604934692383,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.029038602486252785,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.43759298324585,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.029576119035482407,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.474579811096191,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.02897879108786583,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.440483093261719,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.029799485579133034,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.4839277267456055,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.030710609629750252,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.466696739196777,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.030270954594016075,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.455197334289551,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.03170182928442955,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.454664707183838,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030684730038046837,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.437619209289551,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.030972573906183243,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.431796073913574,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028914963826537132,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.460618019104004,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.029810264706611633,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.407312393188477,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.029806939885020256,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.466520309448242,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.03020886518061161,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.427349090576172,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030598826706409454,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.427891731262207,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03192012757062912,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.448113918304443,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.030492989346385002,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.450060844421387,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03188254311680794,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.43007230758667,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02081947773694992,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.782536506652832,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5860040187835693,
+ "eval_runtime": 60.6725,
+ "eval_samples_per_second": 40.249,
+ "eval_steps_per_second": 0.643,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.036721356213092804,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.358441352844238,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.056633755564689636,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.361527919769287,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.06175022944808006,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.3827924728393555,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05159056559205055,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.354856491088867,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.04885871335864067,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.359966278076172,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.04407693073153496,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.364612579345703,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.04891050234436989,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.344021320343018,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.04643657058477402,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.366066932678223,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.05241481214761734,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.3390607833862305,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.05837438255548477,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.390261650085449,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.058482322841882706,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.3856201171875,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.06507553160190582,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.414223670959473,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.0575365275144577,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.402576446533203,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05055592209100723,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.389563083648682,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.047167956829071045,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.402750015258789,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.04473405331373215,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.358315467834473,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.042440690100193024,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.392551898956299,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04465719312429428,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.373558044433594,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.04337528720498085,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.37944221496582,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.04322915896773338,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.38493013381958,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.039873406291007996,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.389793395996094,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.0413178876042366,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.427964210510254,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.04567280039191246,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.404172897338867,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04154926910996437,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.391359329223633,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.03774876892566681,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.381457328796387,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03588850423693657,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.392526626586914,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.03757261857390404,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.39273738861084,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03849692642688751,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.409083366394043,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03787294402718544,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.3776044845581055,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.034496866166591644,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.413462162017822,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.03973205015063286,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.359318733215332,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.036192040890455246,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.392844200134277,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.03436632081866264,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.421384811401367,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.035658642649650574,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.386757850646973,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.03434896469116211,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.364978313446045,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03642290085554123,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.385326385498047,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.03427736461162567,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.406352996826172,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.034845270216464996,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.370953559875488,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03390975296497345,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.3799943923950195,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.03134961053729057,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.419156551361084,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.03512260690331459,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.404660701751709,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03319067507982254,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.388535022735596,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03113507106900215,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.403522968292236,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03559132665395737,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.382323265075684,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.03579750657081604,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.37620210647583,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.033675212413072586,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.411443710327148,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.03220067173242569,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.443470001220703,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031610481441020966,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.406038284301758,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.02958764135837555,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.41604471206665,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.0318237729370594,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4253387451171875,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03260074555873871,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.370792388916016,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.03359333053231239,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.397639274597168,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032516930252313614,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.4140777587890625,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.03203396871685982,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.408872604370117,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.03201935440301895,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.402975082397461,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.03330211713910103,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.394018173217773,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.0348181314766407,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.380587100982666,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.03174913674592972,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.372378349304199,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.029864462092518806,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.42018985748291,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.033312153071165085,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.385973930358887,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03203144669532776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.395167827606201,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.031578030437231064,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.3982062339782715,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.030885562300682068,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.400731086730957,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03235667943954468,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.389830589294434,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.032664328813552856,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.401277542114258,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.030354196205735207,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.422520637512207,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.031162403523921967,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.393032073974609,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.032187070697546005,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.464164733886719,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.03445067256689072,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.368991851806641,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03398612141609192,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.418752670288086,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.029730670154094696,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.41429328918457,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022440090775489807,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7315595149993896,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5853753089904785,
+ "eval_runtime": 60.7174,
+ "eval_samples_per_second": 40.219,
+ "eval_steps_per_second": 0.642,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.03754177689552307,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.304028511047363,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.051616180688142776,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.34659481048584,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055258627980947495,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.310445785522461,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.049489911645650864,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3416619300842285,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05377533286809921,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.3143310546875,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.05443088337779045,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.320600986480713,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.050222691148519516,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.334789752960205,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.051904670894145966,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.316756248474121,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.05376426503062248,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.274116516113281,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.053313180804252625,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.321177005767822,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.058234598487615585,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.3169264793396,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05730101093649864,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.307013511657715,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.058140840381383896,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.332961082458496,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05654815956950188,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.315462112426758,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.05129384249448776,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.338476181030273,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.047652702778577805,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.321232795715332,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.050912003964185715,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.318971633911133,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.0514022521674633,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.3334150314331055,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.05457981303334236,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.325802803039551,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04678885266184807,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.307281494140625,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.05048161372542381,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.299831390380859,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.05334261804819107,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.313231468200684,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.05055771768093109,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.36859130859375,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04955561086535454,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.301530361175537,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.04548008739948273,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.339517593383789,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.046585116535425186,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.333036422729492,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.0503716766834259,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.364710807800293,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.040569406002759933,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.369089126586914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.03754054754972458,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.3349432945251465,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03775954991579056,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.333134174346924,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.04066266492009163,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.354358673095703,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.040490977466106415,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.360981464385986,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.0426657609641552,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.357726097106934,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.0415605828166008,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.363457679748535,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.04020509123802185,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.372447967529297,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.04101516678929329,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.292452812194824,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.03546427935361862,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.33892297744751,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.03661105036735535,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.329778671264648,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03565476834774017,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.351012229919434,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.03591311350464821,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.376980781555176,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.03665630891919136,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.354767799377441,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.0371236577630043,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.374916076660156,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.03464849665760994,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.3713059425354,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03345152735710144,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.350670337677002,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.03357568010687828,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.3573713302612305,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03503486514091492,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.362038612365723,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.03455005958676338,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.3263773918151855,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.03173859417438507,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.351345062255859,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03301836550235748,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.398364067077637,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.033733855932950974,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.335268020629883,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034080155193805695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.371492385864258,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.034396376460790634,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.382076740264893,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.03303401917219162,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.36623477935791,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03297369182109833,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3148393630981445,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.034819211810827255,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.358110427856445,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.033389780670404434,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.367427825927734,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.032363444566726685,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.3813676834106445,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03509577363729477,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.386520862579346,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.034685730934143066,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.33738374710083,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.034755002707242966,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.360980987548828,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03461688384413719,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.396650314331055,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03392560034990311,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.356130123138428,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.03342147544026375,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.359145164489746,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.032852694392204285,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.3796162605285645,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03472859039902687,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.361110687255859,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03277098760008812,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.361811637878418,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.03341882303357124,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.369524955749512,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.032347049564123154,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.369096755981445,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03161193057894707,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.339844703674316,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.03322301432490349,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.376675605773926,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03419092297554016,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.346864223480225,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02461826801300049,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7236275672912598,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.5859372019767761,
+ "eval_runtime": 59.0139,
+ "eval_samples_per_second": 41.38,
+ "eval_steps_per_second": 0.661,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013961605584642,
+ "grad_norm": 0.039559684693813324,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 4.289841651916504,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027923211169284,
+ "grad_norm": 0.05554218962788582,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 4.255925178527832,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041884816753926,
+ "grad_norm": 0.05330996215343475,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 4.272125244140625,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055846422338568,
+ "grad_norm": 0.053247954696416855,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 4.290765285491943,
+ "step": 1084
+ },
+ {
+ "epoch": 15.06980802792321,
+ "grad_norm": 0.05471150949597359,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 4.271744728088379,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083769633507853,
+ "grad_norm": 0.05976719409227371,
+ "learning_rate": 0.000491285979372924,
+ "loss": 4.28203821182251,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097731239092496,
+ "grad_norm": 0.05369880795478821,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 4.275313377380371,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111692844677139,
+ "grad_norm": 0.052529267966747284,
+ "learning_rate": 0.000490873338088198,
+ "loss": 4.25678825378418,
+ "step": 1088
+ },
+ {
+ "epoch": 15.12565445026178,
+ "grad_norm": 0.05184522643685341,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 4.260510444641113,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139616055846423,
+ "grad_norm": 0.051606856286525726,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 4.283233642578125,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153577661431065,
+ "grad_norm": 0.05318903923034668,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 4.286595344543457,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167539267015707,
+ "grad_norm": 0.058525051921606064,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 4.282374382019043,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181500872600349,
+ "grad_norm": 0.054776228964328766,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 4.2913126945495605,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195462478184991,
+ "grad_norm": 0.04907633364200592,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 4.297187805175781,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209424083769633,
+ "grad_norm": 0.05163136497139931,
+ "learning_rate": 0.000489424334303338,
+ "loss": 4.2968645095825195,
+ "step": 1095
+ },
+ {
+ "epoch": 15.223385689354275,
+ "grad_norm": 0.053157489746809006,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 4.285839080810547,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237347294938917,
+ "grad_norm": 0.05251535773277283,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 4.311246871948242,
+ "step": 1097
+ },
+ {
+ "epoch": 15.25130890052356,
+ "grad_norm": 0.048785097897052765,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 4.300364971160889,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265270506108202,
+ "grad_norm": 0.050393108278512955,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 4.299890995025635,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279232111692846,
+ "grad_norm": 0.04894037917256355,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 4.312449932098389,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293193717277488,
+ "grad_norm": 0.05274952948093414,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 4.2889814376831055,
+ "step": 1101
+ },
+ {
+ "epoch": 15.30715532286213,
+ "grad_norm": 0.05946570262312889,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 4.2731170654296875,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321116928446772,
+ "grad_norm": 0.054245397448539734,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 4.272463798522949,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335078534031414,
+ "grad_norm": 0.042092181742191315,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 4.299953937530518,
+ "step": 1104
+ },
+ {
+ "epoch": 15.349040139616056,
+ "grad_norm": 0.040222618728876114,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 4.292524814605713,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363001745200698,
+ "grad_norm": 0.04298696666955948,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 4.31342887878418,
+ "step": 1106
+ },
+ {
+ "epoch": 15.37696335078534,
+ "grad_norm": 0.04271050915122032,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 4.277663707733154,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390924956369982,
+ "grad_norm": 0.042364370077848434,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 4.311102390289307,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404886561954624,
+ "grad_norm": 0.04343319311738014,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 4.311610221862793,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418848167539267,
+ "grad_norm": 0.041321512311697006,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 4.29677677154541,
+ "step": 1110
+ },
+ {
+ "epoch": 15.432809773123909,
+ "grad_norm": 0.04107942432165146,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 4.3163018226623535,
+ "step": 1111
+ },
+ {
+ "epoch": 15.44677137870855,
+ "grad_norm": 0.03844515234231949,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 4.311301231384277,
+ "step": 1112
+ },
+ {
+ "epoch": 15.460732984293193,
+ "grad_norm": 0.03731007128953934,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 4.31325626373291,
+ "step": 1113
+ },
+ {
+ "epoch": 15.474694589877837,
+ "grad_norm": 0.03807826340198517,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 4.296201705932617,
+ "step": 1114
+ },
+ {
+ "epoch": 15.488656195462479,
+ "grad_norm": 0.040137600153684616,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 4.320395469665527,
+ "step": 1115
+ },
+ {
+ "epoch": 15.502617801047121,
+ "grad_norm": 0.038484394550323486,
+ "learning_rate": 0.000485033394934934,
+ "loss": 4.313858985900879,
+ "step": 1116
+ },
+ {
+ "epoch": 15.516579406631763,
+ "grad_norm": 0.03603338822722435,
+ "learning_rate": 0.000484822676912974,
+ "loss": 4.297666549682617,
+ "step": 1117
+ },
+ {
+ "epoch": 15.530541012216405,
+ "grad_norm": 0.038647182285785675,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 4.305002689361572,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544502617801047,
+ "grad_norm": 0.03907139599323273,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 4.329583644866943,
+ "step": 1119
+ },
+ {
+ "epoch": 15.55846422338569,
+ "grad_norm": 0.03621622920036316,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 4.290896415710449,
+ "step": 1120
+ },
+ {
+ "epoch": 15.572425828970331,
+ "grad_norm": 0.037496887147426605,
+ "learning_rate": 0.000483978341612154,
+ "loss": 4.33336877822876,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586387434554974,
+ "grad_norm": 0.03778311610221863,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 4.285354137420654,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600349040139616,
+ "grad_norm": 0.03782997280359268,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 4.29697847366333,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614310645724258,
+ "grad_norm": 0.03665657714009285,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 4.320103168487549,
+ "step": 1124
+ },
+ {
+ "epoch": 15.6282722513089,
+ "grad_norm": 0.03796127438545227,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 4.317153453826904,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642233856893542,
+ "grad_norm": 0.036425214260816574,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 4.293642520904541,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656195462478184,
+ "grad_norm": 0.03629784658551216,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 4.291100025177002,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670157068062828,
+ "grad_norm": 0.03674762696027756,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 4.295144557952881,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68411867364747,
+ "grad_norm": 0.036291178315877914,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 4.313133239746094,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698080279232112,
+ "grad_norm": 0.03817860782146454,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 4.294533729553223,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712041884816754,
+ "grad_norm": 0.036819685250520706,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 4.299263954162598,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726003490401396,
+ "grad_norm": 0.035103823989629745,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 4.287391185760498,
+ "step": 1132
+ },
+ {
+ "epoch": 15.739965095986038,
+ "grad_norm": 0.035701096057891846,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 4.318216323852539,
+ "step": 1133
+ },
+ {
+ "epoch": 15.75392670157068,
+ "grad_norm": 0.03639856353402138,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 4.301465034484863,
+ "step": 1134
+ },
+ {
+ "epoch": 15.767888307155323,
+ "grad_norm": 0.03593013063073158,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 4.311532974243164,
+ "step": 1135
+ },
+ {
+ "epoch": 15.781849912739965,
+ "grad_norm": 0.03564770147204399,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 4.3400726318359375,
+ "step": 1136
+ },
+ {
+ "epoch": 15.795811518324607,
+ "grad_norm": 0.03598494827747345,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 4.30976676940918,
+ "step": 1137
+ },
+ {
+ "epoch": 15.809773123909249,
+ "grad_norm": 0.034391652792692184,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 4.328464508056641,
+ "step": 1138
+ },
+ {
+ "epoch": 15.823734729493891,
+ "grad_norm": 0.03445672616362572,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 4.310579299926758,
+ "step": 1139
+ },
+ {
+ "epoch": 15.837696335078533,
+ "grad_norm": 0.03643282502889633,
+ "learning_rate": 0.000479936075927473,
+ "loss": 4.320905685424805,
+ "step": 1140
+ },
+ {
+ "epoch": 15.851657940663177,
+ "grad_norm": 0.03489850088953972,
+ "learning_rate": 0.000479721889242305,
+ "loss": 4.317229270935059,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86561954624782,
+ "grad_norm": 0.03609171137213707,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 4.309203147888184,
+ "step": 1142
+ },
+ {
+ "epoch": 15.879581151832461,
+ "grad_norm": 0.03705338388681412,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 4.287387371063232,
+ "step": 1143
+ },
+ {
+ "epoch": 15.893542757417103,
+ "grad_norm": 0.03697717934846878,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 4.302860260009766,
+ "step": 1144
+ },
+ {
+ "epoch": 15.907504363001745,
+ "grad_norm": 0.035338208079338074,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 4.3152265548706055,
+ "step": 1145
+ },
+ {
+ "epoch": 15.921465968586388,
+ "grad_norm": 0.03548634052276611,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 4.327335357666016,
+ "step": 1146
+ },
+ {
+ "epoch": 15.93542757417103,
+ "grad_norm": 0.035512328147888184,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 4.3067803382873535,
+ "step": 1147
+ },
+ {
+ "epoch": 15.949389179755672,
+ "grad_norm": 0.03799289092421532,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 4.320701599121094,
+ "step": 1148
+ },
+ {
+ "epoch": 15.963350785340314,
+ "grad_norm": 0.03866339474916458,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 4.321087837219238,
+ "step": 1149
+ },
+ {
+ "epoch": 15.977312390924956,
+ "grad_norm": 0.036002084612846375,
+ "learning_rate": 0.000477787822648007,
+ "loss": 4.287045478820801,
+ "step": 1150
+ },
+ {
+ "epoch": 15.991273996509598,
+ "grad_norm": 0.03433441370725632,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 4.2562665939331055,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.02580358274281025,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 2.67783260345459,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.5870338082313538,
+ "eval_runtime": 60.0469,
+ "eval_samples_per_second": 40.668,
+ "eval_steps_per_second": 0.649,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013961605584644,
+ "grad_norm": 0.04105810821056366,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 4.224295616149902,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027923211169284,
+ "grad_norm": 0.049538832157850266,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 4.222184181213379,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041884816753928,
+ "grad_norm": 0.05089569091796875,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 4.2067060470581055,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05584642233857,
+ "grad_norm": 0.04600589722394943,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 4.238527774810791,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069808027923212,
+ "grad_norm": 0.043311409652233124,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 4.242960453033447,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083769633507853,
+ "grad_norm": 0.047897014766931534,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 4.199069023132324,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097731239092496,
+ "grad_norm": 0.05633604899048805,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 4.211665153503418,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111692844677137,
+ "grad_norm": 0.056110311299562454,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 4.20163631439209,
+ "step": 1160
+ },
+ {
+ "epoch": 16.12565445026178,
+ "grad_norm": 0.05426819995045662,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 4.21889591217041,
+ "step": 1161
+ },
+ {
+ "epoch": 16.13961605584642,
+ "grad_norm": 0.05489835515618324,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 4.220406532287598,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153577661431065,
+ "grad_norm": 0.053941354155540466,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 4.231909275054932,
+ "step": 1163
+ },
+ {
+ "epoch": 16.167539267015705,
+ "grad_norm": 0.058548808097839355,
+ "learning_rate": 0.000474756648870413,
+ "loss": 4.237003326416016,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18150087260035,
+ "grad_norm": 0.05370597913861275,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 4.240085124969482,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195462478184993,
+ "grad_norm": 0.05684991553425789,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 4.203842639923096,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209424083769633,
+ "grad_norm": 0.060745254158973694,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 4.247998237609863,
+ "step": 1167
+ },
+ {
+ "epoch": 16.223385689354277,
+ "grad_norm": 0.058296725153923035,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 4.226161003112793,
+ "step": 1168
+ },
+ {
+ "epoch": 16.237347294938917,
+ "grad_norm": 0.05310584232211113,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 4.229870796203613,
+ "step": 1169
+ },
+ {
+ "epoch": 16.25130890052356,
+ "grad_norm": 0.04710390046238899,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 4.240084648132324,
+ "step": 1170
+ },
+ {
+ "epoch": 16.2652705061082,
+ "grad_norm": 0.05003240332007408,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 4.240972518920898,
+ "step": 1171
+ },
+ {
+ "epoch": 16.279232111692846,
+ "grad_norm": 0.05552851781249046,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 4.248024940490723,
+ "step": 1172
+ },
+ {
+ "epoch": 16.293193717277486,
+ "grad_norm": 0.049967192113399506,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 4.279229164123535,
+ "step": 1173
+ },
+ {
+ "epoch": 16.30715532286213,
+ "grad_norm": 0.04883681237697601,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 4.269807815551758,
+ "step": 1174
+ },
+ {
+ "epoch": 16.32111692844677,
+ "grad_norm": 0.04931480810046196,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 4.232419490814209,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335078534031414,
+ "grad_norm": 0.048746511340141296,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 4.267911911010742,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349040139616054,
+ "grad_norm": 0.04678189381957054,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 4.262979507446289,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363001745200698,
+ "grad_norm": 0.048893529921770096,
+ "learning_rate": 0.000471698352726896,
+ "loss": 4.254408359527588,
+ "step": 1178
+ },
+ {
+ "epoch": 16.376963350785342,
+ "grad_norm": 0.045027732849121094,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 4.2445902824401855,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390924956369982,
+ "grad_norm": 0.04324870929121971,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 4.2685136795043945,
+ "step": 1180
+ },
+ {
+ "epoch": 16.404886561954626,
+ "grad_norm": 0.0470522940158844,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 4.237505912780762,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418848167539267,
+ "grad_norm": 0.04896962642669678,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 4.2523345947265625,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43280977312391,
+ "grad_norm": 0.04562787711620331,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 4.2192792892456055,
+ "step": 1183
+ },
+ {
+ "epoch": 16.44677137870855,
+ "grad_norm": 0.04414575919508934,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 4.250238418579102,
+ "step": 1184
+ },
+ {
+ "epoch": 16.460732984293195,
+ "grad_norm": 0.04363945871591568,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 4.256250381469727,
+ "step": 1185
+ },
+ {
+ "epoch": 16.474694589877835,
+ "grad_norm": 0.04159797728061676,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 4.274903297424316,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48865619546248,
+ "grad_norm": 0.04248544201254845,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 4.278222560882568,
+ "step": 1187
+ },
+ {
+ "epoch": 16.50261780104712,
+ "grad_norm": 0.04379890486598015,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 4.267505645751953,
+ "step": 1188
+ },
+ {
+ "epoch": 16.516579406631763,
+ "grad_norm": 0.04319411888718605,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 4.2607269287109375,
+ "step": 1189
+ },
+ {
+ "epoch": 16.530541012216403,
+ "grad_norm": 0.04143810644745827,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 4.267977714538574,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544502617801047,
+ "grad_norm": 0.03914840891957283,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 4.260498523712158,
+ "step": 1191
+ },
+ {
+ "epoch": 16.55846422338569,
+ "grad_norm": 0.04107543081045151,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 4.2550482749938965,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57242582897033,
+ "grad_norm": 0.041316065937280655,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 4.3075456619262695,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586387434554975,
+ "grad_norm": 0.04069342464208603,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 4.234392166137695,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600349040139616,
+ "grad_norm": 0.03816857561469078,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 4.251038551330566,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61431064572426,
+ "grad_norm": 0.04095374792814255,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 4.284790992736816,
+ "step": 1196
+ },
+ {
+ "epoch": 16.6282722513089,
+ "grad_norm": 0.03788160905241966,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 4.255716323852539,
+ "step": 1197
+ },
+ {
+ "epoch": 16.642233856893544,
+ "grad_norm": 0.03784061595797539,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 4.269606590270996,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656195462478184,
+ "grad_norm": 0.04087168723344803,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 4.231165885925293,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670157068062828,
+ "grad_norm": 0.040546853095293045,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 4.288422584533691,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68411867364747,
+ "grad_norm": 0.03777262568473816,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 4.277492523193359,
+ "step": 1201
+ },
+ {
+ "epoch": 16.698080279232112,
+ "grad_norm": 0.039857733994722366,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 4.239383697509766,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712041884816752,
+ "grad_norm": 0.04387792944908142,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 4.300265312194824,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726003490401396,
+ "grad_norm": 0.04214833676815033,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 4.255545616149902,
+ "step": 1204
+ },
+ {
+ "epoch": 16.739965095986037,
+ "grad_norm": 0.04290056228637695,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 4.275530815124512,
+ "step": 1205
+ },
+ {
+ "epoch": 16.75392670157068,
+ "grad_norm": 0.03975554555654526,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 4.282685279846191,
+ "step": 1206
+ },
+ {
+ "epoch": 16.767888307155324,
+ "grad_norm": 0.039464011788368225,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 4.285504341125488,
+ "step": 1207
+ },
+ {
+ "epoch": 16.781849912739965,
+ "grad_norm": 0.039714012295007706,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 4.273477554321289,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79581151832461,
+ "grad_norm": 0.03936111181974411,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 4.304205894470215,
+ "step": 1209
+ },
+ {
+ "epoch": 16.80977312390925,
+ "grad_norm": 0.03782064467668533,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 4.307382583618164,
+ "step": 1210
+ },
+ {
+ "epoch": 16.823734729493893,
+ "grad_norm": 0.040340717881917953,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 4.308620452880859,
+ "step": 1211
+ },
+ {
+ "epoch": 16.837696335078533,
+ "grad_norm": 0.0403442308306694,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 4.248626708984375,
+ "step": 1212
+ },
+ {
+ "epoch": 16.851657940663177,
+ "grad_norm": 0.038351960480213165,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 4.261248588562012,
+ "step": 1213
+ },
+ {
+ "epoch": 16.865619546247817,
+ "grad_norm": 0.03965333476662636,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 4.278291702270508,
+ "step": 1214
+ },
+ {
+ "epoch": 16.87958115183246,
+ "grad_norm": 0.040909092873334885,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 4.241847991943359,
+ "step": 1215
+ },
+ {
+ "epoch": 16.8935427574171,
+ "grad_norm": 0.038684047758579254,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 4.281346321105957,
+ "step": 1216
+ },
+ {
+ "epoch": 16.907504363001745,
+ "grad_norm": 0.03698103874921799,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 4.3085432052612305,
+ "step": 1217
+ },
+ {
+ "epoch": 16.921465968586386,
+ "grad_norm": 0.03908613696694374,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 4.24220085144043,
+ "step": 1218
+ },
+ {
+ "epoch": 16.93542757417103,
+ "grad_norm": 0.03873388469219208,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 4.282486915588379,
+ "step": 1219
+ },
+ {
+ "epoch": 16.949389179755673,
+ "grad_norm": 0.03709883987903595,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 4.279009819030762,
+ "step": 1220
+ },
+ {
+ "epoch": 16.963350785340314,
+ "grad_norm": 0.03937513753771782,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 4.273211479187012,
+ "step": 1221
+ },
+ {
+ "epoch": 16.977312390924958,
+ "grad_norm": 0.038478728383779526,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 4.281154155731201,
+ "step": 1222
+ },
+ {
+ "epoch": 16.991273996509598,
+ "grad_norm": 0.03913680836558342,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 4.269218444824219,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.02641749382019043,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 2.6905035972595215,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.5879416465759277,
+ "eval_runtime": 60.2807,
+ "eval_samples_per_second": 40.511,
+ "eval_steps_per_second": 0.647,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013961605584644,
+ "grad_norm": 0.048571839928627014,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 4.2008538246154785,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027923211169284,
+ "grad_norm": 0.0640900656580925,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 4.180826663970947,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041884816753928,
+ "grad_norm": 0.06458134204149246,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 4.199562072753906,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05584642233857,
+ "grad_norm": 0.0563119500875473,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 4.171557426452637,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069808027923212,
+ "grad_norm": 0.050914082676172256,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 4.183539390563965,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083769633507853,
+ "grad_norm": 0.053884491324424744,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 4.210450172424316,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097731239092496,
+ "grad_norm": 0.05301345884799957,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 4.1553192138671875,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111692844677137,
+ "grad_norm": 0.051138825714588165,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 4.158510208129883,
+ "step": 1232
+ },
+ {
+ "epoch": 17.12565445026178,
+ "grad_norm": 0.05355675891041756,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 4.167689323425293,
+ "step": 1233
+ },
+ {
+ "epoch": 17.13961605584642,
+ "grad_norm": 0.06062275916337967,
+ "learning_rate": 0.000459203665939781,
+ "loss": 4.170838356018066,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153577661431065,
+ "grad_norm": 0.058666616678237915,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 4.187441825866699,
+ "step": 1235
+ },
+ {
+ "epoch": 17.167539267015705,
+ "grad_norm": 0.057622168213129044,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 4.190746784210205,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18150087260035,
+ "grad_norm": 0.069090336561203,
+ "learning_rate": 0.000458522886595554,
+ "loss": 4.224946022033691,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195462478184993,
+ "grad_norm": 0.06552847474813461,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 4.204067230224609,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209424083769633,
+ "grad_norm": 0.06602177768945694,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 4.189286708831787,
+ "step": 1239
+ },
+ {
+ "epoch": 17.223385689354277,
+ "grad_norm": 0.06751416623592377,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 4.208927154541016,
+ "step": 1240
+ },
+ {
+ "epoch": 17.237347294938917,
+ "grad_norm": 0.0706794261932373,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 4.212923049926758,
+ "step": 1241
+ },
+ {
+ "epoch": 17.25130890052356,
+ "grad_norm": 0.05346836522221565,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 4.2221784591674805,
+ "step": 1242
+ },
+ {
+ "epoch": 17.2652705061082,
+ "grad_norm": 0.05395734682679176,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 4.196428298950195,
+ "step": 1243
+ },
+ {
+ "epoch": 17.279232111692846,
+ "grad_norm": 0.05496550351381302,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 4.208031177520752,
+ "step": 1244
+ },
+ {
+ "epoch": 17.293193717277486,
+ "grad_norm": 0.05320895463228226,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 4.232810974121094,
+ "step": 1245
+ },
+ {
+ "epoch": 17.30715532286213,
+ "grad_norm": 0.05064759403467178,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 4.207826614379883,
+ "step": 1246
+ },
+ {
+ "epoch": 17.32111692844677,
+ "grad_norm": 0.04287239536643028,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 4.195648670196533,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335078534031414,
+ "grad_norm": 0.04527301341295242,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 4.19954776763916,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349040139616054,
+ "grad_norm": 0.04421805590391159,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 4.248671054840088,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363001745200698,
+ "grad_norm": 0.04487175494432449,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 4.2072954177856445,
+ "step": 1250
+ },
+ {
+ "epoch": 17.376963350785342,
+ "grad_norm": 0.04415794834494591,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 4.222837448120117,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390924956369982,
+ "grad_norm": 0.04474013298749924,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 4.213104248046875,
+ "step": 1252
+ },
+ {
+ "epoch": 17.404886561954626,
+ "grad_norm": 0.042631037533283234,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 4.209726333618164,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418848167539267,
+ "grad_norm": 0.04424271360039711,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 4.202699661254883,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43280977312391,
+ "grad_norm": 0.04151812940835953,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 4.190732955932617,
+ "step": 1255
+ },
+ {
+ "epoch": 17.44677137870855,
+ "grad_norm": 0.041588131338357925,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 4.222360610961914,
+ "step": 1256
+ },
+ {
+ "epoch": 17.460732984293195,
+ "grad_norm": 0.042378414422273636,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 4.198424339294434,
+ "step": 1257
+ },
+ {
+ "epoch": 17.474694589877835,
+ "grad_norm": 0.041907574981451035,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 4.224804401397705,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48865619546248,
+ "grad_norm": 0.043774355202913284,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 4.2276611328125,
+ "step": 1259
+ },
+ {
+ "epoch": 17.50261780104712,
+ "grad_norm": 0.04238573834300041,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 4.228109359741211,
+ "step": 1260
+ },
+ {
+ "epoch": 17.516579406631763,
+ "grad_norm": 0.045351386070251465,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 4.208807468414307,
+ "step": 1261
+ },
+ {
+ "epoch": 17.530541012216403,
+ "grad_norm": 0.04300729185342789,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 4.257306098937988,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544502617801047,
+ "grad_norm": 0.041442908346652985,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 4.208004951477051,
+ "step": 1263
+ },
+ {
+ "epoch": 17.55846422338569,
+ "grad_norm": 0.04337337985634804,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 4.23393440246582,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57242582897033,
+ "grad_norm": 0.04335462674498558,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 4.228087425231934,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586387434554975,
+ "grad_norm": 0.03987223282456398,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 4.213971138000488,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600349040139616,
+ "grad_norm": 0.041760094463825226,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 4.248366355895996,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61431064572426,
+ "grad_norm": 0.04420867934823036,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 4.278865814208984,
+ "step": 1268
+ },
+ {
+ "epoch": 17.6282722513089,
+ "grad_norm": 0.04244982823729515,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 4.223642349243164,
+ "step": 1269
+ },
+ {
+ "epoch": 17.642233856893544,
+ "grad_norm": 0.04280664771795273,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 4.231023788452148,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656195462478184,
+ "grad_norm": 0.04294133558869362,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 4.222380638122559,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670157068062828,
+ "grad_norm": 0.04179249703884125,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 4.215641975402832,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68411867364747,
+ "grad_norm": 0.04038798063993454,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 4.228580474853516,
+ "step": 1273
+ },
+ {
+ "epoch": 17.698080279232112,
+ "grad_norm": 0.04211528226733208,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 4.245234489440918,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712041884816752,
+ "grad_norm": 0.04102989658713341,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 4.218495845794678,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726003490401396,
+ "grad_norm": 0.04038692265748978,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 4.2466511726379395,
+ "step": 1276
+ },
+ {
+ "epoch": 17.739965095986037,
+ "grad_norm": 0.03958791121840477,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 4.237762451171875,
+ "step": 1277
+ },
+ {
+ "epoch": 17.75392670157068,
+ "grad_norm": 0.04129907861351967,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 4.238739013671875,
+ "step": 1278
+ },
+ {
+ "epoch": 17.767888307155324,
+ "grad_norm": 0.042581796646118164,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 4.235558032989502,
+ "step": 1279
+ },
+ {
+ "epoch": 17.781849912739965,
+ "grad_norm": 0.04339978098869324,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 4.226593971252441,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79581151832461,
+ "grad_norm": 0.03808023780584335,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 4.238213062286377,
+ "step": 1281
+ },
+ {
+ "epoch": 17.80977312390925,
+ "grad_norm": 0.04022229090332985,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 4.214224815368652,
+ "step": 1282
+ },
+ {
+ "epoch": 17.823734729493893,
+ "grad_norm": 0.04379156231880188,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 4.253068447113037,
+ "step": 1283
+ },
+ {
+ "epoch": 17.837696335078533,
+ "grad_norm": 0.04232964664697647,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 4.249903202056885,
+ "step": 1284
+ },
+ {
+ "epoch": 17.851657940663177,
+ "grad_norm": 0.040980905294418335,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 4.252640724182129,
+ "step": 1285
+ },
+ {
+ "epoch": 17.865619546247817,
+ "grad_norm": 0.04015526548027992,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 4.234862327575684,
+ "step": 1286
+ },
+ {
+ "epoch": 17.87958115183246,
+ "grad_norm": 0.04273828864097595,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 4.216028213500977,
+ "step": 1287
+ },
+ {
+ "epoch": 17.8935427574171,
+ "grad_norm": 0.04210563376545906,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 4.23359489440918,
+ "step": 1288
+ },
+ {
+ "epoch": 17.907504363001745,
+ "grad_norm": 0.038766685873270035,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 4.220360279083252,
+ "step": 1289
+ },
+ {
+ "epoch": 17.921465968586386,
+ "grad_norm": 0.03921609744429588,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 4.253805160522461,
+ "step": 1290
+ },
+ {
+ "epoch": 17.93542757417103,
+ "grad_norm": 0.04071544483304024,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 4.211221694946289,
+ "step": 1291
+ },
+ {
+ "epoch": 17.949389179755673,
+ "grad_norm": 0.03996877372264862,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 4.241739273071289,
+ "step": 1292
+ },
+ {
+ "epoch": 17.963350785340314,
+ "grad_norm": 0.03878853842616081,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 4.237001419067383,
+ "step": 1293
+ },
+ {
+ "epoch": 17.977312390924958,
+ "grad_norm": 0.03889666125178337,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 4.244929790496826,
+ "step": 1294
+ },
+ {
+ "epoch": 17.991273996509598,
+ "grad_norm": 0.03876441717147827,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 4.236083030700684,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.027286000549793243,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 2.636739730834961,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.5896694660186768,
+ "eval_runtime": 60.4742,
+ "eval_samples_per_second": 40.381,
+ "eval_steps_per_second": 0.645,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013961605584644,
+ "grad_norm": 0.04279542714357376,
+ "learning_rate": 0.000444677103403854,
+ "loss": 4.136401653289795,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027923211169284,
+ "grad_norm": 0.051685452461242676,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 4.159584045410156,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041884816753928,
+ "grad_norm": 0.05411457270383835,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 4.160367965698242,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05584642233857,
+ "grad_norm": 0.048574138432741165,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 4.134189128875732,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069808027923212,
+ "grad_norm": 0.04528478533029556,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 4.1022725105285645,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083769633507853,
+ "grad_norm": 0.04474090039730072,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 4.16152286529541,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097731239092496,
+ "grad_norm": 0.04882791265845299,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 4.1553144454956055,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111692844677137,
+ "grad_norm": 0.04831654578447342,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 4.147022724151611,
+ "step": 1304
+ },
+ {
+ "epoch": 18.12565445026178,
+ "grad_norm": 0.047365427017211914,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 4.139405250549316,
+ "step": 1305
+ },
+ {
+ "epoch": 18.13961605584642,
+ "grad_norm": 0.0514012910425663,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 4.115568161010742,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153577661431065,
+ "grad_norm": 0.05225161090493202,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 4.166429042816162,
+ "step": 1307
+ },
+ {
+ "epoch": 18.167539267015705,
+ "grad_norm": 0.05575249716639519,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 4.1414289474487305,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18150087260035,
+ "grad_norm": 0.05405383184552193,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 4.138680934906006,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195462478184993,
+ "grad_norm": 0.05185031518340111,
+ "learning_rate": 0.000441621628895411,
+ "loss": 4.159243583679199,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209424083769633,
+ "grad_norm": 0.05070916935801506,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 4.1844892501831055,
+ "step": 1311
+ },
+ {
+ "epoch": 18.223385689354277,
+ "grad_norm": 0.05037331581115723,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 4.132781028747559,
+ "step": 1312
+ },
+ {
+ "epoch": 18.237347294938917,
+ "grad_norm": 0.05082111805677414,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 4.143031120300293,
+ "step": 1313
+ },
+ {
+ "epoch": 18.25130890052356,
+ "grad_norm": 0.04742169380187988,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 4.137580871582031,
+ "step": 1314
+ },
+ {
+ "epoch": 18.2652705061082,
+ "grad_norm": 0.049042731523513794,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 4.159721374511719,
+ "step": 1315
+ },
+ {
+ "epoch": 18.279232111692846,
+ "grad_norm": 0.045649368315935135,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 4.144008636474609,
+ "step": 1316
+ },
+ {
+ "epoch": 18.293193717277486,
+ "grad_norm": 0.0473787859082222,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 4.164243698120117,
+ "step": 1317
+ },
+ {
+ "epoch": 18.30715532286213,
+ "grad_norm": 0.047626055777072906,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 4.164441108703613,
+ "step": 1318
+ },
+ {
+ "epoch": 18.32111692844677,
+ "grad_norm": 0.04738573357462883,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 4.180338382720947,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335078534031414,
+ "grad_norm": 0.04712498560547829,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 4.146285533905029,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349040139616054,
+ "grad_norm": 0.045505691319704056,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 4.162572860717773,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363001745200698,
+ "grad_norm": 0.0489528588950634,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 4.160312175750732,
+ "step": 1322
+ },
+ {
+ "epoch": 18.376963350785342,
+ "grad_norm": 0.047517575323581696,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 4.164361953735352,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390924956369982,
+ "grad_norm": 0.044214535504579544,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 4.182154655456543,
+ "step": 1324
+ },
+ {
+ "epoch": 18.404886561954626,
+ "grad_norm": 0.043775733560323715,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 4.130522727966309,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418848167539267,
+ "grad_norm": 0.044364530593156815,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 4.151247024536133,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43280977312391,
+ "grad_norm": 0.046320103108882904,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 4.204159736633301,
+ "step": 1327
+ },
+ {
+ "epoch": 18.44677137870855,
+ "grad_norm": 0.04802565276622772,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 4.182550430297852,
+ "step": 1328
+ },
+ {
+ "epoch": 18.460732984293195,
+ "grad_norm": 0.049340926110744476,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 4.163658142089844,
+ "step": 1329
+ },
+ {
+ "epoch": 18.474694589877835,
+ "grad_norm": 0.046701837331056595,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 4.170445442199707,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48865619546248,
+ "grad_norm": 0.04746273159980774,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 4.220232963562012,
+ "step": 1331
+ },
+ {
+ "epoch": 18.50261780104712,
+ "grad_norm": 0.04644489288330078,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 4.190027236938477,
+ "step": 1332
+ },
+ {
+ "epoch": 18.516579406631763,
+ "grad_norm": 0.04394262284040451,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 4.159570693969727,
+ "step": 1333
+ },
+ {
+ "epoch": 18.530541012216403,
+ "grad_norm": 0.04609063267707825,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 4.1657586097717285,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544502617801047,
+ "grad_norm": 0.0450211800634861,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 4.184865474700928,
+ "step": 1335
+ },
+ {
+ "epoch": 18.55846422338569,
+ "grad_norm": 0.0475253090262413,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 4.185857772827148,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57242582897033,
+ "grad_norm": 0.04332203418016434,
+ "learning_rate": 0.000435215814386134,
+ "loss": 4.160165309906006,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586387434554975,
+ "grad_norm": 0.0427570678293705,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 4.1727495193481445,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600349040139616,
+ "grad_norm": 0.04651174694299698,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 4.204797744750977,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61431064572426,
+ "grad_norm": 0.04324701428413391,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 4.186931610107422,
+ "step": 1340
+ },
+ {
+ "epoch": 18.6282722513089,
+ "grad_norm": 0.044644393026828766,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 4.209554195404053,
+ "step": 1341
+ },
+ {
+ "epoch": 18.642233856893544,
+ "grad_norm": 0.042982228100299835,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 4.205570220947266,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656195462478184,
+ "grad_norm": 0.04385090991854668,
+ "learning_rate": 0.000433781621332752,
+ "loss": 4.177936553955078,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670157068062828,
+ "grad_norm": 0.04771510139107704,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 4.1795806884765625,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68411867364747,
+ "grad_norm": 0.04551433399319649,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 4.192194938659668,
+ "step": 1345
+ },
+ {
+ "epoch": 18.698080279232112,
+ "grad_norm": 0.043501488864421844,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 4.177781105041504,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712041884816752,
+ "grad_norm": 0.04394034296274185,
+ "learning_rate": 0.000432823382695327,
+ "loss": 4.206354141235352,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726003490401396,
+ "grad_norm": 0.04911153391003609,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 4.205874919891357,
+ "step": 1348
+ },
+ {
+ "epoch": 18.739965095986037,
+ "grad_norm": 0.04602045193314552,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 4.1947021484375,
+ "step": 1349
+ },
+ {
+ "epoch": 18.75392670157068,
+ "grad_norm": 0.04393904283642769,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 4.188337802886963,
+ "step": 1350
+ },
+ {
+ "epoch": 18.767888307155324,
+ "grad_norm": 0.04526393488049507,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 4.175235748291016,
+ "step": 1351
+ },
+ {
+ "epoch": 18.781849912739965,
+ "grad_norm": 0.041930247098207474,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 4.210986137390137,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79581151832461,
+ "grad_norm": 0.044262565672397614,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 4.176301002502441,
+ "step": 1353
+ },
+ {
+ "epoch": 18.80977312390925,
+ "grad_norm": 0.044190581887960434,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 4.1888837814331055,
+ "step": 1354
+ },
+ {
+ "epoch": 18.823734729493893,
+ "grad_norm": 0.045371655374765396,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 4.182244777679443,
+ "step": 1355
+ },
+ {
+ "epoch": 18.837696335078533,
+ "grad_norm": 0.04302282631397247,
+ "learning_rate": 0.000430661245733797,
+ "loss": 4.189350128173828,
+ "step": 1356
+ },
+ {
+ "epoch": 18.851657940663177,
+ "grad_norm": 0.04402372986078262,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 4.233372211456299,
+ "step": 1357
+ },
+ {
+ "epoch": 18.865619546247817,
+ "grad_norm": 0.04397045075893402,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 4.21682071685791,
+ "step": 1358
+ },
+ {
+ "epoch": 18.87958115183246,
+ "grad_norm": 0.04319734871387482,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 4.204183578491211,
+ "step": 1359
+ },
+ {
+ "epoch": 18.8935427574171,
+ "grad_norm": 0.04600509628653526,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 4.206524848937988,
+ "step": 1360
+ },
+ {
+ "epoch": 18.907504363001745,
+ "grad_norm": 0.04364705830812454,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 4.213656425476074,
+ "step": 1361
+ },
+ {
+ "epoch": 18.921465968586386,
+ "grad_norm": 0.044004205614328384,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 4.1841583251953125,
+ "step": 1362
+ },
+ {
+ "epoch": 18.93542757417103,
+ "grad_norm": 0.04480595886707306,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 4.166214942932129,
+ "step": 1363
+ },
+ {
+ "epoch": 18.949389179755673,
+ "grad_norm": 0.04260887950658798,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 4.198726654052734,
+ "step": 1364
+ },
+ {
+ "epoch": 18.963350785340314,
+ "grad_norm": 0.04513927176594734,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 4.2122392654418945,
+ "step": 1365
+ },
+ {
+ "epoch": 18.977312390924958,
+ "grad_norm": 0.04286042973399162,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 4.192383766174316,
+ "step": 1366
+ },
+ {
+ "epoch": 18.991273996509598,
+ "grad_norm": 0.04287346079945564,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 4.221860408782959,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.030670644715428352,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 2.624823570251465,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.5910831689834595,
+ "eval_runtime": 60.5517,
+ "eval_samples_per_second": 40.329,
+ "eval_steps_per_second": 0.644,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013961605584644,
+ "grad_norm": 0.046196095645427704,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 4.104501724243164,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027923211169284,
+ "grad_norm": 0.05348595231771469,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 4.109098434448242,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041884816753928,
+ "grad_norm": 0.05256905406713486,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 4.115832328796387,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05584642233857,
+ "grad_norm": 0.05241430550813675,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 4.079610824584961,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069808027923212,
+ "grad_norm": 0.04983370006084442,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 4.095335006713867,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083769633507853,
+ "grad_norm": 0.04900007322430611,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 4.093306541442871,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097731239092496,
+ "grad_norm": 0.04995666816830635,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 4.111114501953125,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111692844677137,
+ "grad_norm": 0.04875681549310684,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 4.091944694519043,
+ "step": 1376
+ },
+ {
+ "epoch": 19.12565445026178,
+ "grad_norm": 0.05283811315894127,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 4.112549304962158,
+ "step": 1377
+ },
+ {
+ "epoch": 19.13961605584642,
+ "grad_norm": 0.04801992326974869,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 4.108818054199219,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153577661431065,
+ "grad_norm": 0.04751930385828018,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 4.089441299438477,
+ "step": 1379
+ },
+ {
+ "epoch": 19.167539267015705,
+ "grad_norm": 0.045653726905584335,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 4.07722282409668,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18150087260035,
+ "grad_norm": 0.045616209506988525,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 4.097578048706055,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195462478184993,
+ "grad_norm": 0.04929365590214729,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 4.082221984863281,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209424083769633,
+ "grad_norm": 0.04915834963321686,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 4.123330116271973,
+ "step": 1383
+ },
+ {
+ "epoch": 19.223385689354277,
+ "grad_norm": 0.0470132976770401,
+ "learning_rate": 0.000423881957237287,
+ "loss": 4.113927841186523,
+ "step": 1384
+ },
+ {
+ "epoch": 19.237347294938917,
+ "grad_norm": 0.04695259779691696,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 4.096814155578613,
+ "step": 1385
+ },
+ {
+ "epoch": 19.25130890052356,
+ "grad_norm": 0.04730398207902908,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 4.122525215148926,
+ "step": 1386
+ },
+ {
+ "epoch": 19.2652705061082,
+ "grad_norm": 0.050295207649469376,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 4.117993354797363,
+ "step": 1387
+ },
+ {
+ "epoch": 19.279232111692846,
+ "grad_norm": 0.05082492157816887,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 4.14082145690918,
+ "step": 1388
+ },
+ {
+ "epoch": 19.293193717277486,
+ "grad_norm": 0.04960232973098755,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 4.139887809753418,
+ "step": 1389
+ },
+ {
+ "epoch": 19.30715532286213,
+ "grad_norm": 0.04851173236966133,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 4.138492584228516,
+ "step": 1390
+ },
+ {
+ "epoch": 19.32111692844677,
+ "grad_norm": 0.047838468104600906,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 4.118064880371094,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335078534031414,
+ "grad_norm": 0.05148220434784889,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 4.136533737182617,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349040139616054,
+ "grad_norm": 0.053177181631326675,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 4.1125168800354,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363001745200698,
+ "grad_norm": 0.04861171543598175,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 4.1437153816223145,
+ "step": 1394
+ },
+ {
+ "epoch": 19.376963350785342,
+ "grad_norm": 0.05194924771785736,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 4.097813606262207,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390924956369982,
+ "grad_norm": 0.049609847366809845,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 4.068940162658691,
+ "step": 1396
+ },
+ {
+ "epoch": 19.404886561954626,
+ "grad_norm": 0.047042325139045715,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 4.121333122253418,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418848167539267,
+ "grad_norm": 0.047646693885326385,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 4.114943981170654,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43280977312391,
+ "grad_norm": 0.04676460474729538,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 4.139252185821533,
+ "step": 1399
+ },
+ {
+ "epoch": 19.44677137870855,
+ "grad_norm": 0.04905956983566284,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 4.135970115661621,
+ "step": 1400
+ },
+ {
+ "epoch": 19.460732984293195,
+ "grad_norm": 0.050025034695863724,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 4.15662956237793,
+ "step": 1401
+ },
+ {
+ "epoch": 19.474694589877835,
+ "grad_norm": 0.04740425944328308,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 4.140295505523682,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48865619546248,
+ "grad_norm": 0.046904709190130234,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 4.14083194732666,
+ "step": 1403
+ },
+ {
+ "epoch": 19.50261780104712,
+ "grad_norm": 0.048393651843070984,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 4.161823272705078,
+ "step": 1404
+ },
+ {
+ "epoch": 19.516579406631763,
+ "grad_norm": 0.04766222462058067,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 4.151817798614502,
+ "step": 1405
+ },
+ {
+ "epoch": 19.530541012216403,
+ "grad_norm": 0.04797709733247757,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 4.173160552978516,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544502617801047,
+ "grad_norm": 0.05012711137533188,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 4.161989688873291,
+ "step": 1407
+ },
+ {
+ "epoch": 19.55846422338569,
+ "grad_norm": 0.04726000875234604,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 4.136838436126709,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57242582897033,
+ "grad_norm": 0.046251531690359116,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 4.1526994705200195,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586387434554975,
+ "grad_norm": 0.047958049923181534,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 4.13817024230957,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600349040139616,
+ "grad_norm": 0.046333350241184235,
+ "learning_rate": 0.000417272427439646,
+ "loss": 4.160187721252441,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61431064572426,
+ "grad_norm": 0.0488678440451622,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 4.166190147399902,
+ "step": 1412
+ },
+ {
+ "epoch": 19.6282722513089,
+ "grad_norm": 0.04784323647618294,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 4.164414405822754,
+ "step": 1413
+ },
+ {
+ "epoch": 19.642233856893544,
+ "grad_norm": 0.05064673349261284,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 4.149814128875732,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656195462478184,
+ "grad_norm": 0.05072085186839104,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 4.167759895324707,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670157068062828,
+ "grad_norm": 0.0499395951628685,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 4.162792205810547,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68411867364747,
+ "grad_norm": 0.046766918152570724,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 4.1416521072387695,
+ "step": 1417
+ },
+ {
+ "epoch": 19.698080279232112,
+ "grad_norm": 0.04523273929953575,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 4.161751747131348,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712041884816752,
+ "grad_norm": 0.046106480062007904,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 4.166394233703613,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726003490401396,
+ "grad_norm": 0.04666793718934059,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 4.124961853027344,
+ "step": 1420
+ },
+ {
+ "epoch": 19.739965095986037,
+ "grad_norm": 0.047407012432813644,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 4.1794328689575195,
+ "step": 1421
+ },
+ {
+ "epoch": 19.75392670157068,
+ "grad_norm": 0.04690873995423317,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 4.160114765167236,
+ "step": 1422
+ },
+ {
+ "epoch": 19.767888307155324,
+ "grad_norm": 0.04589436575770378,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 4.137969970703125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.781849912739965,
+ "grad_norm": 0.04641961678862572,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 4.161533355712891,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79581151832461,
+ "grad_norm": 0.045799050480127335,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 4.187039375305176,
+ "step": 1425
+ },
+ {
+ "epoch": 19.80977312390925,
+ "grad_norm": 0.04881894588470459,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 4.158107757568359,
+ "step": 1426
+ },
+ {
+ "epoch": 19.823734729493893,
+ "grad_norm": 0.04506978392601013,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 4.143360137939453,
+ "step": 1427
+ },
+ {
+ "epoch": 19.837696335078533,
+ "grad_norm": 0.04504861310124397,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 4.174564361572266,
+ "step": 1428
+ },
+ {
+ "epoch": 19.851657940663177,
+ "grad_norm": 0.04639774188399315,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 4.162899971008301,
+ "step": 1429
+ },
+ {
+ "epoch": 19.865619546247817,
+ "grad_norm": 0.047444432973861694,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 4.143390655517578,
+ "step": 1430
+ },
+ {
+ "epoch": 19.87958115183246,
+ "grad_norm": 0.04461613669991493,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 4.171284198760986,
+ "step": 1431
+ },
+ {
+ "epoch": 19.8935427574171,
+ "grad_norm": 0.046383798122406006,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 4.14809513092041,
+ "step": 1432
+ },
+ {
+ "epoch": 19.907504363001745,
+ "grad_norm": 0.04642366245388985,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 4.178903579711914,
+ "step": 1433
+ },
+ {
+ "epoch": 19.921465968586386,
+ "grad_norm": 0.0451771542429924,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 4.171825408935547,
+ "step": 1434
+ },
+ {
+ "epoch": 19.93542757417103,
+ "grad_norm": 0.04767893627285957,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 4.169270038604736,
+ "step": 1435
+ },
+ {
+ "epoch": 19.949389179755673,
+ "grad_norm": 0.04673412814736366,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 4.175506591796875,
+ "step": 1436
+ },
+ {
+ "epoch": 19.963350785340314,
+ "grad_norm": 0.04465145245194435,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 4.149707794189453,
+ "step": 1437
+ },
+ {
+ "epoch": 19.977312390924958,
+ "grad_norm": 0.04734131321310997,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 4.166044235229492,
+ "step": 1438
+ },
+ {
+ "epoch": 19.991273996509598,
+ "grad_norm": 0.04706178605556488,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 4.152081489562988,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.031556785106658936,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 2.60109543800354,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013961605584644,
+ "grad_norm": 0.05152537673711777,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 4.084830284118652,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027923211169284,
+ "grad_norm": 0.05918274074792862,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 4.045302391052246,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041884816753928,
+ "grad_norm": 0.05093073844909668,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 4.070842742919922,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05584642233857,
+ "grad_norm": 0.05095985531806946,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 4.083273410797119,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069808027923212,
+ "grad_norm": 0.05502687394618988,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 4.072626113891602,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083769633507853,
+ "grad_norm": 0.05037364363670349,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 4.075478553771973,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097731239092496,
+ "grad_norm": 0.04983789846301079,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 4.066207408905029,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111692844677137,
+ "grad_norm": 0.049478255212306976,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 4.069969654083252,
+ "step": 1448
+ },
+ {
+ "epoch": 20.12565445026178,
+ "grad_norm": 0.051651593297719955,
+ "learning_rate": 0.000407857329622967,
+ "loss": 4.050713062286377,
+ "step": 1449
+ },
+ {
+ "epoch": 20.13961605584642,
+ "grad_norm": 0.05436555668711662,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 4.069849967956543,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153577661431065,
+ "grad_norm": 0.055190350860357285,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 4.081624984741211,
+ "step": 1451
+ },
+ {
+ "epoch": 20.167539267015705,
+ "grad_norm": 0.053246211260557175,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 4.076076507568359,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18150087260035,
+ "grad_norm": 0.051757775247097015,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 4.049139022827148,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195462478184993,
+ "grad_norm": 0.05064203217625618,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 4.068076133728027,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209424083769633,
+ "grad_norm": 0.05114085227251053,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 4.092901229858398,
+ "step": 1455
+ },
+ {
+ "epoch": 20.223385689354277,
+ "grad_norm": 0.047997843474149704,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 4.077084541320801,
+ "step": 1456
+ },
+ {
+ "epoch": 20.237347294938917,
+ "grad_norm": 0.04880451038479805,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 4.059187889099121,
+ "step": 1457
+ },
+ {
+ "epoch": 20.25130890052356,
+ "grad_norm": 0.05036284029483795,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 4.102115154266357,
+ "step": 1458
+ },
+ {
+ "epoch": 20.2652705061082,
+ "grad_norm": 0.04916979745030403,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 4.098356246948242,
+ "step": 1459
+ },
+ {
+ "epoch": 20.279232111692846,
+ "grad_norm": 0.04713984206318855,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 4.057218074798584,
+ "step": 1460
+ },
+ {
+ "epoch": 20.293193717277486,
+ "grad_norm": 0.04982447996735573,
+ "learning_rate": 0.000404858275823277,
+ "loss": 4.070363998413086,
+ "step": 1461
+ },
+ {
+ "epoch": 20.30715532286213,
+ "grad_norm": 0.0475616492331028,
+ "learning_rate": 0.000404607816444578,
+ "loss": 4.074314117431641,
+ "step": 1462
+ },
+ {
+ "epoch": 20.32111692844677,
+ "grad_norm": 0.046537093818187714,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 4.079501152038574,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335078534031414,
+ "grad_norm": 0.04889644682407379,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 4.0731425285339355,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349040139616054,
+ "grad_norm": 0.05095166340470314,
+ "learning_rate": 0.000403855947934478,
+ "loss": 4.08765983581543,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363001745200698,
+ "grad_norm": 0.049677830189466476,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 4.128052711486816,
+ "step": 1466
+ },
+ {
+ "epoch": 20.376963350785342,
+ "grad_norm": 0.05221942439675331,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 4.094407081604004,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390924956369982,
+ "grad_norm": 0.052042894065380096,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 4.088282585144043,
+ "step": 1468
+ },
+ {
+ "epoch": 20.404886561954626,
+ "grad_norm": 0.04822041094303131,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 4.1070990562438965,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418848167539267,
+ "grad_norm": 0.047207772731781006,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 4.102565765380859,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43280977312391,
+ "grad_norm": 0.049633774906396866,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 4.097216606140137,
+ "step": 1471
+ },
+ {
+ "epoch": 20.44677137870855,
+ "grad_norm": 0.05135185644030571,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 4.095193862915039,
+ "step": 1472
+ },
+ {
+ "epoch": 20.460732984293195,
+ "grad_norm": 0.05215826258063316,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 4.085207939147949,
+ "step": 1473
+ },
+ {
+ "epoch": 20.474694589877835,
+ "grad_norm": 0.050653353333473206,
+ "learning_rate": 0.000401595976318565,
+ "loss": 4.085142612457275,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48865619546248,
+ "grad_norm": 0.04795852303504944,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 4.1158270835876465,
+ "step": 1475
+ },
+ {
+ "epoch": 20.50261780104712,
+ "grad_norm": 0.047416940331459045,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 4.096989154815674,
+ "step": 1476
+ },
+ {
+ "epoch": 20.516579406631763,
+ "grad_norm": 0.049753669649362564,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 4.118251800537109,
+ "step": 1477
+ },
+ {
+ "epoch": 20.530541012216403,
+ "grad_norm": 0.04889055714011192,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 4.081287384033203,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544502617801047,
+ "grad_norm": 0.04857790097594261,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 4.084896564483643,
+ "step": 1479
+ },
+ {
+ "epoch": 20.55846422338569,
+ "grad_norm": 0.047250498086214066,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 4.113656997680664,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57242582897033,
+ "grad_norm": 0.048608481884002686,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 4.113809108734131,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586387434554975,
+ "grad_norm": 0.048189081251621246,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 4.113949775695801,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600349040139616,
+ "grad_norm": 0.049929242581129074,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 4.104221820831299,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61431064572426,
+ "grad_norm": 0.048482850193977356,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 4.091039180755615,
+ "step": 1484
+ },
+ {
+ "epoch": 20.6282722513089,
+ "grad_norm": 0.04810645803809166,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 4.130088806152344,
+ "step": 1485
+ },
+ {
+ "epoch": 20.642233856893544,
+ "grad_norm": 0.04745674133300781,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 4.106161117553711,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656195462478184,
+ "grad_norm": 0.048513609915971756,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 4.119222164154053,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670157068062828,
+ "grad_norm": 0.04506325349211693,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 4.115543842315674,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68411867364747,
+ "grad_norm": 0.04726741835474968,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 4.084345817565918,
+ "step": 1489
+ },
+ {
+ "epoch": 20.698080279232112,
+ "grad_norm": 0.04873545467853546,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 4.116580009460449,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712041884816752,
+ "grad_norm": 0.051252465695142746,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 4.129733085632324,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726003490401396,
+ "grad_norm": 0.0489567406475544,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 4.103477954864502,
+ "step": 1492
+ },
+ {
+ "epoch": 20.739965095986037,
+ "grad_norm": 0.04539056867361069,
+ "learning_rate": 0.000396803974909638,
+ "loss": 4.1327667236328125,
+ "step": 1493
+ },
+ {
+ "epoch": 20.75392670157068,
+ "grad_norm": 0.04954095557332039,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 4.084133625030518,
+ "step": 1494
+ },
+ {
+ "epoch": 20.767888307155324,
+ "grad_norm": 0.05041591823101044,
+ "learning_rate": 0.000396297942393265,
+ "loss": 4.12288761138916,
+ "step": 1495
+ },
+ {
+ "epoch": 20.781849912739965,
+ "grad_norm": 0.046549517661333084,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 4.1259918212890625,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79581151832461,
+ "grad_norm": 0.04704999923706055,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 4.102563858032227,
+ "step": 1497
+ },
+ {
+ "epoch": 20.80977312390925,
+ "grad_norm": 0.05025320500135422,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 4.135492324829102,
+ "step": 1498
+ },
+ {
+ "epoch": 20.823734729493893,
+ "grad_norm": 0.04559441655874252,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 4.118548393249512,
+ "step": 1499
+ },
+ {
+ "epoch": 20.837696335078533,
+ "grad_norm": 0.04913092032074928,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 4.129302024841309,
+ "step": 1500
+ },
+ {
+ "epoch": 20.851657940663177,
+ "grad_norm": 0.04846089705824852,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 4.166059970855713,
+ "step": 1501
+ },
+ {
+ "epoch": 20.865619546247817,
+ "grad_norm": 0.04710553213953972,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 4.104712009429932,
+ "step": 1502
+ },
+ {
+ "epoch": 20.87958115183246,
+ "grad_norm": 0.04896444454789162,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 4.140282154083252,
+ "step": 1503
+ },
+ {
+ "epoch": 20.8935427574171,
+ "grad_norm": 0.0469740591943264,
+ "learning_rate": 0.000394017098798633,
+ "loss": 4.125306129455566,
+ "step": 1504
+ },
+ {
+ "epoch": 20.907504363001745,
+ "grad_norm": 0.04638523980975151,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 4.133161544799805,
+ "step": 1505
+ },
+ {
+ "epoch": 20.921465968586386,
+ "grad_norm": 0.04814442619681358,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 4.093092918395996,
+ "step": 1506
+ },
+ {
+ "epoch": 20.93542757417103,
+ "grad_norm": 0.049911532551050186,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 4.1001715660095215,
+ "step": 1507
+ },
+ {
+ "epoch": 20.949389179755673,
+ "grad_norm": 0.0446670837700367,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 4.140350341796875,
+ "step": 1508
+ },
+ {
+ "epoch": 20.963350785340314,
+ "grad_norm": 0.04789397493004799,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 4.163155555725098,
+ "step": 1509
+ },
+ {
+ "epoch": 20.977312390924958,
+ "grad_norm": 0.050148315727710724,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 4.118380069732666,
+ "step": 1510
+ },
+ {
+ "epoch": 20.991273996509598,
+ "grad_norm": 0.047446198761463165,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 4.116507530212402,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.03415544703602791,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 2.5738282203674316,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.5933796167373657,
+ "eval_runtime": 60.4011,
+ "eval_samples_per_second": 40.43,
+ "eval_steps_per_second": 0.646,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013961605584644,
+ "grad_norm": 0.052276045083999634,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 4.0301923751831055,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027923211169284,
+ "grad_norm": 0.06514342874288559,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 4.031280517578125,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041884816753928,
+ "grad_norm": 0.05948134884238243,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 4.060680389404297,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05584642233857,
+ "grad_norm": 0.05251640826463699,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 4.0218400955200195,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069808027923212,
+ "grad_norm": 0.05426032841205597,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 4.026187896728516,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083769633507853,
+ "grad_norm": 0.052583567798137665,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 4.016767501831055,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097731239092496,
+ "grad_norm": 0.04966135695576668,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 4.043508529663086,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111692844677137,
+ "grad_norm": 0.053838398307561874,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 4.054832458496094,
+ "step": 1520
+ },
+ {
+ "epoch": 21.12565445026178,
+ "grad_norm": 0.05304184556007385,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 4.03280782699585,
+ "step": 1521
+ },
+ {
+ "epoch": 21.13961605584642,
+ "grad_norm": 0.05252664163708687,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 4.039841175079346,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153577661431065,
+ "grad_norm": 0.054593686014413834,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 4.048308372497559,
+ "step": 1523
+ },
+ {
+ "epoch": 21.167539267015705,
+ "grad_norm": 0.050722211599349976,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 4.051860332489014,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18150087260035,
+ "grad_norm": 0.05248095095157623,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 4.056846618652344,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195462478184993,
+ "grad_norm": 0.05166107788681984,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 4.045178413391113,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209424083769633,
+ "grad_norm": 0.04987792298197746,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 4.014827728271484,
+ "step": 1527
+ },
+ {
+ "epoch": 21.223385689354277,
+ "grad_norm": 0.05316167324781418,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 4.033143997192383,
+ "step": 1528
+ },
+ {
+ "epoch": 21.237347294938917,
+ "grad_norm": 0.05228257551789284,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 4.033842086791992,
+ "step": 1529
+ },
+ {
+ "epoch": 21.25130890052356,
+ "grad_norm": 0.049349382519721985,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 4.025982856750488,
+ "step": 1530
+ },
+ {
+ "epoch": 21.2652705061082,
+ "grad_norm": 0.055104926228523254,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 4.059151649475098,
+ "step": 1531
+ },
+ {
+ "epoch": 21.279232111692846,
+ "grad_norm": 0.052629340440034866,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 4.032320976257324,
+ "step": 1532
+ },
+ {
+ "epoch": 21.293193717277486,
+ "grad_norm": 0.05396415665745735,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 4.0425567626953125,
+ "step": 1533
+ },
+ {
+ "epoch": 21.30715532286213,
+ "grad_norm": 0.054417286068201065,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 4.03759241104126,
+ "step": 1534
+ },
+ {
+ "epoch": 21.32111692844677,
+ "grad_norm": 0.061248697340488434,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 4.0441083908081055,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335078534031414,
+ "grad_norm": 0.05934970825910568,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 4.050718307495117,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349040139616054,
+ "grad_norm": 0.052805181592702866,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 4.069460868835449,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363001745200698,
+ "grad_norm": 0.052820611745119095,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 4.063657283782959,
+ "step": 1538
+ },
+ {
+ "epoch": 21.376963350785342,
+ "grad_norm": 0.05043052136898041,
+ "learning_rate": 0.00038509205478744,
+ "loss": 4.029552459716797,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390924956369982,
+ "grad_norm": 0.04828709363937378,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 4.05656623840332,
+ "step": 1540
+ },
+ {
+ "epoch": 21.404886561954626,
+ "grad_norm": 0.049760788679122925,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 4.066338539123535,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418848167539267,
+ "grad_norm": 0.05037159472703934,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 4.06976842880249,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43280977312391,
+ "grad_norm": 0.05276317894458771,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 4.062975883483887,
+ "step": 1543
+ },
+ {
+ "epoch": 21.44677137870855,
+ "grad_norm": 0.05571872368454933,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 4.065622329711914,
+ "step": 1544
+ },
+ {
+ "epoch": 21.460732984293195,
+ "grad_norm": 0.05438469722867012,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 4.074963569641113,
+ "step": 1545
+ },
+ {
+ "epoch": 21.474694589877835,
+ "grad_norm": 0.053415946662425995,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 4.080055236816406,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48865619546248,
+ "grad_norm": 0.05216296389698982,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 4.091355800628662,
+ "step": 1547
+ },
+ {
+ "epoch": 21.50261780104712,
+ "grad_norm": 0.051212459802627563,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 4.058528900146484,
+ "step": 1548
+ },
+ {
+ "epoch": 21.516579406631763,
+ "grad_norm": 0.05033959075808525,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 4.064615726470947,
+ "step": 1549
+ },
+ {
+ "epoch": 21.530541012216403,
+ "grad_norm": 0.04665188118815422,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 4.059698104858398,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544502617801047,
+ "grad_norm": 0.050221361219882965,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 4.068614959716797,
+ "step": 1551
+ },
+ {
+ "epoch": 21.55846422338569,
+ "grad_norm": 0.05089551955461502,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 4.0833892822265625,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57242582897033,
+ "grad_norm": 0.051966603845357895,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 4.083354949951172,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586387434554975,
+ "grad_norm": 0.04981927201151848,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 4.083340644836426,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600349040139616,
+ "grad_norm": 0.05141765624284744,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 4.066590309143066,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61431064572426,
+ "grad_norm": 0.048985838890075684,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 4.088237762451172,
+ "step": 1556
+ },
+ {
+ "epoch": 21.6282722513089,
+ "grad_norm": 0.05013054236769676,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 4.081377983093262,
+ "step": 1557
+ },
+ {
+ "epoch": 21.642233856893544,
+ "grad_norm": 0.05170975625514984,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 4.054303169250488,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656195462478184,
+ "grad_norm": 0.048257142305374146,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 4.103672504425049,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670157068062828,
+ "grad_norm": 0.049942679703235626,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 4.029951572418213,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68411867364747,
+ "grad_norm": 0.05083315819501877,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 4.048798084259033,
+ "step": 1561
+ },
+ {
+ "epoch": 21.698080279232112,
+ "grad_norm": 0.04984128847718239,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 4.079655647277832,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712041884816752,
+ "grad_norm": 0.04985307902097702,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 4.071877956390381,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726003490401396,
+ "grad_norm": 0.051137227565050125,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 4.076539993286133,
+ "step": 1564
+ },
+ {
+ "epoch": 21.739965095986037,
+ "grad_norm": 0.04968905821442604,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 4.084498882293701,
+ "step": 1565
+ },
+ {
+ "epoch": 21.75392670157068,
+ "grad_norm": 0.049119509756565094,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 4.066773414611816,
+ "step": 1566
+ },
+ {
+ "epoch": 21.767888307155324,
+ "grad_norm": 0.05107425898313522,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 4.065309047698975,
+ "step": 1567
+ },
+ {
+ "epoch": 21.781849912739965,
+ "grad_norm": 0.049012504518032074,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 4.062110900878906,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79581151832461,
+ "grad_norm": 0.04972011595964432,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 4.0630903244018555,
+ "step": 1569
+ },
+ {
+ "epoch": 21.80977312390925,
+ "grad_norm": 0.04822700098156929,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 4.106127738952637,
+ "step": 1570
+ },
+ {
+ "epoch": 21.823734729493893,
+ "grad_norm": 0.04808267205953598,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 4.085200309753418,
+ "step": 1571
+ },
+ {
+ "epoch": 21.837696335078533,
+ "grad_norm": 0.048644762486219406,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 4.0856170654296875,
+ "step": 1572
+ },
+ {
+ "epoch": 21.851657940663177,
+ "grad_norm": 0.05232283100485802,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 4.126387119293213,
+ "step": 1573
+ },
+ {
+ "epoch": 21.865619546247817,
+ "grad_norm": 0.05095643922686577,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 4.108254432678223,
+ "step": 1574
+ },
+ {
+ "epoch": 21.87958115183246,
+ "grad_norm": 0.045972101390361786,
+ "learning_rate": 0.000375827577564042,
+ "loss": 4.058515548706055,
+ "step": 1575
+ },
+ {
+ "epoch": 21.8935427574171,
+ "grad_norm": 0.04974910616874695,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 4.089197158813477,
+ "step": 1576
+ },
+ {
+ "epoch": 21.907504363001745,
+ "grad_norm": 0.04890159144997597,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 4.1078410148620605,
+ "step": 1577
+ },
+ {
+ "epoch": 21.921465968586386,
+ "grad_norm": 0.04738525673747063,
+ "learning_rate": 0.000375051971546236,
+ "loss": 4.1219282150268555,
+ "step": 1578
+ },
+ {
+ "epoch": 21.93542757417103,
+ "grad_norm": 0.04713096842169762,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 4.10295295715332,
+ "step": 1579
+ },
+ {
+ "epoch": 21.949389179755673,
+ "grad_norm": 0.04834369197487831,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 4.102747917175293,
+ "step": 1580
+ },
+ {
+ "epoch": 21.963350785340314,
+ "grad_norm": 0.047073688358068466,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 4.08024787902832,
+ "step": 1581
+ },
+ {
+ "epoch": 21.977312390924958,
+ "grad_norm": 0.04568164795637131,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 4.078930854797363,
+ "step": 1582
+ },
+ {
+ "epoch": 21.991273996509598,
+ "grad_norm": 0.04822581633925438,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 4.089601516723633,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.03507270663976669,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 2.5586814880371094,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.595272958278656,
+ "eval_runtime": 60.7797,
+ "eval_samples_per_second": 40.178,
+ "eval_steps_per_second": 0.642,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013961605584644,
+ "grad_norm": 0.05232533439993858,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 4.011311054229736,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027923211169284,
+ "grad_norm": 0.058515798300504684,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 4.0049638748168945,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041884816753928,
+ "grad_norm": 0.055082619190216064,
+ "learning_rate": 0.000372722041257003,
+ "loss": 3.9904184341430664,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05584642233857,
+ "grad_norm": 0.05213981494307518,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 4.003454685211182,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069808027923212,
+ "grad_norm": 0.051619648933410645,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 3.9900388717651367,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083769633507853,
+ "grad_norm": 0.055212799459695816,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 3.973494052886963,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097731239092496,
+ "grad_norm": 0.05262092128396034,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 4.00933837890625,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111692844677137,
+ "grad_norm": 0.05012505501508713,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 4.01861047744751,
+ "step": 1592
+ },
+ {
+ "epoch": 22.12565445026178,
+ "grad_norm": 0.049657098948955536,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 4.011143684387207,
+ "step": 1593
+ },
+ {
+ "epoch": 22.13961605584642,
+ "grad_norm": 0.05187768116593361,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 3.9922969341278076,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153577661431065,
+ "grad_norm": 0.05053112655878067,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 4.007021903991699,
+ "step": 1595
+ },
+ {
+ "epoch": 22.167539267015705,
+ "grad_norm": 0.048542529344558716,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 3.9966416358947754,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18150087260035,
+ "grad_norm": 0.050972841680049896,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 3.9803225994110107,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195462478184993,
+ "grad_norm": 0.05001165345311165,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 4.0140509605407715,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209424083769633,
+ "grad_norm": 0.05100536718964577,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 3.9906058311462402,
+ "step": 1599
+ },
+ {
+ "epoch": 22.223385689354277,
+ "grad_norm": 0.052977193146944046,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 4.012383460998535,
+ "step": 1600
+ },
+ {
+ "epoch": 22.237347294938917,
+ "grad_norm": 0.051219623535871506,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 3.982931613922119,
+ "step": 1601
+ },
+ {
+ "epoch": 22.25130890052356,
+ "grad_norm": 0.05275029316544533,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 4.032669544219971,
+ "step": 1602
+ },
+ {
+ "epoch": 22.2652705061082,
+ "grad_norm": 0.04781945049762726,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 3.9933102130889893,
+ "step": 1603
+ },
+ {
+ "epoch": 22.279232111692846,
+ "grad_norm": 0.054087597876787186,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 4.0075201988220215,
+ "step": 1604
+ },
+ {
+ "epoch": 22.293193717277486,
+ "grad_norm": 0.05485915020108223,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 4.004299640655518,
+ "step": 1605
+ },
+ {
+ "epoch": 22.30715532286213,
+ "grad_norm": 0.05148882791399956,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 4.044161796569824,
+ "step": 1606
+ },
+ {
+ "epoch": 22.32111692844677,
+ "grad_norm": 0.051912397146224976,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 3.9777069091796875,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335078534031414,
+ "grad_norm": 0.052630942314863205,
+ "learning_rate": 0.000367268105959126,
+ "loss": 4.029580116271973,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349040139616054,
+ "grad_norm": 0.0512809231877327,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 4.022092342376709,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363001745200698,
+ "grad_norm": 0.05094102770090103,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 4.018102169036865,
+ "step": 1610
+ },
+ {
+ "epoch": 22.376963350785342,
+ "grad_norm": 0.04972759634256363,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 4.027868270874023,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390924956369982,
+ "grad_norm": 0.05072534456849098,
+ "learning_rate": 0.000366226612810024,
+ "loss": 3.995988368988037,
+ "step": 1612
+ },
+ {
+ "epoch": 22.404886561954626,
+ "grad_norm": 0.050623781979084015,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 4.019711494445801,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418848167539267,
+ "grad_norm": 0.05035007372498512,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 4.007514953613281,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43280977312391,
+ "grad_norm": 0.054426416754722595,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 4.0167131423950195,
+ "step": 1615
+ },
+ {
+ "epoch": 22.44677137870855,
+ "grad_norm": 0.05365690588951111,
+ "learning_rate": 0.000365184304613104,
+ "loss": 4.0313920974731445,
+ "step": 1616
+ },
+ {
+ "epoch": 22.460732984293195,
+ "grad_norm": 0.0509161539375782,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 4.022063255310059,
+ "step": 1617
+ },
+ {
+ "epoch": 22.474694589877835,
+ "grad_norm": 0.05309660732746124,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 4.017317771911621,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48865619546248,
+ "grad_norm": 0.0488961897790432,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 4.028288841247559,
+ "step": 1619
+ },
+ {
+ "epoch": 22.50261780104712,
+ "grad_norm": 0.04982095956802368,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 4.007115840911865,
+ "step": 1620
+ },
+ {
+ "epoch": 22.516579406631763,
+ "grad_norm": 0.05136589705944061,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 4.028022766113281,
+ "step": 1621
+ },
+ {
+ "epoch": 22.530541012216403,
+ "grad_norm": 0.04865279048681259,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 4.015928268432617,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544502617801047,
+ "grad_norm": 0.05131891369819641,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 4.041997909545898,
+ "step": 1623
+ },
+ {
+ "epoch": 22.55846422338569,
+ "grad_norm": 0.05063068866729736,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 4.042117118835449,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57242582897033,
+ "grad_norm": 0.049348752945661545,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 4.012609481811523,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586387434554975,
+ "grad_norm": 0.05123874917626381,
+ "learning_rate": 0.000362575056487632,
+ "loss": 4.03358793258667,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600349040139616,
+ "grad_norm": 0.04921206086874008,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 4.039857864379883,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61431064572426,
+ "grad_norm": 0.05011922866106033,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 4.015835762023926,
+ "step": 1628
+ },
+ {
+ "epoch": 22.6282722513089,
+ "grad_norm": 0.049950532615184784,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 4.068299293518066,
+ "step": 1629
+ },
+ {
+ "epoch": 22.642233856893544,
+ "grad_norm": 0.04962172359228134,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 4.03205680847168,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656195462478184,
+ "grad_norm": 0.05059773847460747,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 4.065027236938477,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670157068062828,
+ "grad_norm": 0.04986768960952759,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 4.01540470123291,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68411867364747,
+ "grad_norm": 0.050884976983070374,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 4.060809135437012,
+ "step": 1633
+ },
+ {
+ "epoch": 22.698080279232112,
+ "grad_norm": 0.052478861063718796,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 4.088409423828125,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712041884816752,
+ "grad_norm": 0.05118529871106148,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 4.071645736694336,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726003490401396,
+ "grad_norm": 0.05059075728058815,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 4.028438568115234,
+ "step": 1636
+ },
+ {
+ "epoch": 22.739965095986037,
+ "grad_norm": 0.049222562462091446,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 4.02166223526001,
+ "step": 1637
+ },
+ {
+ "epoch": 22.75392670157068,
+ "grad_norm": 0.049673549830913544,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 4.0511860847473145,
+ "step": 1638
+ },
+ {
+ "epoch": 22.767888307155324,
+ "grad_norm": 0.04954935610294342,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 4.036063194274902,
+ "step": 1639
+ },
+ {
+ "epoch": 22.781849912739965,
+ "grad_norm": 0.05110877379775047,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 4.081101417541504,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79581151832461,
+ "grad_norm": 0.049009695649147034,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 4.042314529418945,
+ "step": 1641
+ },
+ {
+ "epoch": 22.80977312390925,
+ "grad_norm": 0.0496281161904335,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 4.028315544128418,
+ "step": 1642
+ },
+ {
+ "epoch": 22.823734729493893,
+ "grad_norm": 0.04931195452809334,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 4.018457412719727,
+ "step": 1643
+ },
+ {
+ "epoch": 22.837696335078533,
+ "grad_norm": 0.05006880313158035,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 4.0709991455078125,
+ "step": 1644
+ },
+ {
+ "epoch": 22.851657940663177,
+ "grad_norm": 0.04956946149468422,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 4.065585136413574,
+ "step": 1645
+ },
+ {
+ "epoch": 22.865619546247817,
+ "grad_norm": 0.048182398080825806,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 4.0522589683532715,
+ "step": 1646
+ },
+ {
+ "epoch": 22.87958115183246,
+ "grad_norm": 0.04964807257056236,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 4.06989860534668,
+ "step": 1647
+ },
+ {
+ "epoch": 22.8935427574171,
+ "grad_norm": 0.049299128353595734,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 4.082858085632324,
+ "step": 1648
+ },
+ {
+ "epoch": 22.907504363001745,
+ "grad_norm": 0.04893320053815842,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 4.047144889831543,
+ "step": 1649
+ },
+ {
+ "epoch": 22.921465968586386,
+ "grad_norm": 0.04772866889834404,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 4.043060302734375,
+ "step": 1650
+ },
+ {
+ "epoch": 22.93542757417103,
+ "grad_norm": 0.049417730420827866,
+ "learning_rate": 0.000356031397755195,
+ "loss": 4.031186580657959,
+ "step": 1651
+ },
+ {
+ "epoch": 22.949389179755673,
+ "grad_norm": 0.049944665282964706,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 4.074260711669922,
+ "step": 1652
+ },
+ {
+ "epoch": 22.963350785340314,
+ "grad_norm": 0.04856271669268608,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 4.054103851318359,
+ "step": 1653
+ },
+ {
+ "epoch": 22.977312390924958,
+ "grad_norm": 0.0470513179898262,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 4.041568756103516,
+ "step": 1654
+ },
+ {
+ "epoch": 22.991273996509598,
+ "grad_norm": 0.049018729478120804,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 4.066161632537842,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.034226156771183014,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 2.5147223472595215,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.5970026850700378,
+ "eval_runtime": 60.0787,
+ "eval_samples_per_second": 40.647,
+ "eval_steps_per_second": 0.649,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013961605584644,
+ "grad_norm": 0.05091085657477379,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 3.95741868019104,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027923211169284,
+ "grad_norm": 0.05874285474419594,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 3.966327667236328,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041884816753928,
+ "grad_norm": 0.05394266918301582,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.937232732772827,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05584642233857,
+ "grad_norm": 0.05120540037751198,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 3.9637370109558105,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069808027923212,
+ "grad_norm": 0.05310065299272537,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 3.998631477355957,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083769633507853,
+ "grad_norm": 0.05127466097474098,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.949101448059082,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097731239092496,
+ "grad_norm": 0.05201594531536102,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 3.986267566680908,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111692844677137,
+ "grad_norm": 0.051927369087934494,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 3.9604103565216064,
+ "step": 1664
+ },
+ {
+ "epoch": 23.12565445026178,
+ "grad_norm": 0.05125141143798828,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.9339547157287598,
+ "step": 1665
+ },
+ {
+ "epoch": 23.13961605584642,
+ "grad_norm": 0.05241088196635246,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 3.971261978149414,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153577661431065,
+ "grad_norm": 0.050563592463731766,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.9450082778930664,
+ "step": 1667
+ },
+ {
+ "epoch": 23.167539267015705,
+ "grad_norm": 0.05204763635993004,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 3.9696707725524902,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18150087260035,
+ "grad_norm": 0.05518610402941704,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 3.9816057682037354,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195462478184993,
+ "grad_norm": 0.05336211621761322,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.932252883911133,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209424083769633,
+ "grad_norm": 0.052204642444849014,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 3.9818131923675537,
+ "step": 1671
+ },
+ {
+ "epoch": 23.223385689354277,
+ "grad_norm": 0.05299545079469681,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.9488978385925293,
+ "step": 1672
+ },
+ {
+ "epoch": 23.237347294938917,
+ "grad_norm": 0.05143733322620392,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 4.00959587097168,
+ "step": 1673
+ },
+ {
+ "epoch": 23.25130890052356,
+ "grad_norm": 0.05186324194073677,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.9529542922973633,
+ "step": 1674
+ },
+ {
+ "epoch": 23.2652705061082,
+ "grad_norm": 0.05334410071372986,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 3.96693754196167,
+ "step": 1675
+ },
+ {
+ "epoch": 23.279232111692846,
+ "grad_norm": 0.05345796048641205,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 3.9674174785614014,
+ "step": 1676
+ },
+ {
+ "epoch": 23.293193717277486,
+ "grad_norm": 0.050988901406526566,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 3.9698638916015625,
+ "step": 1677
+ },
+ {
+ "epoch": 23.30715532286213,
+ "grad_norm": 0.056467678397893906,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 3.9645955562591553,
+ "step": 1678
+ },
+ {
+ "epoch": 23.32111692844677,
+ "grad_norm": 0.05554647371172905,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 4.0036091804504395,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335078534031414,
+ "grad_norm": 0.05607818067073822,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 3.973773717880249,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349040139616054,
+ "grad_norm": 0.05734477937221527,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 3.987692356109619,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363001745200698,
+ "grad_norm": 0.052175361663103104,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 3.9761803150177,
+ "step": 1682
+ },
+ {
+ "epoch": 23.376963350785342,
+ "grad_norm": 0.053368814289569855,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 3.965020179748535,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390924956369982,
+ "grad_norm": 0.05505364388227463,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 3.987494468688965,
+ "step": 1684
+ },
+ {
+ "epoch": 23.404886561954626,
+ "grad_norm": 0.0524514764547348,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 3.9624319076538086,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418848167539267,
+ "grad_norm": 0.05681568384170532,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 4.008373737335205,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43280977312391,
+ "grad_norm": 0.05344909429550171,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 3.9840354919433594,
+ "step": 1687
+ },
+ {
+ "epoch": 23.44677137870855,
+ "grad_norm": 0.052657466381788254,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 3.992262363433838,
+ "step": 1688
+ },
+ {
+ "epoch": 23.460732984293195,
+ "grad_norm": 0.05526747554540634,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 3.97529935836792,
+ "step": 1689
+ },
+ {
+ "epoch": 23.474694589877835,
+ "grad_norm": 0.05492551624774933,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 4.004708766937256,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48865619546248,
+ "grad_norm": 0.053742099553346634,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 3.994016647338867,
+ "step": 1691
+ },
+ {
+ "epoch": 23.50261780104712,
+ "grad_norm": 0.05363722890615463,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 3.9942378997802734,
+ "step": 1692
+ },
+ {
+ "epoch": 23.516579406631763,
+ "grad_norm": 0.05167972669005394,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 3.998253345489502,
+ "step": 1693
+ },
+ {
+ "epoch": 23.530541012216403,
+ "grad_norm": 0.0518590584397316,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 3.9850645065307617,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544502617801047,
+ "grad_norm": 0.05521136894822121,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 4.02669620513916,
+ "step": 1695
+ },
+ {
+ "epoch": 23.55846422338569,
+ "grad_norm": 0.054986029863357544,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 4.022308826446533,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57242582897033,
+ "grad_norm": 0.05405082181096077,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 4.021487236022949,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586387434554975,
+ "grad_norm": 0.051257841289043427,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 3.9724135398864746,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600349040139616,
+ "grad_norm": 0.05109119415283203,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 3.978914499282837,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61431064572426,
+ "grad_norm": 0.05303420498967171,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 4.027509689331055,
+ "step": 1700
+ },
+ {
+ "epoch": 23.6282722513089,
+ "grad_norm": 0.055543459951877594,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 3.9750852584838867,
+ "step": 1701
+ },
+ {
+ "epoch": 23.642233856893544,
+ "grad_norm": 0.05068540200591087,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 4.029600143432617,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656195462478184,
+ "grad_norm": 0.04991193488240242,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 3.9901509284973145,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670157068062828,
+ "grad_norm": 0.05403713881969452,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 4.017679214477539,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68411867364747,
+ "grad_norm": 0.05203867331147194,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 4.008406639099121,
+ "step": 1705
+ },
+ {
+ "epoch": 23.698080279232112,
+ "grad_norm": 0.05106597766280174,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 4.009983062744141,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712041884816752,
+ "grad_norm": 0.04976251348853111,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 4.007230281829834,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726003490401396,
+ "grad_norm": 0.05199650675058365,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 3.9902141094207764,
+ "step": 1708
+ },
+ {
+ "epoch": 23.739965095986037,
+ "grad_norm": 0.05285118520259857,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 3.9966492652893066,
+ "step": 1709
+ },
+ {
+ "epoch": 23.75392670157068,
+ "grad_norm": 0.051937371492385864,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 3.962397336959839,
+ "step": 1710
+ },
+ {
+ "epoch": 23.767888307155324,
+ "grad_norm": 0.05179062858223915,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 4.027141571044922,
+ "step": 1711
+ },
+ {
+ "epoch": 23.781849912739965,
+ "grad_norm": 0.049667056649923325,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 4.024991512298584,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79581151832461,
+ "grad_norm": 0.05204658955335617,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 4.002647876739502,
+ "step": 1713
+ },
+ {
+ "epoch": 23.80977312390925,
+ "grad_norm": 0.05069981515407562,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 4.035422325134277,
+ "step": 1714
+ },
+ {
+ "epoch": 23.823734729493893,
+ "grad_norm": 0.05044785887002945,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 4.042465686798096,
+ "step": 1715
+ },
+ {
+ "epoch": 23.837696335078533,
+ "grad_norm": 0.053949885070323944,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 4.002717971801758,
+ "step": 1716
+ },
+ {
+ "epoch": 23.851657940663177,
+ "grad_norm": 0.05139258876442909,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 4.026609897613525,
+ "step": 1717
+ },
+ {
+ "epoch": 23.865619546247817,
+ "grad_norm": 0.05316372960805893,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 4.019758224487305,
+ "step": 1718
+ },
+ {
+ "epoch": 23.87958115183246,
+ "grad_norm": 0.050791382789611816,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 3.9859633445739746,
+ "step": 1719
+ },
+ {
+ "epoch": 23.8935427574171,
+ "grad_norm": 0.05124847590923309,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 4.038055419921875,
+ "step": 1720
+ },
+ {
+ "epoch": 23.907504363001745,
+ "grad_norm": 0.0495096817612648,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 4.02764368057251,
+ "step": 1721
+ },
+ {
+ "epoch": 23.921465968586386,
+ "grad_norm": 0.05205567181110382,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 4.003867149353027,
+ "step": 1722
+ },
+ {
+ "epoch": 23.93542757417103,
+ "grad_norm": 0.05233347788453102,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 4.024157524108887,
+ "step": 1723
+ },
+ {
+ "epoch": 23.949389179755673,
+ "grad_norm": 0.05126349255442619,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 4.05349063873291,
+ "step": 1724
+ },
+ {
+ "epoch": 23.963350785340314,
+ "grad_norm": 0.050006967037916183,
+ "learning_rate": 0.000336518346307424,
+ "loss": 4.024704933166504,
+ "step": 1725
+ },
+ {
+ "epoch": 23.977312390924958,
+ "grad_norm": 0.052343279123306274,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 4.008549213409424,
+ "step": 1726
+ },
+ {
+ "epoch": 23.991273996509598,
+ "grad_norm": 0.05126652494072914,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 4.002162933349609,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.03560106083750725,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 2.511049270629883,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.5985628962516785,
+ "eval_runtime": 59.0712,
+ "eval_samples_per_second": 41.34,
+ "eval_steps_per_second": 0.66,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013961605584644,
+ "grad_norm": 0.05205906555056572,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.936516284942627,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027923211169284,
+ "grad_norm": 0.056368257850408554,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.95200777053833,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041884816753928,
+ "grad_norm": 0.05075085908174515,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.940810441970825,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05584642233857,
+ "grad_norm": 0.05119713768362999,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.90130615234375,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069808027923212,
+ "grad_norm": 0.05254847928881645,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.9491095542907715,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083769633507853,
+ "grad_norm": 0.053025662899017334,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.9371917247772217,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097731239092496,
+ "grad_norm": 0.05402665585279465,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 3.9651317596435547,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111692844677137,
+ "grad_norm": 0.0519571453332901,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.9189062118530273,
+ "step": 1736
+ },
+ {
+ "epoch": 24.12565445026178,
+ "grad_norm": 0.05436019226908684,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.948184013366699,
+ "step": 1737
+ },
+ {
+ "epoch": 24.13961605584642,
+ "grad_norm": 0.056572459638118744,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.9076738357543945,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153577661431065,
+ "grad_norm": 0.05553183704614639,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.933659076690674,
+ "step": 1739
+ },
+ {
+ "epoch": 24.167539267015705,
+ "grad_norm": 0.05554374307394028,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.941751480102539,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18150087260035,
+ "grad_norm": 0.05451614409685135,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.928593635559082,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195462478184993,
+ "grad_norm": 0.055458780378103256,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 3.961820602416992,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209424083769633,
+ "grad_norm": 0.054980456829071045,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 3.9736528396606445,
+ "step": 1743
+ },
+ {
+ "epoch": 24.223385689354277,
+ "grad_norm": 0.05538604408502579,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.935093641281128,
+ "step": 1744
+ },
+ {
+ "epoch": 24.237347294938917,
+ "grad_norm": 0.05580678582191467,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.9240689277648926,
+ "step": 1745
+ },
+ {
+ "epoch": 24.25130890052356,
+ "grad_norm": 0.052294690161943436,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.937718152999878,
+ "step": 1746
+ },
+ {
+ "epoch": 24.2652705061082,
+ "grad_norm": 0.052018795162439346,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.942049503326416,
+ "step": 1747
+ },
+ {
+ "epoch": 24.279232111692846,
+ "grad_norm": 0.0550091527402401,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.9215564727783203,
+ "step": 1748
+ },
+ {
+ "epoch": 24.293193717277486,
+ "grad_norm": 0.05447716638445854,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.9154953956604004,
+ "step": 1749
+ },
+ {
+ "epoch": 24.30715532286213,
+ "grad_norm": 0.05300029739737511,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 3.9642622470855713,
+ "step": 1750
+ },
+ {
+ "epoch": 24.32111692844677,
+ "grad_norm": 0.05312013998627663,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.938897132873535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335078534031414,
+ "grad_norm": 0.053737837821245193,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.94972562789917,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349040139616054,
+ "grad_norm": 0.05087469890713692,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.9301164150238037,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363001745200698,
+ "grad_norm": 0.05179893225431442,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.9393205642700195,
+ "step": 1754
+ },
+ {
+ "epoch": 24.376963350785342,
+ "grad_norm": 0.05277438461780548,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 3.960968494415283,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390924956369982,
+ "grad_norm": 0.05090474337339401,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 3.9516637325286865,
+ "step": 1756
+ },
+ {
+ "epoch": 24.404886561954626,
+ "grad_norm": 0.052786506712436676,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.9332518577575684,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418848167539267,
+ "grad_norm": 0.052095942199230194,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.9108662605285645,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43280977312391,
+ "grad_norm": 0.05310450494289398,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 3.9746041297912598,
+ "step": 1759
+ },
+ {
+ "epoch": 24.44677137870855,
+ "grad_norm": 0.05078566074371338,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.93316650390625,
+ "step": 1760
+ },
+ {
+ "epoch": 24.460732984293195,
+ "grad_norm": 0.0519949309527874,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.9489870071411133,
+ "step": 1761
+ },
+ {
+ "epoch": 24.474694589877835,
+ "grad_norm": 0.05239349976181984,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 3.955118417739868,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48865619546248,
+ "grad_norm": 0.05193360522389412,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 3.9707555770874023,
+ "step": 1763
+ },
+ {
+ "epoch": 24.50261780104712,
+ "grad_norm": 0.05397362262010574,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 3.973053455352783,
+ "step": 1764
+ },
+ {
+ "epoch": 24.516579406631763,
+ "grad_norm": 0.05379117652773857,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 3.957289695739746,
+ "step": 1765
+ },
+ {
+ "epoch": 24.530541012216403,
+ "grad_norm": 0.0526256300508976,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 3.9595577716827393,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544502617801047,
+ "grad_norm": 0.05377693101763725,
+ "learning_rate": 0.000325372061241796,
+ "loss": 3.97849178314209,
+ "step": 1767
+ },
+ {
+ "epoch": 24.55846422338569,
+ "grad_norm": 0.05182809755206108,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 3.9714043140411377,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57242582897033,
+ "grad_norm": 0.05364793911576271,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 3.9582200050354004,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586387434554975,
+ "grad_norm": 0.051558300852775574,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 3.9747936725616455,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600349040139616,
+ "grad_norm": 0.053487639874219894,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 3.9556500911712646,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61431064572426,
+ "grad_norm": 0.05146970972418785,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.94858455657959,
+ "step": 1772
+ },
+ {
+ "epoch": 24.6282722513089,
+ "grad_norm": 0.05099836364388466,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 3.9716954231262207,
+ "step": 1773
+ },
+ {
+ "epoch": 24.642233856893544,
+ "grad_norm": 0.054634734988212585,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 3.974881887435913,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656195462478184,
+ "grad_norm": 0.054601747542619705,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 3.9746241569519043,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670157068062828,
+ "grad_norm": 0.052188221365213394,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 3.957106113433838,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68411867364747,
+ "grad_norm": 0.055673446506261826,
+ "learning_rate": 0.000322712903929477,
+ "loss": 3.9803967475891113,
+ "step": 1777
+ },
+ {
+ "epoch": 24.698080279232112,
+ "grad_norm": 0.05397328734397888,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 3.967238426208496,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712041884816752,
+ "grad_norm": 0.05384642630815506,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 3.949392795562744,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726003490401396,
+ "grad_norm": 0.05452995374798775,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.9432482719421387,
+ "step": 1780
+ },
+ {
+ "epoch": 24.739965095986037,
+ "grad_norm": 0.052430473268032074,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 3.9944653511047363,
+ "step": 1781
+ },
+ {
+ "epoch": 24.75392670157068,
+ "grad_norm": 0.05514240264892578,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 3.97713565826416,
+ "step": 1782
+ },
+ {
+ "epoch": 24.767888307155324,
+ "grad_norm": 0.05189919099211693,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 3.983150005340576,
+ "step": 1783
+ },
+ {
+ "epoch": 24.781849912739965,
+ "grad_norm": 0.05590776726603508,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 3.9548301696777344,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79581151832461,
+ "grad_norm": 0.05564190074801445,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 3.9725170135498047,
+ "step": 1785
+ },
+ {
+ "epoch": 24.80977312390925,
+ "grad_norm": 0.05121326446533203,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 3.9722347259521484,
+ "step": 1786
+ },
+ {
+ "epoch": 24.823734729493893,
+ "grad_norm": 0.05444202199578285,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 3.9634029865264893,
+ "step": 1787
+ },
+ {
+ "epoch": 24.837696335078533,
+ "grad_norm": 0.05242442339658737,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 3.973270893096924,
+ "step": 1788
+ },
+ {
+ "epoch": 24.851657940663177,
+ "grad_norm": 0.05497463420033455,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 3.9851887226104736,
+ "step": 1789
+ },
+ {
+ "epoch": 24.865619546247817,
+ "grad_norm": 0.05482025071978569,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 4.001110076904297,
+ "step": 1790
+ },
+ {
+ "epoch": 24.87958115183246,
+ "grad_norm": 0.054101862013339996,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 3.9782891273498535,
+ "step": 1791
+ },
+ {
+ "epoch": 24.8935427574171,
+ "grad_norm": 0.05538944527506828,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 3.973541498184204,
+ "step": 1792
+ },
+ {
+ "epoch": 24.907504363001745,
+ "grad_norm": 0.0502893440425396,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 3.9875426292419434,
+ "step": 1793
+ },
+ {
+ "epoch": 24.921465968586386,
+ "grad_norm": 0.05576152354478836,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.947035551071167,
+ "step": 1794
+ },
+ {
+ "epoch": 24.93542757417103,
+ "grad_norm": 0.05356278270483017,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 3.9683666229248047,
+ "step": 1795
+ },
+ {
+ "epoch": 24.949389179755673,
+ "grad_norm": 0.05377693101763725,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 3.9973440170288086,
+ "step": 1796
+ },
+ {
+ "epoch": 24.963350785340314,
+ "grad_norm": 0.05334382876753807,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 3.9843404293060303,
+ "step": 1797
+ },
+ {
+ "epoch": 24.977312390924958,
+ "grad_norm": 0.05514024198055267,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 3.9920568466186523,
+ "step": 1798
+ },
+ {
+ "epoch": 24.991273996509598,
+ "grad_norm": 0.05631054937839508,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 4.005859375,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.03653077408671379,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 2.503864049911499,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 0.5995591878890991,
+ "eval_runtime": 59.5213,
+ "eval_samples_per_second": 41.027,
+ "eval_steps_per_second": 0.655,
+ "step": 1800
+ },
+ {
+ "epoch": 25.013961605584644,
+ "grad_norm": 0.0576021745800972,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 3.894834518432617,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027923211169284,
+ "grad_norm": 0.05842140316963196,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 3.8879270553588867,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041884816753928,
+ "grad_norm": 0.051821205765008926,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 3.9011340141296387,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05584642233857,
+ "grad_norm": 0.053872160613536835,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 3.877183437347412,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069808027923212,
+ "grad_norm": 0.05487574636936188,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 3.9126815795898438,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083769633507853,
+ "grad_norm": 0.054469428956508636,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 3.894768238067627,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097731239092496,
+ "grad_norm": 0.05124446377158165,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 3.916809320449829,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111692844677137,
+ "grad_norm": 0.05193263292312622,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 3.917140483856201,
+ "step": 1808
+ },
+ {
+ "epoch": 25.12565445026178,
+ "grad_norm": 0.0519125871360302,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 3.903704881668091,
+ "step": 1809
+ },
+ {
+ "epoch": 25.13961605584642,
+ "grad_norm": 0.052205223590135574,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 3.892639636993408,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153577661431065,
+ "grad_norm": 0.052036914974451065,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 3.895559787750244,
+ "step": 1811
+ },
+ {
+ "epoch": 25.167539267015705,
+ "grad_norm": 0.054616156965494156,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 3.8976142406463623,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18150087260035,
+ "grad_norm": 0.05603514611721039,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 3.914311170578003,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195462478184993,
+ "grad_norm": 0.05170242115855217,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 3.9141643047332764,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209424083769633,
+ "grad_norm": 0.05476640164852142,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 3.8952436447143555,
+ "step": 1815
+ },
+ {
+ "epoch": 25.223385689354277,
+ "grad_norm": 0.055114105343818665,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 3.872978687286377,
+ "step": 1816
+ },
+ {
+ "epoch": 25.237347294938917,
+ "grad_norm": 0.051593828946352005,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 3.893911600112915,
+ "step": 1817
+ },
+ {
+ "epoch": 25.25130890052356,
+ "grad_norm": 0.05363611876964569,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 3.9062986373901367,
+ "step": 1818
+ },
+ {
+ "epoch": 25.2652705061082,
+ "grad_norm": 0.054656051099300385,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 3.9145383834838867,
+ "step": 1819
+ },
+ {
+ "epoch": 25.279232111692846,
+ "grad_norm": 0.05251947045326233,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 3.912965774536133,
+ "step": 1820
+ },
+ {
+ "epoch": 25.293193717277486,
+ "grad_norm": 0.05575991049408913,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 3.901510238647461,
+ "step": 1821
+ },
+ {
+ "epoch": 25.30715532286213,
+ "grad_norm": 0.05540676787495613,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 3.8659887313842773,
+ "step": 1822
+ },
+ {
+ "epoch": 25.32111692844677,
+ "grad_norm": 0.0577131025493145,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 3.884077548980713,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335078534031414,
+ "grad_norm": 0.055947501212358475,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 3.9274377822875977,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349040139616054,
+ "grad_norm": 0.05328478664159775,
+ "learning_rate": 0.000309930621834911,
+ "loss": 3.8783111572265625,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363001745200698,
+ "grad_norm": 0.057696059346199036,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 3.903721570968628,
+ "step": 1826
+ },
+ {
+ "epoch": 25.376963350785342,
+ "grad_norm": 0.05717543140053749,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 3.9259400367736816,
+ "step": 1827
+ },
+ {
+ "epoch": 25.390924956369982,
+ "grad_norm": 0.05328287556767464,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 3.9450511932373047,
+ "step": 1828
+ },
+ {
+ "epoch": 25.404886561954626,
+ "grad_norm": 0.05414140224456787,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 3.9231765270233154,
+ "step": 1829
+ },
+ {
+ "epoch": 25.418848167539267,
+ "grad_norm": 0.05535200983285904,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 3.8928630352020264,
+ "step": 1830
+ },
+ {
+ "epoch": 25.43280977312391,
+ "grad_norm": 0.05468215420842171,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 3.933401584625244,
+ "step": 1831
+ },
+ {
+ "epoch": 25.44677137870855,
+ "grad_norm": 0.056624677032232285,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 3.922624111175537,
+ "step": 1832
+ },
+ {
+ "epoch": 25.460732984293195,
+ "grad_norm": 0.0545354001224041,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 3.916515827178955,
+ "step": 1833
+ },
+ {
+ "epoch": 25.474694589877835,
+ "grad_norm": 0.053979165852069855,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 3.891812324523926,
+ "step": 1834
+ },
+ {
+ "epoch": 25.48865619546248,
+ "grad_norm": 0.05619875714182854,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 3.959524631500244,
+ "step": 1835
+ },
+ {
+ "epoch": 25.50261780104712,
+ "grad_norm": 0.053561773151159286,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 3.9109320640563965,
+ "step": 1836
+ },
+ {
+ "epoch": 25.516579406631763,
+ "grad_norm": 0.05566735938191414,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 3.910492420196533,
+ "step": 1837
+ },
+ {
+ "epoch": 25.530541012216403,
+ "grad_norm": 0.05757030472159386,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 3.9716763496398926,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544502617801047,
+ "grad_norm": 0.05121897906064987,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 3.9114646911621094,
+ "step": 1839
+ },
+ {
+ "epoch": 25.55846422338569,
+ "grad_norm": 0.05664774775505066,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 3.9465560913085938,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57242582897033,
+ "grad_norm": 0.0569215826690197,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 3.908374309539795,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586387434554975,
+ "grad_norm": 0.05345458909869194,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 3.9336414337158203,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600349040139616,
+ "grad_norm": 0.0565933920443058,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 3.9478759765625,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61431064572426,
+ "grad_norm": 0.05887870490550995,
+ "learning_rate": 0.000304866093757771,
+ "loss": 3.9018349647521973,
+ "step": 1844
+ },
+ {
+ "epoch": 25.6282722513089,
+ "grad_norm": 0.05515308678150177,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 3.9446825981140137,
+ "step": 1845
+ },
+ {
+ "epoch": 25.642233856893544,
+ "grad_norm": 0.058602213859558105,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 3.9497885704040527,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656195462478184,
+ "grad_norm": 0.053806282579898834,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 3.9247193336486816,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670157068062828,
+ "grad_norm": 0.05495952442288399,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 3.912296772003174,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68411867364747,
+ "grad_norm": 0.05479473993182182,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 3.9471306800842285,
+ "step": 1849
+ },
+ {
+ "epoch": 25.698080279232112,
+ "grad_norm": 0.053319111466407776,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 3.955942153930664,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712041884816752,
+ "grad_norm": 0.05832696333527565,
+ "learning_rate": 0.000303,
+ "loss": 3.959169626235962,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726003490401396,
+ "grad_norm": 0.057019371539354324,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 3.9682188034057617,
+ "step": 1852
+ },
+ {
+ "epoch": 25.739965095986037,
+ "grad_norm": 0.05099530518054962,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 3.921377658843994,
+ "step": 1853
+ },
+ {
+ "epoch": 25.75392670157068,
+ "grad_norm": 0.05466439574956894,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 3.946341037750244,
+ "step": 1854
+ },
+ {
+ "epoch": 25.767888307155324,
+ "grad_norm": 0.05742308124899864,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 3.948031425476074,
+ "step": 1855
+ },
+ {
+ "epoch": 25.781849912739965,
+ "grad_norm": 0.057581741362810135,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 3.91162109375,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79581151832461,
+ "grad_norm": 0.05520833283662796,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 3.937495231628418,
+ "step": 1857
+ },
+ {
+ "epoch": 25.80977312390925,
+ "grad_norm": 0.05763327330350876,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 3.9565534591674805,
+ "step": 1858
+ },
+ {
+ "epoch": 25.823734729493893,
+ "grad_norm": 0.0544707253575325,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 3.965190887451172,
+ "step": 1859
+ },
+ {
+ "epoch": 25.837696335078533,
+ "grad_norm": 0.05445104092359543,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 3.938063859939575,
+ "step": 1860
+ },
+ {
+ "epoch": 25.851657940663177,
+ "grad_norm": 0.05505113676190376,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 3.928392171859741,
+ "step": 1861
+ },
+ {
+ "epoch": 25.865619546247817,
+ "grad_norm": 0.05574212223291397,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 3.959397077560425,
+ "step": 1862
+ },
+ {
+ "epoch": 25.87958115183246,
+ "grad_norm": 0.053950466215610504,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 3.9532437324523926,
+ "step": 1863
+ },
+ {
+ "epoch": 25.8935427574171,
+ "grad_norm": 0.05495546758174896,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 3.946746826171875,
+ "step": 1864
+ },
+ {
+ "epoch": 25.907504363001745,
+ "grad_norm": 0.05328625068068504,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 3.9494776725769043,
+ "step": 1865
+ },
+ {
+ "epoch": 25.921465968586386,
+ "grad_norm": 0.05593104660511017,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 3.959989547729492,
+ "step": 1866
+ },
+ {
+ "epoch": 25.93542757417103,
+ "grad_norm": 0.05442969873547554,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 3.952371597290039,
+ "step": 1867
+ },
+ {
+ "epoch": 25.949389179755673,
+ "grad_norm": 0.05446415767073631,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 3.9538888931274414,
+ "step": 1868
+ },
+ {
+ "epoch": 25.963350785340314,
+ "grad_norm": 0.054761797189712524,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 3.9386215209960938,
+ "step": 1869
+ },
+ {
+ "epoch": 25.977312390924958,
+ "grad_norm": 0.054201845079660416,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 3.9252963066101074,
+ "step": 1870
+ },
+ {
+ "epoch": 25.991273996509598,
+ "grad_norm": 0.05364517495036125,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 3.9733197689056396,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.038914378732442856,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 2.4261250495910645,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 0.6009622812271118,
+ "eval_runtime": 60.6864,
+ "eval_samples_per_second": 40.24,
+ "eval_steps_per_second": 0.643,
+ "step": 1872
+ },
+ {
+ "epoch": 26.013961605584644,
+ "grad_norm": 0.05345558747649193,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 3.8902673721313477,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027923211169284,
+ "grad_norm": 0.05672457069158554,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 3.8676247596740723,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041884816753928,
+ "grad_norm": 0.051981374621391296,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 3.872048854827881,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05584642233857,
+ "grad_norm": 0.05262405797839165,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 3.8582329750061035,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069808027923212,
+ "grad_norm": 0.054473262280225754,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 3.8352112770080566,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083769633507853,
+ "grad_norm": 0.053167302161455154,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 3.881760835647583,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097731239092496,
+ "grad_norm": 0.053720805794000626,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 3.891775608062744,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111692844677137,
+ "grad_norm": 0.055102936923503876,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 3.8872156143188477,
+ "step": 1880
+ },
+ {
+ "epoch": 26.12565445026178,
+ "grad_norm": 0.05108853057026863,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 3.8625407218933105,
+ "step": 1881
+ },
+ {
+ "epoch": 26.13961605584642,
+ "grad_norm": 0.05526706948876381,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 3.886127233505249,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153577661431065,
+ "grad_norm": 0.05458226427435875,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 3.8722105026245117,
+ "step": 1883
+ },
+ {
+ "epoch": 26.167539267015705,
+ "grad_norm": 0.05342113599181175,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 3.853677749633789,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18150087260035,
+ "grad_norm": 0.053665641695261,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 3.849315881729126,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195462478184993,
+ "grad_norm": 0.054922688752412796,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 3.8599777221679688,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209424083769633,
+ "grad_norm": 0.05412537604570389,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 3.8767571449279785,
+ "step": 1887
+ },
+ {
+ "epoch": 26.223385689354277,
+ "grad_norm": 0.05467838793992996,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 3.8436965942382812,
+ "step": 1888
+ },
+ {
+ "epoch": 26.237347294938917,
+ "grad_norm": 0.054886411875486374,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 3.8796300888061523,
+ "step": 1889
+ },
+ {
+ "epoch": 26.25130890052356,
+ "grad_norm": 0.05437029153108597,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 3.858133316040039,
+ "step": 1890
+ },
+ {
+ "epoch": 26.2652705061082,
+ "grad_norm": 0.05273836478590965,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 3.839015483856201,
+ "step": 1891
+ },
+ {
+ "epoch": 26.279232111692846,
+ "grad_norm": 0.0546451173722744,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 3.8723373413085938,
+ "step": 1892
+ },
+ {
+ "epoch": 26.293193717277486,
+ "grad_norm": 0.05240490287542343,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 3.877549886703491,
+ "step": 1893
+ },
+ {
+ "epoch": 26.30715532286213,
+ "grad_norm": 0.05420667678117752,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 3.8643171787261963,
+ "step": 1894
+ },
+ {
+ "epoch": 26.32111692844677,
+ "grad_norm": 0.053440142422914505,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 3.8681447505950928,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335078534031414,
+ "grad_norm": 0.052396632730960846,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 3.8691887855529785,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349040139616054,
+ "grad_norm": 0.05318925902247429,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 3.897003650665283,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363001745200698,
+ "grad_norm": 0.05328533053398132,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 3.9079437255859375,
+ "step": 1898
+ },
+ {
+ "epoch": 26.376963350785342,
+ "grad_norm": 0.05196928232908249,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 3.8392016887664795,
+ "step": 1899
+ },
+ {
+ "epoch": 26.390924956369982,
+ "grad_norm": 0.053326766937971115,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 3.8930301666259766,
+ "step": 1900
+ },
+ {
+ "epoch": 26.404886561954626,
+ "grad_norm": 0.052926234900951385,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 3.866997718811035,
+ "step": 1901
+ },
+ {
+ "epoch": 26.418848167539267,
+ "grad_norm": 0.05289287120103836,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 3.896690607070923,
+ "step": 1902
+ },
+ {
+ "epoch": 26.43280977312391,
+ "grad_norm": 0.052162300795316696,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 3.8642401695251465,
+ "step": 1903
+ },
+ {
+ "epoch": 26.44677137870855,
+ "grad_norm": 0.05495629832148552,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 3.8982696533203125,
+ "step": 1904
+ },
+ {
+ "epoch": 26.460732984293195,
+ "grad_norm": 0.055102866142988205,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 3.8725972175598145,
+ "step": 1905
+ },
+ {
+ "epoch": 26.474694589877835,
+ "grad_norm": 0.05537770688533783,
+ "learning_rate": 0.000288343693342466,
+ "loss": 3.9050111770629883,
+ "step": 1906
+ },
+ {
+ "epoch": 26.48865619546248,
+ "grad_norm": 0.05490661785006523,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 3.90586256980896,
+ "step": 1907
+ },
+ {
+ "epoch": 26.50261780104712,
+ "grad_norm": 0.054393451660871506,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 3.8841705322265625,
+ "step": 1908
+ },
+ {
+ "epoch": 26.516579406631763,
+ "grad_norm": 0.0553058385848999,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 3.883500576019287,
+ "step": 1909
+ },
+ {
+ "epoch": 26.530541012216403,
+ "grad_norm": 0.053976885974407196,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 3.894948959350586,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544502617801047,
+ "grad_norm": 0.05290692299604416,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 3.8880579471588135,
+ "step": 1911
+ },
+ {
+ "epoch": 26.55846422338569,
+ "grad_norm": 0.05570613220334053,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 3.9035487174987793,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57242582897033,
+ "grad_norm": 0.054136309772729874,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 3.9137563705444336,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586387434554975,
+ "grad_norm": 0.05328850820660591,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 3.87838077545166,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600349040139616,
+ "grad_norm": 0.05169905349612236,
+ "learning_rate": 0.000285947841605349,
+ "loss": 3.882963180541992,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61431064572426,
+ "grad_norm": 0.05304642394185066,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 3.9250073432922363,
+ "step": 1916
+ },
+ {
+ "epoch": 26.6282722513089,
+ "grad_norm": 0.051766350865364075,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 3.8915140628814697,
+ "step": 1917
+ },
+ {
+ "epoch": 26.642233856893544,
+ "grad_norm": 0.054485417902469635,
+ "learning_rate": 0.000285149463934261,
+ "loss": 3.8973851203918457,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656195462478184,
+ "grad_norm": 0.05437815189361572,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 3.9054274559020996,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670157068062828,
+ "grad_norm": 0.05402572825551033,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 3.9060683250427246,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68411867364747,
+ "grad_norm": 0.05332810431718826,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 3.9053144454956055,
+ "step": 1921
+ },
+ {
+ "epoch": 26.698080279232112,
+ "grad_norm": 0.0557575598359108,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 3.9026694297790527,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712041884816752,
+ "grad_norm": 0.05650033429265022,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 3.8795413970947266,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726003490401396,
+ "grad_norm": 0.054456304758787155,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 3.932950496673584,
+ "step": 1924
+ },
+ {
+ "epoch": 26.739965095986037,
+ "grad_norm": 0.05676668882369995,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 3.878704071044922,
+ "step": 1925
+ },
+ {
+ "epoch": 26.75392670157068,
+ "grad_norm": 0.05875959247350693,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 3.920717239379883,
+ "step": 1926
+ },
+ {
+ "epoch": 26.767888307155324,
+ "grad_norm": 0.054490379989147186,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 3.912592887878418,
+ "step": 1927
+ },
+ {
+ "epoch": 26.781849912739965,
+ "grad_norm": 0.055268481373786926,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 3.900418758392334,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79581151832461,
+ "grad_norm": 0.05617086961865425,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 3.899095296859741,
+ "step": 1929
+ },
+ {
+ "epoch": 26.80977312390925,
+ "grad_norm": 0.05310523882508278,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 3.917501449584961,
+ "step": 1930
+ },
+ {
+ "epoch": 26.823734729493893,
+ "grad_norm": 0.05476578325033188,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 3.867847442626953,
+ "step": 1931
+ },
+ {
+ "epoch": 26.837696335078533,
+ "grad_norm": 0.05456431582570076,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 3.8760366439819336,
+ "step": 1932
+ },
+ {
+ "epoch": 26.851657940663177,
+ "grad_norm": 0.055269334465265274,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 3.910797119140625,
+ "step": 1933
+ },
+ {
+ "epoch": 26.865619546247817,
+ "grad_norm": 0.054999858140945435,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 3.946951389312744,
+ "step": 1934
+ },
+ {
+ "epoch": 26.87958115183246,
+ "grad_norm": 0.05520344525575638,
+ "learning_rate": 0.000280627938758204,
+ "loss": 3.9161229133605957,
+ "step": 1935
+ },
+ {
+ "epoch": 26.8935427574171,
+ "grad_norm": 0.05451727285981178,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 3.908015727996826,
+ "step": 1936
+ },
+ {
+ "epoch": 26.907504363001745,
+ "grad_norm": 0.05561735853552818,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 3.944608449935913,
+ "step": 1937
+ },
+ {
+ "epoch": 26.921465968586386,
+ "grad_norm": 0.054743457585573196,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 3.906559467315674,
+ "step": 1938
+ },
+ {
+ "epoch": 26.93542757417103,
+ "grad_norm": 0.05497661978006363,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 3.900981903076172,
+ "step": 1939
+ },
+ {
+ "epoch": 26.949389179755673,
+ "grad_norm": 0.05452290549874306,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 3.910825252532959,
+ "step": 1940
+ },
+ {
+ "epoch": 26.963350785340314,
+ "grad_norm": 0.05409001186490059,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 3.921509265899658,
+ "step": 1941
+ },
+ {
+ "epoch": 26.977312390924958,
+ "grad_norm": 0.05467002093791962,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 3.9039864540100098,
+ "step": 1942
+ },
+ {
+ "epoch": 26.991273996509598,
+ "grad_norm": 0.05442725867033005,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 3.9153316020965576,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.039271868765354156,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 2.451029062271118,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 0.6021912097930908,
+ "eval_runtime": 61.3232,
+ "eval_samples_per_second": 39.822,
+ "eval_steps_per_second": 0.636,
+ "step": 1944
+ },
+ {
+ "epoch": 27.013961605584644,
+ "grad_norm": 0.05251244455575943,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 3.8143270015716553,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027923211169284,
+ "grad_norm": 0.056699808686971664,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 3.8331453800201416,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041884816753928,
+ "grad_norm": 0.055446673184633255,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 3.833841323852539,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05584642233857,
+ "grad_norm": 0.053186994045972824,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 3.82643985748291,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069808027923212,
+ "grad_norm": 0.05483868718147278,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 3.8209316730499268,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083769633507853,
+ "grad_norm": 0.05387664586305618,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 3.8402364253997803,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097731239092496,
+ "grad_norm": 0.05422234162688255,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 3.8125908374786377,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111692844677137,
+ "grad_norm": 0.05522442236542702,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 3.830599784851074,
+ "step": 1952
+ },
+ {
+ "epoch": 27.12565445026178,
+ "grad_norm": 0.05163128301501274,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 3.833247184753418,
+ "step": 1953
+ },
+ {
+ "epoch": 27.13961605584642,
+ "grad_norm": 0.053523529320955276,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 3.812887191772461,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153577661431065,
+ "grad_norm": 0.05362863466143608,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 3.8325610160827637,
+ "step": 1955
+ },
+ {
+ "epoch": 27.167539267015705,
+ "grad_norm": 0.054255858063697815,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 3.807152271270752,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18150087260035,
+ "grad_norm": 0.05259740352630615,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 3.843714475631714,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195462478184993,
+ "grad_norm": 0.05457160621881485,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 3.847249984741211,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209424083769633,
+ "grad_norm": 0.05517619848251343,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 3.864114284515381,
+ "step": 1959
+ },
+ {
+ "epoch": 27.223385689354277,
+ "grad_norm": 0.052946750074625015,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 3.8219656944274902,
+ "step": 1960
+ },
+ {
+ "epoch": 27.237347294938917,
+ "grad_norm": 0.053545452654361725,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 3.8502726554870605,
+ "step": 1961
+ },
+ {
+ "epoch": 27.25130890052356,
+ "grad_norm": 0.054657287895679474,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 3.837273120880127,
+ "step": 1962
+ },
+ {
+ "epoch": 27.2652705061082,
+ "grad_norm": 0.05385086312890053,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 3.837573528289795,
+ "step": 1963
+ },
+ {
+ "epoch": 27.279232111692846,
+ "grad_norm": 0.05512048304080963,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 3.842432737350464,
+ "step": 1964
+ },
+ {
+ "epoch": 27.293193717277486,
+ "grad_norm": 0.05286296457052231,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 3.845123767852783,
+ "step": 1965
+ },
+ {
+ "epoch": 27.30715532286213,
+ "grad_norm": 0.05396866053342819,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 3.836874485015869,
+ "step": 1966
+ },
+ {
+ "epoch": 27.32111692844677,
+ "grad_norm": 0.053341615945100784,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 3.8860960006713867,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335078534031414,
+ "grad_norm": 0.0531761609017849,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 3.828686237335205,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349040139616054,
+ "grad_norm": 0.054875023663043976,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 3.83782696723938,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363001745200698,
+ "grad_norm": 0.057710494846105576,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 3.8841819763183594,
+ "step": 1970
+ },
+ {
+ "epoch": 27.376963350785342,
+ "grad_norm": 0.05443769320845604,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 3.856564521789551,
+ "step": 1971
+ },
+ {
+ "epoch": 27.390924956369982,
+ "grad_norm": 0.05571551248431206,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 3.847541570663452,
+ "step": 1972
+ },
+ {
+ "epoch": 27.404886561954626,
+ "grad_norm": 0.05770037695765495,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 3.843390941619873,
+ "step": 1973
+ },
+ {
+ "epoch": 27.418848167539267,
+ "grad_norm": 0.05414504557847977,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 3.8786306381225586,
+ "step": 1974
+ },
+ {
+ "epoch": 27.43280977312391,
+ "grad_norm": 0.058058444410562515,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 3.8680999279022217,
+ "step": 1975
+ },
+ {
+ "epoch": 27.44677137870855,
+ "grad_norm": 0.056312717497348785,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 3.858879566192627,
+ "step": 1976
+ },
+ {
+ "epoch": 27.460732984293195,
+ "grad_norm": 0.055812690407037735,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 3.870741844177246,
+ "step": 1977
+ },
+ {
+ "epoch": 27.474694589877835,
+ "grad_norm": 0.057005349546670914,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 3.8850440979003906,
+ "step": 1978
+ },
+ {
+ "epoch": 27.48865619546248,
+ "grad_norm": 0.055214181542396545,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 3.8485918045043945,
+ "step": 1979
+ },
+ {
+ "epoch": 27.50261780104712,
+ "grad_norm": 0.054300058633089066,
+ "learning_rate": 0.00026868712586269,
+ "loss": 3.8759729862213135,
+ "step": 1980
+ },
+ {
+ "epoch": 27.516579406631763,
+ "grad_norm": 0.05370737239718437,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 3.8498623371124268,
+ "step": 1981
+ },
+ {
+ "epoch": 27.530541012216403,
+ "grad_norm": 0.05481775104999542,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 3.8408432006835938,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544502617801047,
+ "grad_norm": 0.05360998213291168,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 3.8800129890441895,
+ "step": 1983
+ },
+ {
+ "epoch": 27.55846422338569,
+ "grad_norm": 0.05510537698864937,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 3.8329358100891113,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57242582897033,
+ "grad_norm": 0.057950302958488464,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 3.870562791824341,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586387434554975,
+ "grad_norm": 0.0546487458050251,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 3.85231351852417,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600349040139616,
+ "grad_norm": 0.056359753012657166,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 3.8782615661621094,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61431064572426,
+ "grad_norm": 0.05409953370690346,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 3.8209736347198486,
+ "step": 1988
+ },
+ {
+ "epoch": 27.6282722513089,
+ "grad_norm": 0.05549091845750809,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 3.9041504859924316,
+ "step": 1989
+ },
+ {
+ "epoch": 27.642233856893544,
+ "grad_norm": 0.05437365546822548,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 3.8530654907226562,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656195462478184,
+ "grad_norm": 0.055456943809986115,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 3.862886428833008,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670157068062828,
+ "grad_norm": 0.05614335089921951,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 3.867595672607422,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68411867364747,
+ "grad_norm": 0.05374857038259506,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 3.8731391429901123,
+ "step": 1993
+ },
+ {
+ "epoch": 27.698080279232112,
+ "grad_norm": 0.05544714257121086,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 3.837608575820923,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712041884816752,
+ "grad_norm": 0.05495234578847885,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 3.8587915897369385,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726003490401396,
+ "grad_norm": 0.05354582890868187,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 3.8618052005767822,
+ "step": 1996
+ },
+ {
+ "epoch": 27.739965095986037,
+ "grad_norm": 0.05556262657046318,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 3.8630597591400146,
+ "step": 1997
+ },
+ {
+ "epoch": 27.75392670157068,
+ "grad_norm": 0.05576452240347862,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 3.8570914268493652,
+ "step": 1998
+ },
+ {
+ "epoch": 27.767888307155324,
+ "grad_norm": 0.053394027054309845,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 3.868783950805664,
+ "step": 1999
+ },
+ {
+ "epoch": 27.781849912739965,
+ "grad_norm": 0.05506644397974014,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 3.871555805206299,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79581151832461,
+ "grad_norm": 0.054367583245038986,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 3.8487513065338135,
+ "step": 2001
+ },
+ {
+ "epoch": 27.80977312390925,
+ "grad_norm": 0.05482787266373634,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 3.88199520111084,
+ "step": 2002
+ },
+ {
+ "epoch": 27.823734729493893,
+ "grad_norm": 0.05416601523756981,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 3.8876938819885254,
+ "step": 2003
+ },
+ {
+ "epoch": 27.837696335078533,
+ "grad_norm": 0.05612047761678696,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 3.8772940635681152,
+ "step": 2004
+ },
+ {
+ "epoch": 27.851657940663177,
+ "grad_norm": 0.05662519112229347,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 3.8429012298583984,
+ "step": 2005
+ },
+ {
+ "epoch": 27.865619546247817,
+ "grad_norm": 0.055557794868946075,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 3.8712010383605957,
+ "step": 2006
+ },
+ {
+ "epoch": 27.87958115183246,
+ "grad_norm": 0.054365914314985275,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 3.884854793548584,
+ "step": 2007
+ },
+ {
+ "epoch": 27.8935427574171,
+ "grad_norm": 0.05622334033250809,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 3.853909492492676,
+ "step": 2008
+ },
+ {
+ "epoch": 27.907504363001745,
+ "grad_norm": 0.05513857677578926,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 3.888460159301758,
+ "step": 2009
+ },
+ {
+ "epoch": 27.921465968586386,
+ "grad_norm": 0.05537945404648781,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 3.891343593597412,
+ "step": 2010
+ },
+ {
+ "epoch": 27.93542757417103,
+ "grad_norm": 0.05649508908390999,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 3.8807168006896973,
+ "step": 2011
+ },
+ {
+ "epoch": 27.949389179755673,
+ "grad_norm": 0.054261550307273865,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 3.8827662467956543,
+ "step": 2012
+ },
+ {
+ "epoch": 27.963350785340314,
+ "grad_norm": 0.05615224689245224,
+ "learning_rate": 0.00025996500713765,
+ "loss": 3.8720929622650146,
+ "step": 2013
+ },
+ {
+ "epoch": 27.977312390924958,
+ "grad_norm": 0.05571882426738739,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 3.8978285789489746,
+ "step": 2014
+ },
+ {
+ "epoch": 27.991273996509598,
+ "grad_norm": 0.05477411672472954,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 3.875284194946289,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.03992476686835289,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 2.404466152191162,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 0.6039577126502991,
+ "eval_runtime": 59.8223,
+ "eval_samples_per_second": 40.821,
+ "eval_steps_per_second": 0.652,
+ "step": 2016
+ },
+ {
+ "epoch": 28.013961605584644,
+ "grad_norm": 0.05395985022187233,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 3.78012752532959,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027923211169284,
+ "grad_norm": 0.05445432290434837,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 3.794297218322754,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041884816753928,
+ "grad_norm": 0.05048619955778122,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 3.8059635162353516,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05584642233857,
+ "grad_norm": 0.054649483412504196,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 3.8080105781555176,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069808027923212,
+ "grad_norm": 0.054571982473134995,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 3.8171639442443848,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083769633507853,
+ "grad_norm": 0.05476690083742142,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 3.7937488555908203,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097731239092496,
+ "grad_norm": 0.05586248263716698,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 3.804347515106201,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111692844677137,
+ "grad_norm": 0.05466577783226967,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 3.8114442825317383,
+ "step": 2024
+ },
+ {
+ "epoch": 28.12565445026178,
+ "grad_norm": 0.05342315137386322,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 3.8107991218566895,
+ "step": 2025
+ },
+ {
+ "epoch": 28.13961605584642,
+ "grad_norm": 0.05318617820739746,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 3.8390064239501953,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153577661431065,
+ "grad_norm": 0.05371803045272827,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 3.8021721839904785,
+ "step": 2027
+ },
+ {
+ "epoch": 28.167539267015705,
+ "grad_norm": 0.05319817736744881,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 3.844275712966919,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18150087260035,
+ "grad_norm": 0.05202893167734146,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 3.8008670806884766,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195462478184993,
+ "grad_norm": 0.054026514291763306,
+ "learning_rate": 0.000255486047794226,
+ "loss": 3.81990909576416,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209424083769633,
+ "grad_norm": 0.05195610597729683,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 3.8142457008361816,
+ "step": 2031
+ },
+ {
+ "epoch": 28.223385689354277,
+ "grad_norm": 0.05212428793311119,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 3.8139138221740723,
+ "step": 2032
+ },
+ {
+ "epoch": 28.237347294938917,
+ "grad_norm": 0.05503816530108452,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 3.8072667121887207,
+ "step": 2033
+ },
+ {
+ "epoch": 28.25130890052356,
+ "grad_norm": 0.05305452272295952,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 3.8146510124206543,
+ "step": 2034
+ },
+ {
+ "epoch": 28.2652705061082,
+ "grad_norm": 0.05301898717880249,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 3.7936806678771973,
+ "step": 2035
+ },
+ {
+ "epoch": 28.279232111692846,
+ "grad_norm": 0.0542520172894001,
+ "learning_rate": 0.000253907826333243,
+ "loss": 3.7920427322387695,
+ "step": 2036
+ },
+ {
+ "epoch": 28.293193717277486,
+ "grad_norm": 0.05349896103143692,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 3.819732189178467,
+ "step": 2037
+ },
+ {
+ "epoch": 28.30715532286213,
+ "grad_norm": 0.05463109537959099,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 3.804321765899658,
+ "step": 2038
+ },
+ {
+ "epoch": 28.32111692844677,
+ "grad_norm": 0.05445708706974983,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 3.8128511905670166,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335078534031414,
+ "grad_norm": 0.05655837804079056,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 3.798435688018799,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349040139616054,
+ "grad_norm": 0.05543001368641853,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 3.779137134552002,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363001745200698,
+ "grad_norm": 0.05458657816052437,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 3.833167552947998,
+ "step": 2042
+ },
+ {
+ "epoch": 28.376963350785342,
+ "grad_norm": 0.056820500642061234,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 3.8207151889801025,
+ "step": 2043
+ },
+ {
+ "epoch": 28.390924956369982,
+ "grad_norm": 0.05464114993810654,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 3.788215160369873,
+ "step": 2044
+ },
+ {
+ "epoch": 28.404886561954626,
+ "grad_norm": 0.053662411868572235,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 3.8059287071228027,
+ "step": 2045
+ },
+ {
+ "epoch": 28.418848167539267,
+ "grad_norm": 0.055799756199121475,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 3.8050360679626465,
+ "step": 2046
+ },
+ {
+ "epoch": 28.43280977312391,
+ "grad_norm": 0.05412062630057335,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 3.8110790252685547,
+ "step": 2047
+ },
+ {
+ "epoch": 28.44677137870855,
+ "grad_norm": 0.05672532692551613,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 3.841888904571533,
+ "step": 2048
+ },
+ {
+ "epoch": 28.460732984293195,
+ "grad_norm": 0.05622367933392525,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 3.8200669288635254,
+ "step": 2049
+ },
+ {
+ "epoch": 28.474694589877835,
+ "grad_norm": 0.0544997975230217,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 3.8031818866729736,
+ "step": 2050
+ },
+ {
+ "epoch": 28.48865619546248,
+ "grad_norm": 0.058867089450359344,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 3.856843948364258,
+ "step": 2051
+ },
+ {
+ "epoch": 28.50261780104712,
+ "grad_norm": 0.055324845016002655,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 3.807135820388794,
+ "step": 2052
+ },
+ {
+ "epoch": 28.516579406631763,
+ "grad_norm": 0.055414266884326935,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 3.8380215167999268,
+ "step": 2053
+ },
+ {
+ "epoch": 28.530541012216403,
+ "grad_norm": 0.05787660554051399,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 3.831721782684326,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544502617801047,
+ "grad_norm": 0.05551844462752342,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 3.8572306632995605,
+ "step": 2055
+ },
+ {
+ "epoch": 28.55846422338569,
+ "grad_norm": 0.05452537164092064,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 3.840963840484619,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57242582897033,
+ "grad_norm": 0.05530945956707001,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 3.840442419052124,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586387434554975,
+ "grad_norm": 0.05631886050105095,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 3.7851438522338867,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600349040139616,
+ "grad_norm": 0.05489915981888771,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 3.7870893478393555,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61431064572426,
+ "grad_norm": 0.055331867188215256,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 3.8441572189331055,
+ "step": 2060
+ },
+ {
+ "epoch": 28.6282722513089,
+ "grad_norm": 0.05580870062112808,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 3.834148645401001,
+ "step": 2061
+ },
+ {
+ "epoch": 28.642233856893544,
+ "grad_norm": 0.05681769549846649,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 3.8349759578704834,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656195462478184,
+ "grad_norm": 0.05484442040324211,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 3.798792600631714,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670157068062828,
+ "grad_norm": 0.05657925084233284,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 3.853915214538574,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68411867364747,
+ "grad_norm": 0.05419609695672989,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 3.8423995971679688,
+ "step": 2065
+ },
+ {
+ "epoch": 28.698080279232112,
+ "grad_norm": 0.05700303241610527,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 3.822547674179077,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712041884816752,
+ "grad_norm": 0.055846985429525375,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 3.830044984817505,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726003490401396,
+ "grad_norm": 0.05754954740405083,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 3.868316173553467,
+ "step": 2068
+ },
+ {
+ "epoch": 28.739965095986037,
+ "grad_norm": 0.056265417486429214,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 3.8297243118286133,
+ "step": 2069
+ },
+ {
+ "epoch": 28.75392670157068,
+ "grad_norm": 0.05429232493042946,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 3.872836112976074,
+ "step": 2070
+ },
+ {
+ "epoch": 28.767888307155324,
+ "grad_norm": 0.05589764565229416,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 3.838311195373535,
+ "step": 2071
+ },
+ {
+ "epoch": 28.781849912739965,
+ "grad_norm": 0.05460003390908241,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 3.8316597938537598,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79581151832461,
+ "grad_norm": 0.054542504251003265,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 3.816542625427246,
+ "step": 2073
+ },
+ {
+ "epoch": 28.80977312390925,
+ "grad_norm": 0.0549996979534626,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 3.796245574951172,
+ "step": 2074
+ },
+ {
+ "epoch": 28.823734729493893,
+ "grad_norm": 0.05521940439939499,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 3.8112473487854004,
+ "step": 2075
+ },
+ {
+ "epoch": 28.837696335078533,
+ "grad_norm": 0.05690402165055275,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 3.8242125511169434,
+ "step": 2076
+ },
+ {
+ "epoch": 28.851657940663177,
+ "grad_norm": 0.055383820086717606,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 3.84938383102417,
+ "step": 2077
+ },
+ {
+ "epoch": 28.865619546247817,
+ "grad_norm": 0.05612736567854881,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 3.8383970260620117,
+ "step": 2078
+ },
+ {
+ "epoch": 28.87958115183246,
+ "grad_norm": 0.05723752826452255,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 3.828171730041504,
+ "step": 2079
+ },
+ {
+ "epoch": 28.8935427574171,
+ "grad_norm": 0.0576927475631237,
+ "learning_rate": 0.000242380656502223,
+ "loss": 3.8756160736083984,
+ "step": 2080
+ },
+ {
+ "epoch": 28.907504363001745,
+ "grad_norm": 0.05609026178717613,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 3.8703267574310303,
+ "step": 2081
+ },
+ {
+ "epoch": 28.921465968586386,
+ "grad_norm": 0.057038433849811554,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 3.8566956520080566,
+ "step": 2082
+ },
+ {
+ "epoch": 28.93542757417103,
+ "grad_norm": 0.05683749541640282,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 3.8409035205841064,
+ "step": 2083
+ },
+ {
+ "epoch": 28.949389179755673,
+ "grad_norm": 0.05682830885052681,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 3.8435707092285156,
+ "step": 2084
+ },
+ {
+ "epoch": 28.963350785340314,
+ "grad_norm": 0.05645477771759033,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 3.8142364025115967,
+ "step": 2085
+ },
+ {
+ "epoch": 28.977312390924958,
+ "grad_norm": 0.05519694834947586,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 3.830415725708008,
+ "step": 2086
+ },
+ {
+ "epoch": 28.991273996509598,
+ "grad_norm": 0.05785970389842987,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 3.8413705825805664,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.04018386825919151,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 2.3790979385375977,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 0.6047304272651672,
+ "eval_runtime": 61.1049,
+ "eval_samples_per_second": 39.964,
+ "eval_steps_per_second": 0.638,
+ "step": 2088
+ },
+ {
+ "epoch": 29.013961605584644,
+ "grad_norm": 0.058019861578941345,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 3.793783664703369,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027923211169284,
+ "grad_norm": 0.058003105223178864,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 3.783708095550537,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041884816753928,
+ "grad_norm": 0.05485067516565323,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 3.7829298973083496,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05584642233857,
+ "grad_norm": 0.056009069085121155,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 3.7747392654418945,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069808027923212,
+ "grad_norm": 0.05480542406439781,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 3.774707317352295,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083769633507853,
+ "grad_norm": 0.05384726822376251,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 3.75443959236145,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097731239092496,
+ "grad_norm": 0.053818102926015854,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 3.7790050506591797,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111692844677137,
+ "grad_norm": 0.05595344305038452,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 3.74898624420166,
+ "step": 2096
+ },
+ {
+ "epoch": 29.12565445026178,
+ "grad_norm": 0.05394545570015907,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 3.7751331329345703,
+ "step": 2097
+ },
+ {
+ "epoch": 29.13961605584642,
+ "grad_norm": 0.05346408113837242,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 3.7671167850494385,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153577661431065,
+ "grad_norm": 0.0541236512362957,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 3.749098062515259,
+ "step": 2099
+ },
+ {
+ "epoch": 29.167539267015705,
+ "grad_norm": 0.05369208753108978,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 3.7778515815734863,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18150087260035,
+ "grad_norm": 0.055362965911626816,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 3.747697591781616,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195462478184993,
+ "grad_norm": 0.05389911308884621,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 3.7759058475494385,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209424083769633,
+ "grad_norm": 0.05315527692437172,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 3.7832705974578857,
+ "step": 2103
+ },
+ {
+ "epoch": 29.223385689354277,
+ "grad_norm": 0.05377589911222458,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 3.7743210792541504,
+ "step": 2104
+ },
+ {
+ "epoch": 29.237347294938917,
+ "grad_norm": 0.05307800695300102,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 3.7654497623443604,
+ "step": 2105
+ },
+ {
+ "epoch": 29.25130890052356,
+ "grad_norm": 0.056719157844781876,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 3.794179916381836,
+ "step": 2106
+ },
+ {
+ "epoch": 29.2652705061082,
+ "grad_norm": 0.053281862288713455,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 3.7620739936828613,
+ "step": 2107
+ },
+ {
+ "epoch": 29.279232111692846,
+ "grad_norm": 0.05470574274659157,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 3.788421869277954,
+ "step": 2108
+ },
+ {
+ "epoch": 29.293193717277486,
+ "grad_norm": 0.05293853580951691,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 3.7655067443847656,
+ "step": 2109
+ },
+ {
+ "epoch": 29.30715532286213,
+ "grad_norm": 0.05524440482258797,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 3.7701616287231445,
+ "step": 2110
+ },
+ {
+ "epoch": 29.32111692844677,
+ "grad_norm": 0.05508396402001381,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 3.7993319034576416,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335078534031414,
+ "grad_norm": 0.05559603124856949,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 3.7799668312072754,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349040139616054,
+ "grad_norm": 0.05530122295022011,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 3.751032829284668,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363001745200698,
+ "grad_norm": 0.05408896878361702,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 3.7761611938476562,
+ "step": 2114
+ },
+ {
+ "epoch": 29.376963350785342,
+ "grad_norm": 0.05530737340450287,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 3.745713233947754,
+ "step": 2115
+ },
+ {
+ "epoch": 29.390924956369982,
+ "grad_norm": 0.05424542352557182,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 3.8041834831237793,
+ "step": 2116
+ },
+ {
+ "epoch": 29.404886561954626,
+ "grad_norm": 0.054800380021333694,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 3.7995352745056152,
+ "step": 2117
+ },
+ {
+ "epoch": 29.418848167539267,
+ "grad_norm": 0.05685900151729584,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 3.7970776557922363,
+ "step": 2118
+ },
+ {
+ "epoch": 29.43280977312391,
+ "grad_norm": 0.05555475875735283,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 3.7950801849365234,
+ "step": 2119
+ },
+ {
+ "epoch": 29.44677137870855,
+ "grad_norm": 0.05670571327209473,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 3.7967746257781982,
+ "step": 2120
+ },
+ {
+ "epoch": 29.460732984293195,
+ "grad_norm": 0.05598503723740578,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 3.809349298477173,
+ "step": 2121
+ },
+ {
+ "epoch": 29.474694589877835,
+ "grad_norm": 0.05501085892319679,
+ "learning_rate": 0.000231465389734324,
+ "loss": 3.803968906402588,
+ "step": 2122
+ },
+ {
+ "epoch": 29.48865619546248,
+ "grad_norm": 0.05519728362560272,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 3.7615163326263428,
+ "step": 2123
+ },
+ {
+ "epoch": 29.50261780104712,
+ "grad_norm": 0.055866893380880356,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 3.7939207553863525,
+ "step": 2124
+ },
+ {
+ "epoch": 29.516579406631763,
+ "grad_norm": 0.05460986867547035,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 3.7740321159362793,
+ "step": 2125
+ },
+ {
+ "epoch": 29.530541012216403,
+ "grad_norm": 0.055752526968717575,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 3.82424259185791,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544502617801047,
+ "grad_norm": 0.056792002171278,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 3.8006062507629395,
+ "step": 2127
+ },
+ {
+ "epoch": 29.55846422338569,
+ "grad_norm": 0.054986681789159775,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 3.7713897228240967,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57242582897033,
+ "grad_norm": 0.054215673357248306,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 3.8005690574645996,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586387434554975,
+ "grad_norm": 0.05700439214706421,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 3.825161933898926,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600349040139616,
+ "grad_norm": 0.055078551173210144,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 3.786716938018799,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61431064572426,
+ "grad_norm": 0.055569086223840714,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 3.8020172119140625,
+ "step": 2132
+ },
+ {
+ "epoch": 29.6282722513089,
+ "grad_norm": 0.0548536442220211,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 3.8038570880889893,
+ "step": 2133
+ },
+ {
+ "epoch": 29.642233856893544,
+ "grad_norm": 0.05604325234889984,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 3.796006441116333,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656195462478184,
+ "grad_norm": 0.05614005774259567,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 3.7938876152038574,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670157068062828,
+ "grad_norm": 0.05761241167783737,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 3.813568592071533,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68411867364747,
+ "grad_norm": 0.05579913780093193,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 3.826862335205078,
+ "step": 2137
+ },
+ {
+ "epoch": 29.698080279232112,
+ "grad_norm": 0.05784283205866814,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 3.808281421661377,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712041884816752,
+ "grad_norm": 0.05596453696489334,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 3.8119921684265137,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726003490401396,
+ "grad_norm": 0.05699361115694046,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 3.822702407836914,
+ "step": 2140
+ },
+ {
+ "epoch": 29.739965095986037,
+ "grad_norm": 0.05375700071454048,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 3.78023099899292,
+ "step": 2141
+ },
+ {
+ "epoch": 29.75392670157068,
+ "grad_norm": 0.05588206648826599,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 3.791456699371338,
+ "step": 2142
+ },
+ {
+ "epoch": 29.767888307155324,
+ "grad_norm": 0.05579312890768051,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 3.799713134765625,
+ "step": 2143
+ },
+ {
+ "epoch": 29.781849912739965,
+ "grad_norm": 0.05713854730129242,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 3.8211703300476074,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79581151832461,
+ "grad_norm": 0.05645165592432022,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 3.7947046756744385,
+ "step": 2145
+ },
+ {
+ "epoch": 29.80977312390925,
+ "grad_norm": 0.056201402097940445,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 3.811009407043457,
+ "step": 2146
+ },
+ {
+ "epoch": 29.823734729493893,
+ "grad_norm": 0.05659479275345802,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 3.8009982109069824,
+ "step": 2147
+ },
+ {
+ "epoch": 29.837696335078533,
+ "grad_norm": 0.05797034874558449,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 3.7994565963745117,
+ "step": 2148
+ },
+ {
+ "epoch": 29.851657940663177,
+ "grad_norm": 0.05734949931502342,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 3.7696728706359863,
+ "step": 2149
+ },
+ {
+ "epoch": 29.865619546247817,
+ "grad_norm": 0.060396358370780945,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 3.8269689083099365,
+ "step": 2150
+ },
+ {
+ "epoch": 29.87958115183246,
+ "grad_norm": 0.057348065078258514,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 3.7925050258636475,
+ "step": 2151
+ },
+ {
+ "epoch": 29.8935427574171,
+ "grad_norm": 0.058563247323036194,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 3.82887864112854,
+ "step": 2152
+ },
+ {
+ "epoch": 29.907504363001745,
+ "grad_norm": 0.06011585891246796,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 3.8061296939849854,
+ "step": 2153
+ },
+ {
+ "epoch": 29.921465968586386,
+ "grad_norm": 0.058440592139959335,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 3.792586326599121,
+ "step": 2154
+ },
+ {
+ "epoch": 29.93542757417103,
+ "grad_norm": 0.057050157338380814,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 3.7885212898254395,
+ "step": 2155
+ },
+ {
+ "epoch": 29.949389179755673,
+ "grad_norm": 0.05920996889472008,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 3.813566207885742,
+ "step": 2156
+ },
+ {
+ "epoch": 29.963350785340314,
+ "grad_norm": 0.057791322469711304,
+ "learning_rate": 0.00022244633283095,
+ "loss": 3.8399810791015625,
+ "step": 2157
+ },
+ {
+ "epoch": 29.977312390924958,
+ "grad_norm": 0.05644001066684723,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 3.79464054107666,
+ "step": 2158
+ },
+ {
+ "epoch": 29.991273996509598,
+ "grad_norm": 0.06021320819854736,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 3.804086208343506,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.040621865540742874,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 2.386565923690796,
+ "step": 2160
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 9.216882739091866e+17,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2160/training_args.bin b/runs/l2r90-wd0033-ssl/checkpoint-2160/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..313c41733ff55284c06034a59bb885b30e40ffc8
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2160/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7ddd0bdddba68d3e19ab33477000f3fc684fe4ee8e3580eeb8c02646e5fc27ac
+size 4856
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2520/chat_template.jinja b/runs/l2r90-wd0033-ssl/checkpoint-2520/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2520/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2520/config.json b/runs/l2r90-wd0033-ssl/checkpoint-2520/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2520/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2520/generation_config.json b/runs/l2r90-wd0033-ssl/checkpoint-2520/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2520/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2520/model.safetensors b/runs/l2r90-wd0033-ssl/checkpoint-2520/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..d29a1fcabe8d175ea0fa3432cd75abe7ac39e43d
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2520/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:20c647cbf8ef68a7c1da4054b7156fb87dea1a5be3ad7704a6a2b788a03007ea
+size 583356232
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2520/optimizer.pt b/runs/l2r90-wd0033-ssl/checkpoint-2520/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..67847788913041b0060a343ff98ca70942162483
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2520/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3a0fa9f7a2247fb0ead7a5f36db12ecbbc51e628696d1d4283334a995d9ce989
+size 1166827898
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2520/rng_state_0.pth b/runs/l2r90-wd0033-ssl/checkpoint-2520/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..5a1ed108c5e3dd5a0d01545311a2fbfc85ac080c
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2520/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7e38b28450b484bae9066d3591189b4df81a5b8528d06e5a2b200106829c1ea0
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2520/rng_state_1.pth b/runs/l2r90-wd0033-ssl/checkpoint-2520/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..bd3984d5a741ff19bdaf4610d6ecd59df0a8dab0
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2520/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5e143c6e2599c38a9fc0815513d0cafda411c6b0eef6f87171d11fee664205d0
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2520/scheduler.pt b/runs/l2r90-wd0033-ssl/checkpoint-2520/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..b53f13f877641b522f2c435934f587fdad6161cc
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2520/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6f776ea81adcefd29eff4d7371e55373e8890bd750697e1c8b781d131e0a39a6
+size 1064
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2520/tokenizer.json b/runs/l2r90-wd0033-ssl/checkpoint-2520/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2520/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2520/tokenizer_config.json b/runs/l2r90-wd0033-ssl/checkpoint-2520/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2520/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2520/trainer_state.json b/runs/l2r90-wd0033-ssl/checkpoint-2520/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..e44af58d930fdf374979f52065f7db30636167a1
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2520/trainer_state.json
@@ -0,0 +1,17938 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 35.0,
+ "eval_steps": 500,
+ "global_step": 2520,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013961605584642234,
+ "grad_norm": 0.26733073592185974,
+ "learning_rate": 0.0,
+ "loss": 7.990724086761475,
+ "step": 1
+ },
+ {
+ "epoch": 0.027923211169284468,
+ "grad_norm": 0.26712340116500854,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 7.990110874176025,
+ "step": 2
+ },
+ {
+ "epoch": 0.041884816753926704,
+ "grad_norm": 0.2646932601928711,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 7.956935405731201,
+ "step": 3
+ },
+ {
+ "epoch": 0.055846422338568937,
+ "grad_norm": 0.24684545397758484,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 7.890539646148682,
+ "step": 4
+ },
+ {
+ "epoch": 0.06980802792321117,
+ "grad_norm": 0.21016643941402435,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 7.8146138191223145,
+ "step": 5
+ },
+ {
+ "epoch": 0.08376963350785341,
+ "grad_norm": 0.17992742359638214,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 7.724307060241699,
+ "step": 6
+ },
+ {
+ "epoch": 0.09773123909249563,
+ "grad_norm": 0.14356893301010132,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 7.645662307739258,
+ "step": 7
+ },
+ {
+ "epoch": 0.11169284467713787,
+ "grad_norm": 0.10830964893102646,
+ "learning_rate": 4.2e-05,
+ "loss": 7.582019329071045,
+ "step": 8
+ },
+ {
+ "epoch": 0.1256544502617801,
+ "grad_norm": 0.08265161514282227,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 7.5230793952941895,
+ "step": 9
+ },
+ {
+ "epoch": 0.13961605584642234,
+ "grad_norm": 0.06423710286617279,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 7.473812103271484,
+ "step": 10
+ },
+ {
+ "epoch": 0.15357766143106458,
+ "grad_norm": 0.05132246017456055,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 7.440786838531494,
+ "step": 11
+ },
+ {
+ "epoch": 0.16753926701570682,
+ "grad_norm": 0.040572404861450195,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 7.41290283203125,
+ "step": 12
+ },
+ {
+ "epoch": 0.18150087260034903,
+ "grad_norm": 0.03138705715537071,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 7.39749002456665,
+ "step": 13
+ },
+ {
+ "epoch": 0.19546247818499127,
+ "grad_norm": 0.02553008496761322,
+ "learning_rate": 7.8e-05,
+ "loss": 7.387092590332031,
+ "step": 14
+ },
+ {
+ "epoch": 0.2094240837696335,
+ "grad_norm": 0.022178977727890015,
+ "learning_rate": 8.4e-05,
+ "loss": 7.378993511199951,
+ "step": 15
+ },
+ {
+ "epoch": 0.22338568935427575,
+ "grad_norm": 0.019915293902158737,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 7.37220573425293,
+ "step": 16
+ },
+ {
+ "epoch": 0.23734729493891799,
+ "grad_norm": 0.017390629276633263,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 7.366759300231934,
+ "step": 17
+ },
+ {
+ "epoch": 0.2513089005235602,
+ "grad_norm": 0.015472771599888802,
+ "learning_rate": 0.000102,
+ "loss": 7.3652663230896,
+ "step": 18
+ },
+ {
+ "epoch": 0.26527050610820246,
+ "grad_norm": 0.013599214144051075,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 7.359714984893799,
+ "step": 19
+ },
+ {
+ "epoch": 0.2792321116928447,
+ "grad_norm": 0.012442861683666706,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 7.353302001953125,
+ "step": 20
+ },
+ {
+ "epoch": 0.2931937172774869,
+ "grad_norm": 0.010471965186297894,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 7.357417583465576,
+ "step": 21
+ },
+ {
+ "epoch": 0.30715532286212915,
+ "grad_norm": 0.009647224098443985,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 7.351763725280762,
+ "step": 22
+ },
+ {
+ "epoch": 0.32111692844677137,
+ "grad_norm": 0.008228210732340813,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 7.345434188842773,
+ "step": 23
+ },
+ {
+ "epoch": 0.33507853403141363,
+ "grad_norm": 0.007454514969140291,
+ "learning_rate": 0.000138,
+ "loss": 7.342520236968994,
+ "step": 24
+ },
+ {
+ "epoch": 0.34904013961605584,
+ "grad_norm": 0.0069856527261435986,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 7.338870048522949,
+ "step": 25
+ },
+ {
+ "epoch": 0.36300174520069806,
+ "grad_norm": 0.006091665010899305,
+ "learning_rate": 0.00015,
+ "loss": 7.328121662139893,
+ "step": 26
+ },
+ {
+ "epoch": 0.3769633507853403,
+ "grad_norm": 0.005525320768356323,
+ "learning_rate": 0.000156,
+ "loss": 7.335029602050781,
+ "step": 27
+ },
+ {
+ "epoch": 0.39092495636998253,
+ "grad_norm": 0.0061133443377912045,
+ "learning_rate": 0.000162,
+ "loss": 7.319417953491211,
+ "step": 28
+ },
+ {
+ "epoch": 0.4048865619546248,
+ "grad_norm": 0.006040909793227911,
+ "learning_rate": 0.000168,
+ "loss": 7.326376438140869,
+ "step": 29
+ },
+ {
+ "epoch": 0.418848167539267,
+ "grad_norm": 0.006003216840326786,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 7.319552421569824,
+ "step": 30
+ },
+ {
+ "epoch": 0.4328097731239092,
+ "grad_norm": 0.007096003741025925,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 7.323486328125,
+ "step": 31
+ },
+ {
+ "epoch": 0.4467713787085515,
+ "grad_norm": 0.007725459989160299,
+ "learning_rate": 0.000186,
+ "loss": 7.3190717697143555,
+ "step": 32
+ },
+ {
+ "epoch": 0.4607329842931937,
+ "grad_norm": 0.007338670082390308,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 7.317154884338379,
+ "step": 33
+ },
+ {
+ "epoch": 0.47469458987783597,
+ "grad_norm": 0.007709544617682695,
+ "learning_rate": 0.000198,
+ "loss": 7.311826705932617,
+ "step": 34
+ },
+ {
+ "epoch": 0.4886561954624782,
+ "grad_norm": 0.009249784983694553,
+ "learning_rate": 0.000204,
+ "loss": 7.322498798370361,
+ "step": 35
+ },
+ {
+ "epoch": 0.5026178010471204,
+ "grad_norm": 0.006950767710804939,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 7.314571380615234,
+ "step": 36
+ },
+ {
+ "epoch": 0.5165794066317626,
+ "grad_norm": 0.007583985570818186,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 7.309540748596191,
+ "step": 37
+ },
+ {
+ "epoch": 0.5305410122164049,
+ "grad_norm": 0.014731820672750473,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 7.307004928588867,
+ "step": 38
+ },
+ {
+ "epoch": 0.5445026178010471,
+ "grad_norm": 0.016276473179459572,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 7.299642562866211,
+ "step": 39
+ },
+ {
+ "epoch": 0.5584642233856894,
+ "grad_norm": 0.020018689334392548,
+ "learning_rate": 0.000234,
+ "loss": 7.298466682434082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5724258289703316,
+ "grad_norm": 0.014941536821424961,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 7.289968967437744,
+ "step": 41
+ },
+ {
+ "epoch": 0.5863874345549738,
+ "grad_norm": 0.018299812451004982,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 7.276564598083496,
+ "step": 42
+ },
+ {
+ "epoch": 0.6003490401396161,
+ "grad_norm": 0.01615075021982193,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.2575554847717285,
+ "step": 43
+ },
+ {
+ "epoch": 0.6143106457242583,
+ "grad_norm": 0.013063831254839897,
+ "learning_rate": 0.000258,
+ "loss": 7.253885269165039,
+ "step": 44
+ },
+ {
+ "epoch": 0.6282722513089005,
+ "grad_norm": 0.015211676247417927,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.245779037475586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6422338568935427,
+ "grad_norm": 0.02093500830233097,
+ "learning_rate": 0.00027,
+ "loss": 7.225130081176758,
+ "step": 46
+ },
+ {
+ "epoch": 0.6561954624781849,
+ "grad_norm": 0.01808053068816662,
+ "learning_rate": 0.000276,
+ "loss": 7.209654331207275,
+ "step": 47
+ },
+ {
+ "epoch": 0.6701570680628273,
+ "grad_norm": 0.02209644205868244,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.197124481201172,
+ "step": 48
+ },
+ {
+ "epoch": 0.6841186736474695,
+ "grad_norm": 0.023524878546595573,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.1841936111450195,
+ "step": 49
+ },
+ {
+ "epoch": 0.6980802792321117,
+ "grad_norm": 0.031148087233304977,
+ "learning_rate": 0.000294,
+ "loss": 7.188040733337402,
+ "step": 50
+ },
+ {
+ "epoch": 0.7120418848167539,
+ "grad_norm": 0.02931929938495159,
+ "learning_rate": 0.0003,
+ "loss": 7.17371940612793,
+ "step": 51
+ },
+ {
+ "epoch": 0.7260034904013961,
+ "grad_norm": 0.01611483097076416,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.1714324951171875,
+ "step": 52
+ },
+ {
+ "epoch": 0.7399650959860384,
+ "grad_norm": 0.02619907818734646,
+ "learning_rate": 0.000312,
+ "loss": 7.158261299133301,
+ "step": 53
+ },
+ {
+ "epoch": 0.7539267015706806,
+ "grad_norm": 0.049462560564279556,
+ "learning_rate": 0.000318,
+ "loss": 7.167076110839844,
+ "step": 54
+ },
+ {
+ "epoch": 0.7678883071553229,
+ "grad_norm": 0.09835070371627808,
+ "learning_rate": 0.000324,
+ "loss": 7.154292106628418,
+ "step": 55
+ },
+ {
+ "epoch": 0.7818499127399651,
+ "grad_norm": 0.11406510323286057,
+ "learning_rate": 0.00033,
+ "loss": 7.162016868591309,
+ "step": 56
+ },
+ {
+ "epoch": 0.7958115183246073,
+ "grad_norm": 0.027026303112506866,
+ "learning_rate": 0.000336,
+ "loss": 7.141416549682617,
+ "step": 57
+ },
+ {
+ "epoch": 0.8097731239092496,
+ "grad_norm": 0.06690579652786255,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.13959264755249,
+ "step": 58
+ },
+ {
+ "epoch": 0.8237347294938918,
+ "grad_norm": 0.020919431000947952,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.126513481140137,
+ "step": 59
+ },
+ {
+ "epoch": 0.837696335078534,
+ "grad_norm": 0.05618130415678024,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.124345779418945,
+ "step": 60
+ },
+ {
+ "epoch": 0.8516579406631762,
+ "grad_norm": 0.03402302414178848,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.100615501403809,
+ "step": 61
+ },
+ {
+ "epoch": 0.8656195462478184,
+ "grad_norm": 0.03646906837821007,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.116507530212402,
+ "step": 62
+ },
+ {
+ "epoch": 0.8795811518324608,
+ "grad_norm": 0.03076397068798542,
+ "learning_rate": 0.000372,
+ "loss": 7.097659111022949,
+ "step": 63
+ },
+ {
+ "epoch": 0.893542757417103,
+ "grad_norm": 0.03329646959900856,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.097290992736816,
+ "step": 64
+ },
+ {
+ "epoch": 0.9075043630017452,
+ "grad_norm": 0.01998630352318287,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.092268943786621,
+ "step": 65
+ },
+ {
+ "epoch": 0.9214659685863874,
+ "grad_norm": 0.03278940171003342,
+ "learning_rate": 0.00039,
+ "loss": 7.088292121887207,
+ "step": 66
+ },
+ {
+ "epoch": 0.9354275741710296,
+ "grad_norm": 0.024620421230793,
+ "learning_rate": 0.000396,
+ "loss": 7.078650951385498,
+ "step": 67
+ },
+ {
+ "epoch": 0.9493891797556719,
+ "grad_norm": 0.03750383481383324,
+ "learning_rate": 0.000402,
+ "loss": 7.052778244018555,
+ "step": 68
+ },
+ {
+ "epoch": 0.9633507853403142,
+ "grad_norm": 0.07558075338602066,
+ "learning_rate": 0.000408,
+ "loss": 7.045991897583008,
+ "step": 69
+ },
+ {
+ "epoch": 0.9773123909249564,
+ "grad_norm": 0.16021482646465302,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.023152828216553,
+ "step": 70
+ },
+ {
+ "epoch": 0.9912739965095986,
+ "grad_norm": 0.24981893599033356,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.067540645599365,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.10255520790815353,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 4.425640106201172,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8861758708953857,
+ "eval_runtime": 58.2699,
+ "eval_samples_per_second": 41.908,
+ "eval_steps_per_second": 0.669,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139616055846423,
+ "grad_norm": 0.13374797999858856,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.073373317718506,
+ "step": 73
+ },
+ {
+ "epoch": 1.0279232111692844,
+ "grad_norm": 0.05079387128353119,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.0243306159973145,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418848167539267,
+ "grad_norm": 0.10056068748235703,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.014476299285889,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558464223385688,
+ "grad_norm": 0.06995758414268494,
+ "learning_rate": 0.00045,
+ "loss": 7.011931419372559,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698080279232112,
+ "grad_norm": 0.0483892522752285,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 6.9822797775268555,
+ "step": 77
+ },
+ {
+ "epoch": 1.0837696335078535,
+ "grad_norm": 0.08408170193433762,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 6.964348793029785,
+ "step": 78
+ },
+ {
+ "epoch": 1.0977312390924956,
+ "grad_norm": 0.08027710020542145,
+ "learning_rate": 0.000468,
+ "loss": 6.96852445602417,
+ "step": 79
+ },
+ {
+ "epoch": 1.111692844677138,
+ "grad_norm": 0.05647159367799759,
+ "learning_rate": 0.000474,
+ "loss": 6.944484710693359,
+ "step": 80
+ },
+ {
+ "epoch": 1.12565445026178,
+ "grad_norm": 0.07464330643415451,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 6.9246745109558105,
+ "step": 81
+ },
+ {
+ "epoch": 1.1396160558464223,
+ "grad_norm": 0.04451674595475197,
+ "learning_rate": 0.000486,
+ "loss": 6.895403861999512,
+ "step": 82
+ },
+ {
+ "epoch": 1.1535776614310647,
+ "grad_norm": 0.05820256471633911,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 6.887338638305664,
+ "step": 83
+ },
+ {
+ "epoch": 1.1675392670157068,
+ "grad_norm": 0.02925296127796173,
+ "learning_rate": 0.000498,
+ "loss": 6.857797622680664,
+ "step": 84
+ },
+ {
+ "epoch": 1.181500872600349,
+ "grad_norm": 0.05096987634897232,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 6.854822158813477,
+ "step": 85
+ },
+ {
+ "epoch": 1.1954624781849912,
+ "grad_norm": 0.03483197093009949,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 6.840914726257324,
+ "step": 86
+ },
+ {
+ "epoch": 1.2094240837696335,
+ "grad_norm": 0.03323264792561531,
+ "learning_rate": 0.000516,
+ "loss": 6.834497451782227,
+ "step": 87
+ },
+ {
+ "epoch": 1.2233856893542758,
+ "grad_norm": 0.03290868178009987,
+ "learning_rate": 0.000522,
+ "loss": 6.820746898651123,
+ "step": 88
+ },
+ {
+ "epoch": 1.237347294938918,
+ "grad_norm": 0.025398844853043556,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 6.795746803283691,
+ "step": 89
+ },
+ {
+ "epoch": 1.2513089005235603,
+ "grad_norm": 0.030700596049427986,
+ "learning_rate": 0.000534,
+ "loss": 6.7867889404296875,
+ "step": 90
+ },
+ {
+ "epoch": 1.2652705061082026,
+ "grad_norm": 0.019589420408010483,
+ "learning_rate": 0.00054,
+ "loss": 6.77443790435791,
+ "step": 91
+ },
+ {
+ "epoch": 1.2792321116928447,
+ "grad_norm": 0.03313920646905899,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 6.762271881103516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2931937172774868,
+ "grad_norm": 0.025389468297362328,
+ "learning_rate": 0.000552,
+ "loss": 6.750258445739746,
+ "step": 93
+ },
+ {
+ "epoch": 1.307155322862129,
+ "grad_norm": 0.040296655148267746,
+ "learning_rate": 0.000558,
+ "loss": 6.729808807373047,
+ "step": 94
+ },
+ {
+ "epoch": 1.3211169284467714,
+ "grad_norm": 0.04735933244228363,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 6.737982749938965,
+ "step": 95
+ },
+ {
+ "epoch": 1.3350785340314135,
+ "grad_norm": 0.05529680848121643,
+ "learning_rate": 0.00057,
+ "loss": 6.735714912414551,
+ "step": 96
+ },
+ {
+ "epoch": 1.3490401396160558,
+ "grad_norm": 0.07148554176092148,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 6.708859920501709,
+ "step": 97
+ },
+ {
+ "epoch": 1.3630017452006982,
+ "grad_norm": 0.059708479791879654,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 6.708860397338867,
+ "step": 98
+ },
+ {
+ "epoch": 1.3769633507853403,
+ "grad_norm": 0.027111195027828217,
+ "learning_rate": 0.000588,
+ "loss": 6.686685085296631,
+ "step": 99
+ },
+ {
+ "epoch": 1.3909249563699826,
+ "grad_norm": 0.0500967763364315,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 6.676325798034668,
+ "step": 100
+ },
+ {
+ "epoch": 1.404886561954625,
+ "grad_norm": 0.044565796852111816,
+ "learning_rate": 0.0006,
+ "loss": 6.674530029296875,
+ "step": 101
+ },
+ {
+ "epoch": 1.418848167539267,
+ "grad_norm": 0.03608180209994316,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 6.653741359710693,
+ "step": 102
+ },
+ {
+ "epoch": 1.4328097731239091,
+ "grad_norm": 0.030394982546567917,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 6.643152236938477,
+ "step": 103
+ },
+ {
+ "epoch": 1.4467713787085514,
+ "grad_norm": 0.040185656398534775,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 6.628686904907227,
+ "step": 104
+ },
+ {
+ "epoch": 1.4607329842931938,
+ "grad_norm": 0.021049322560429573,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 6.610864639282227,
+ "step": 105
+ },
+ {
+ "epoch": 1.4746945898778359,
+ "grad_norm": 0.04164277762174606,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 6.602880954742432,
+ "step": 106
+ },
+ {
+ "epoch": 1.4886561954624782,
+ "grad_norm": 0.0389864556491375,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 6.599456787109375,
+ "step": 107
+ },
+ {
+ "epoch": 1.5026178010471205,
+ "grad_norm": 0.05805973336100578,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 6.588700294494629,
+ "step": 108
+ },
+ {
+ "epoch": 1.5165794066317626,
+ "grad_norm": 0.08825036883354187,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 6.587496757507324,
+ "step": 109
+ },
+ {
+ "epoch": 1.530541012216405,
+ "grad_norm": 0.12521523237228394,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 6.626501083374023,
+ "step": 110
+ },
+ {
+ "epoch": 1.5445026178010473,
+ "grad_norm": 0.07429645210504532,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 6.590241432189941,
+ "step": 111
+ },
+ {
+ "epoch": 1.5584642233856894,
+ "grad_norm": 0.06463009119033813,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 6.578714370727539,
+ "step": 112
+ },
+ {
+ "epoch": 1.5724258289703315,
+ "grad_norm": 0.07508203387260437,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 6.561270713806152,
+ "step": 113
+ },
+ {
+ "epoch": 1.5863874345549738,
+ "grad_norm": 0.04125197231769562,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 6.558821678161621,
+ "step": 114
+ },
+ {
+ "epoch": 1.600349040139616,
+ "grad_norm": 0.05457386001944542,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.541611671447754,
+ "step": 115
+ },
+ {
+ "epoch": 1.6143106457242582,
+ "grad_norm": 0.03684385493397713,
+ "learning_rate": 0.000599973080557108,
+ "loss": 6.547738075256348,
+ "step": 116
+ },
+ {
+ "epoch": 1.6282722513089005,
+ "grad_norm": 0.05888746306300163,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.529491424560547,
+ "step": 117
+ },
+ {
+ "epoch": 1.6422338568935428,
+ "grad_norm": 0.056668370962142944,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.519444942474365,
+ "step": 118
+ },
+ {
+ "epoch": 1.656195462478185,
+ "grad_norm": 0.09123896062374115,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.522984981536865,
+ "step": 119
+ },
+ {
+ "epoch": 1.6701570680628273,
+ "grad_norm": 0.12380023300647736,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.515375137329102,
+ "step": 120
+ },
+ {
+ "epoch": 1.6841186736474696,
+ "grad_norm": 0.06953322887420654,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.500491142272949,
+ "step": 121
+ },
+ {
+ "epoch": 1.6980802792321117,
+ "grad_norm": 0.08613869547843933,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.5097198486328125,
+ "step": 122
+ },
+ {
+ "epoch": 1.7120418848167538,
+ "grad_norm": 0.08296383172273636,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.487521171569824,
+ "step": 123
+ },
+ {
+ "epoch": 1.7260034904013961,
+ "grad_norm": 0.05744783207774162,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.485134124755859,
+ "step": 124
+ },
+ {
+ "epoch": 1.7399650959860384,
+ "grad_norm": 0.04108566418290138,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.466182708740234,
+ "step": 125
+ },
+ {
+ "epoch": 1.7539267015706805,
+ "grad_norm": 0.05374931916594505,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.479011535644531,
+ "step": 126
+ },
+ {
+ "epoch": 1.7678883071553229,
+ "grad_norm": 0.03867914527654648,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.465902328491211,
+ "step": 127
+ },
+ {
+ "epoch": 1.7818499127399652,
+ "grad_norm": 0.05477374792098999,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.453082084655762,
+ "step": 128
+ },
+ {
+ "epoch": 1.7958115183246073,
+ "grad_norm": 0.06508298963308334,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.446089744567871,
+ "step": 129
+ },
+ {
+ "epoch": 1.8097731239092496,
+ "grad_norm": 0.06388624757528305,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.451748847961426,
+ "step": 130
+ },
+ {
+ "epoch": 1.823734729493892,
+ "grad_norm": 0.04223188757896423,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.4492387771606445,
+ "step": 131
+ },
+ {
+ "epoch": 1.837696335078534,
+ "grad_norm": 0.04528658092021942,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.409390926361084,
+ "step": 132
+ },
+ {
+ "epoch": 1.8516579406631761,
+ "grad_norm": 0.04962967336177826,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.415774345397949,
+ "step": 133
+ },
+ {
+ "epoch": 1.8656195462478184,
+ "grad_norm": 0.031944390386343,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.411528587341309,
+ "step": 134
+ },
+ {
+ "epoch": 1.8795811518324608,
+ "grad_norm": 0.028411149978637695,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.392329216003418,
+ "step": 135
+ },
+ {
+ "epoch": 1.8935427574171029,
+ "grad_norm": 0.03516148403286934,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.392238616943359,
+ "step": 136
+ },
+ {
+ "epoch": 1.9075043630017452,
+ "grad_norm": 0.027800120413303375,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.39179801940918,
+ "step": 137
+ },
+ {
+ "epoch": 1.9214659685863875,
+ "grad_norm": 0.02362987957894802,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.378960609436035,
+ "step": 138
+ },
+ {
+ "epoch": 1.9354275741710296,
+ "grad_norm": 0.033999308943748474,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.36802339553833,
+ "step": 139
+ },
+ {
+ "epoch": 1.949389179755672,
+ "grad_norm": 0.03290879726409912,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.366308689117432,
+ "step": 140
+ },
+ {
+ "epoch": 1.9633507853403143,
+ "grad_norm": 0.0286843404173851,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.350399971008301,
+ "step": 141
+ },
+ {
+ "epoch": 1.9773123909249564,
+ "grad_norm": 0.027145877480506897,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.33477783203125,
+ "step": 142
+ },
+ {
+ "epoch": 1.9912739965095985,
+ "grad_norm": 0.03530878946185112,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.322859764099121,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.028133150190114975,
+ "learning_rate": 0.000599778806120077,
+ "loss": 3.955288887023926,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7927101254463196,
+ "eval_runtime": 60.8372,
+ "eval_samples_per_second": 40.14,
+ "eval_steps_per_second": 0.641,
+ "step": 144
+ },
+ {
+ "epoch": 2.013961605584642,
+ "grad_norm": 0.03054557554423809,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.311975479125977,
+ "step": 145
+ },
+ {
+ "epoch": 2.0279232111692846,
+ "grad_norm": 0.05406459793448448,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.3030853271484375,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418848167539267,
+ "grad_norm": 0.10911019891500473,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.329056739807129,
+ "step": 147
+ },
+ {
+ "epoch": 2.055846422338569,
+ "grad_norm": 0.13948315382003784,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.362421035766602,
+ "step": 148
+ },
+ {
+ "epoch": 2.0698080279232114,
+ "grad_norm": 0.13176991045475006,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.3888139724731445,
+ "step": 149
+ },
+ {
+ "epoch": 2.0837696335078535,
+ "grad_norm": 0.09615885466337204,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.35154914855957,
+ "step": 150
+ },
+ {
+ "epoch": 2.0977312390924956,
+ "grad_norm": 0.06462561339139938,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.335790634155273,
+ "step": 151
+ },
+ {
+ "epoch": 2.1116928446771377,
+ "grad_norm": 0.05186621472239494,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.324396133422852,
+ "step": 152
+ },
+ {
+ "epoch": 2.1256544502617802,
+ "grad_norm": 0.06095981225371361,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.318549156188965,
+ "step": 153
+ },
+ {
+ "epoch": 2.1396160558464223,
+ "grad_norm": 0.036121442914009094,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.300607681274414,
+ "step": 154
+ },
+ {
+ "epoch": 2.1535776614310644,
+ "grad_norm": 0.055716633796691895,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.307400703430176,
+ "step": 155
+ },
+ {
+ "epoch": 2.167539267015707,
+ "grad_norm": 0.037626396864652634,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.287364959716797,
+ "step": 156
+ },
+ {
+ "epoch": 2.181500872600349,
+ "grad_norm": 0.04329098388552666,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.28244686126709,
+ "step": 157
+ },
+ {
+ "epoch": 2.195462478184991,
+ "grad_norm": 0.031868450343608856,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.2602410316467285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2094240837696333,
+ "grad_norm": 0.03647344186902046,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.263988971710205,
+ "step": 159
+ },
+ {
+ "epoch": 2.223385689354276,
+ "grad_norm": 0.03141765296459198,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.263205051422119,
+ "step": 160
+ },
+ {
+ "epoch": 2.237347294938918,
+ "grad_norm": 0.028253229334950447,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.253107070922852,
+ "step": 161
+ },
+ {
+ "epoch": 2.25130890052356,
+ "grad_norm": 0.03608971834182739,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.2571702003479,
+ "step": 162
+ },
+ {
+ "epoch": 2.2652705061082026,
+ "grad_norm": 0.02644292451441288,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.241732597351074,
+ "step": 163
+ },
+ {
+ "epoch": 2.2792321116928447,
+ "grad_norm": 0.028604455292224884,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.230289459228516,
+ "step": 164
+ },
+ {
+ "epoch": 2.2931937172774868,
+ "grad_norm": 0.023704614490270615,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.226101398468018,
+ "step": 165
+ },
+ {
+ "epoch": 2.3071553228621293,
+ "grad_norm": 0.020691027864813805,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.2096405029296875,
+ "step": 166
+ },
+ {
+ "epoch": 2.3211169284467714,
+ "grad_norm": 0.025598427280783653,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.22075080871582,
+ "step": 167
+ },
+ {
+ "epoch": 2.3350785340314135,
+ "grad_norm": 0.030546002089977264,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.216886520385742,
+ "step": 168
+ },
+ {
+ "epoch": 2.349040139616056,
+ "grad_norm": 0.026066245511174202,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.2022504806518555,
+ "step": 169
+ },
+ {
+ "epoch": 2.363001745200698,
+ "grad_norm": 0.023802239447832108,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.189509391784668,
+ "step": 170
+ },
+ {
+ "epoch": 2.3769633507853403,
+ "grad_norm": 0.029805002734065056,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.186277389526367,
+ "step": 171
+ },
+ {
+ "epoch": 2.3909249563699824,
+ "grad_norm": 0.043327830731868744,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.164876937866211,
+ "step": 172
+ },
+ {
+ "epoch": 2.404886561954625,
+ "grad_norm": 0.058787211775779724,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.179936408996582,
+ "step": 173
+ },
+ {
+ "epoch": 2.418848167539267,
+ "grad_norm": 0.0787397176027298,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.1830244064331055,
+ "step": 174
+ },
+ {
+ "epoch": 2.432809773123909,
+ "grad_norm": 0.0899854376912117,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.190343856811523,
+ "step": 175
+ },
+ {
+ "epoch": 2.4467713787085517,
+ "grad_norm": 0.06520707905292511,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.185057163238525,
+ "step": 176
+ },
+ {
+ "epoch": 2.4607329842931938,
+ "grad_norm": 0.1002504974603653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.171279430389404,
+ "step": 177
+ },
+ {
+ "epoch": 2.474694589877836,
+ "grad_norm": 0.09591004997491837,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.174382209777832,
+ "step": 178
+ },
+ {
+ "epoch": 2.488656195462478,
+ "grad_norm": 0.12127121537923813,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.185717582702637,
+ "step": 179
+ },
+ {
+ "epoch": 2.5026178010471205,
+ "grad_norm": 0.09640590846538544,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.196681976318359,
+ "step": 180
+ },
+ {
+ "epoch": 2.5165794066317626,
+ "grad_norm": 0.048341453075408936,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.158774375915527,
+ "step": 181
+ },
+ {
+ "epoch": 2.530541012216405,
+ "grad_norm": 0.08206334710121155,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.17445707321167,
+ "step": 182
+ },
+ {
+ "epoch": 2.5445026178010473,
+ "grad_norm": 0.038934968411922455,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.170527458190918,
+ "step": 183
+ },
+ {
+ "epoch": 2.5584642233856894,
+ "grad_norm": 0.060987457633018494,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.1501994132995605,
+ "step": 184
+ },
+ {
+ "epoch": 2.5724258289703315,
+ "grad_norm": 0.033269546926021576,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.144186019897461,
+ "step": 185
+ },
+ {
+ "epoch": 2.5863874345549736,
+ "grad_norm": 0.046137236058712006,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.1236677169799805,
+ "step": 186
+ },
+ {
+ "epoch": 2.600349040139616,
+ "grad_norm": 0.038265109062194824,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.121979236602783,
+ "step": 187
+ },
+ {
+ "epoch": 2.614310645724258,
+ "grad_norm": 0.039128001779317856,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.124156951904297,
+ "step": 188
+ },
+ {
+ "epoch": 2.6282722513089007,
+ "grad_norm": 0.025303540751338005,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.108036994934082,
+ "step": 189
+ },
+ {
+ "epoch": 2.642233856893543,
+ "grad_norm": 0.04441308230161667,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.1211161613464355,
+ "step": 190
+ },
+ {
+ "epoch": 2.656195462478185,
+ "grad_norm": 0.03083532303571701,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.094561576843262,
+ "step": 191
+ },
+ {
+ "epoch": 2.670157068062827,
+ "grad_norm": 0.02951296605169773,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.091608047485352,
+ "step": 192
+ },
+ {
+ "epoch": 2.6841186736474696,
+ "grad_norm": 0.02385987527668476,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.087080001831055,
+ "step": 193
+ },
+ {
+ "epoch": 2.6980802792321117,
+ "grad_norm": 0.03315780311822891,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.088132381439209,
+ "step": 194
+ },
+ {
+ "epoch": 2.712041884816754,
+ "grad_norm": 0.02410835586488247,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.081460952758789,
+ "step": 195
+ },
+ {
+ "epoch": 2.7260034904013963,
+ "grad_norm": 0.033069487661123276,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.0696539878845215,
+ "step": 196
+ },
+ {
+ "epoch": 2.7399650959860384,
+ "grad_norm": 0.031288471072912216,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.047612190246582,
+ "step": 197
+ },
+ {
+ "epoch": 2.7539267015706805,
+ "grad_norm": 0.03399677947163582,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.019026756286621,
+ "step": 198
+ },
+ {
+ "epoch": 2.7678883071553226,
+ "grad_norm": 0.031948331743478775,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.050133228302002,
+ "step": 199
+ },
+ {
+ "epoch": 2.781849912739965,
+ "grad_norm": 0.049206435680389404,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.045332431793213,
+ "step": 200
+ },
+ {
+ "epoch": 2.7958115183246073,
+ "grad_norm": 0.07528849691152573,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.059202194213867,
+ "step": 201
+ },
+ {
+ "epoch": 2.80977312390925,
+ "grad_norm": 0.11815470457077026,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.077486038208008,
+ "step": 202
+ },
+ {
+ "epoch": 2.823734729493892,
+ "grad_norm": 0.0867348462343216,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.045670032501221,
+ "step": 203
+ },
+ {
+ "epoch": 2.837696335078534,
+ "grad_norm": 0.05716705322265625,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.034373760223389,
+ "step": 204
+ },
+ {
+ "epoch": 2.851657940663176,
+ "grad_norm": 0.05944284051656723,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.049911975860596,
+ "step": 205
+ },
+ {
+ "epoch": 2.8656195462478182,
+ "grad_norm": 0.03220470994710922,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.039861679077148,
+ "step": 206
+ },
+ {
+ "epoch": 2.8795811518324608,
+ "grad_norm": 0.057081177830696106,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.010714530944824,
+ "step": 207
+ },
+ {
+ "epoch": 2.893542757417103,
+ "grad_norm": 0.04037516191601753,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.018877983093262,
+ "step": 208
+ },
+ {
+ "epoch": 2.9075043630017454,
+ "grad_norm": 0.036492880433797836,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.009493827819824,
+ "step": 209
+ },
+ {
+ "epoch": 2.9214659685863875,
+ "grad_norm": 0.04785023257136345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.0140862464904785,
+ "step": 210
+ },
+ {
+ "epoch": 2.9354275741710296,
+ "grad_norm": 0.07721036672592163,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.030359268188477,
+ "step": 211
+ },
+ {
+ "epoch": 2.9493891797556717,
+ "grad_norm": 0.14220190048217773,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.045989990234375,
+ "step": 212
+ },
+ {
+ "epoch": 2.9633507853403143,
+ "grad_norm": 0.10265158116817474,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.043968200683594,
+ "step": 213
+ },
+ {
+ "epoch": 2.9773123909249564,
+ "grad_norm": 0.051398951560258865,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.022278785705566,
+ "step": 214
+ },
+ {
+ "epoch": 2.9912739965095985,
+ "grad_norm": 0.06044618785381317,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.017296314239502,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.03284750506281853,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 3.7544798851013184,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7531244158744812,
+ "eval_runtime": 60.3986,
+ "eval_samples_per_second": 40.431,
+ "eval_steps_per_second": 0.646,
+ "step": 216
+ },
+ {
+ "epoch": 3.013961605584642,
+ "grad_norm": 0.03895840048789978,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 5.9985857009887695,
+ "step": 217
+ },
+ {
+ "epoch": 3.0279232111692846,
+ "grad_norm": 0.04554992541670799,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 5.984710693359375,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418848167539267,
+ "grad_norm": 0.03359116613864899,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 5.984078407287598,
+ "step": 219
+ },
+ {
+ "epoch": 3.055846422338569,
+ "grad_norm": 0.035171881318092346,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 5.979424476623535,
+ "step": 220
+ },
+ {
+ "epoch": 3.0698080279232114,
+ "grad_norm": 0.032774072140455246,
+ "learning_rate": 0.000598278794687186,
+ "loss": 5.94467830657959,
+ "step": 221
+ },
+ {
+ "epoch": 3.0837696335078535,
+ "grad_norm": 0.032842908054590225,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 5.95986270904541,
+ "step": 222
+ },
+ {
+ "epoch": 3.0977312390924956,
+ "grad_norm": 0.02870488539338112,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 5.951186656951904,
+ "step": 223
+ },
+ {
+ "epoch": 3.1116928446771377,
+ "grad_norm": 0.030777374282479286,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 5.942493438720703,
+ "step": 224
+ },
+ {
+ "epoch": 3.1256544502617802,
+ "grad_norm": 0.026439018547534943,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 5.938275337219238,
+ "step": 225
+ },
+ {
+ "epoch": 3.1396160558464223,
+ "grad_norm": 0.028689472004771233,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 5.931982040405273,
+ "step": 226
+ },
+ {
+ "epoch": 3.1535776614310644,
+ "grad_norm": 0.028100362047553062,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 5.919692039489746,
+ "step": 227
+ },
+ {
+ "epoch": 3.167539267015707,
+ "grad_norm": 0.02407177910208702,
+ "learning_rate": 0.00059807235435179,
+ "loss": 5.909029006958008,
+ "step": 228
+ },
+ {
+ "epoch": 3.181500872600349,
+ "grad_norm": 0.0237748846411705,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 5.9131598472595215,
+ "step": 229
+ },
+ {
+ "epoch": 3.195462478184991,
+ "grad_norm": 0.022418556734919548,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 5.909396171569824,
+ "step": 230
+ },
+ {
+ "epoch": 3.2094240837696333,
+ "grad_norm": 0.024656571447849274,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 5.885243892669678,
+ "step": 231
+ },
+ {
+ "epoch": 3.223385689354276,
+ "grad_norm": 0.02136118710041046,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 5.898958206176758,
+ "step": 232
+ },
+ {
+ "epoch": 3.237347294938918,
+ "grad_norm": 0.025290116667747498,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 5.886986255645752,
+ "step": 233
+ },
+ {
+ "epoch": 3.25130890052356,
+ "grad_norm": 0.02656404674053192,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 5.897647857666016,
+ "step": 234
+ },
+ {
+ "epoch": 3.2652705061082026,
+ "grad_norm": 0.019188793376088142,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 5.875659942626953,
+ "step": 235
+ },
+ {
+ "epoch": 3.2792321116928447,
+ "grad_norm": 0.02308926172554493,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 5.859380722045898,
+ "step": 236
+ },
+ {
+ "epoch": 3.2931937172774868,
+ "grad_norm": 0.029775725677609444,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 5.873961448669434,
+ "step": 237
+ },
+ {
+ "epoch": 3.3071553228621293,
+ "grad_norm": 0.03875259310007095,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 5.863171100616455,
+ "step": 238
+ },
+ {
+ "epoch": 3.3211169284467714,
+ "grad_norm": 0.052446842193603516,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 5.886547088623047,
+ "step": 239
+ },
+ {
+ "epoch": 3.3350785340314135,
+ "grad_norm": 0.05583837628364563,
+ "learning_rate": 0.000597691359743243,
+ "loss": 5.8849287033081055,
+ "step": 240
+ },
+ {
+ "epoch": 3.349040139616056,
+ "grad_norm": 0.034729067236185074,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 5.863419532775879,
+ "step": 241
+ },
+ {
+ "epoch": 3.363001745200698,
+ "grad_norm": 0.03590013459324837,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 5.852169990539551,
+ "step": 242
+ },
+ {
+ "epoch": 3.3769633507853403,
+ "grad_norm": 0.035041507333517075,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 5.836690902709961,
+ "step": 243
+ },
+ {
+ "epoch": 3.3909249563699824,
+ "grad_norm": 0.02857939340174198,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 5.836485862731934,
+ "step": 244
+ },
+ {
+ "epoch": 3.404886561954625,
+ "grad_norm": 0.03176642581820488,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 5.860719203948975,
+ "step": 245
+ },
+ {
+ "epoch": 3.418848167539267,
+ "grad_norm": 0.026105748489499092,
+ "learning_rate": 0.000597488038559129,
+ "loss": 5.8354949951171875,
+ "step": 246
+ },
+ {
+ "epoch": 3.432809773123909,
+ "grad_norm": 0.029094981029629707,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 5.846247673034668,
+ "step": 247
+ },
+ {
+ "epoch": 3.4467713787085517,
+ "grad_norm": 0.02388785220682621,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 5.837607383728027,
+ "step": 248
+ },
+ {
+ "epoch": 3.4607329842931938,
+ "grad_norm": 0.03094472736120224,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 5.832090377807617,
+ "step": 249
+ },
+ {
+ "epoch": 3.474694589877836,
+ "grad_norm": 0.03227658197283745,
+ "learning_rate": 0.000597347745466109,
+ "loss": 5.83722448348999,
+ "step": 250
+ },
+ {
+ "epoch": 3.488656195462478,
+ "grad_norm": 0.042813923209905624,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 5.851722717285156,
+ "step": 251
+ },
+ {
+ "epoch": 3.5026178010471205,
+ "grad_norm": 0.05825525149703026,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.795719146728516,
+ "step": 252
+ },
+ {
+ "epoch": 3.5165794066317626,
+ "grad_norm": 0.0867447778582573,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.813750267028809,
+ "step": 253
+ },
+ {
+ "epoch": 3.530541012216405,
+ "grad_norm": 0.09192710369825363,
+ "learning_rate": 0.000597203657966319,
+ "loss": 5.823488235473633,
+ "step": 254
+ },
+ {
+ "epoch": 3.5445026178010473,
+ "grad_norm": 0.06069248169660568,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 5.814342021942139,
+ "step": 255
+ },
+ {
+ "epoch": 3.5584642233856894,
+ "grad_norm": 0.06259952485561371,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 5.834272384643555,
+ "step": 256
+ },
+ {
+ "epoch": 3.5724258289703315,
+ "grad_norm": 0.0513639897108078,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.807951927185059,
+ "step": 257
+ },
+ {
+ "epoch": 3.5863874345549736,
+ "grad_norm": 0.03327798470854759,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.790238857269287,
+ "step": 258
+ },
+ {
+ "epoch": 3.600349040139616,
+ "grad_norm": 0.03963128477334976,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.7956695556640625,
+ "step": 259
+ },
+ {
+ "epoch": 3.614310645724258,
+ "grad_norm": 0.031153453513979912,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 5.7874531745910645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6282722513089007,
+ "grad_norm": 0.031419433653354645,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 5.779650688171387,
+ "step": 261
+ },
+ {
+ "epoch": 3.642233856893543,
+ "grad_norm": 0.0288684144616127,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.776577949523926,
+ "step": 262
+ },
+ {
+ "epoch": 3.656195462478185,
+ "grad_norm": 0.028315918520092964,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.778571128845215,
+ "step": 263
+ },
+ {
+ "epoch": 3.670157068062827,
+ "grad_norm": 0.024977795779705048,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 5.776470184326172,
+ "step": 264
+ },
+ {
+ "epoch": 3.6841186736474696,
+ "grad_norm": 0.026822632178664207,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.755108833312988,
+ "step": 265
+ },
+ {
+ "epoch": 3.6980802792321117,
+ "grad_norm": 0.02738630585372448,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.74041748046875,
+ "step": 266
+ },
+ {
+ "epoch": 3.712041884816754,
+ "grad_norm": 0.02361280657351017,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.764027118682861,
+ "step": 267
+ },
+ {
+ "epoch": 3.7260034904013963,
+ "grad_norm": 0.029926009476184845,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.780390739440918,
+ "step": 268
+ },
+ {
+ "epoch": 3.7399650959860384,
+ "grad_norm": 0.03519507870078087,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.758105278015137,
+ "step": 269
+ },
+ {
+ "epoch": 3.7539267015706805,
+ "grad_norm": 0.030114633962512016,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.757230758666992,
+ "step": 270
+ },
+ {
+ "epoch": 3.7678883071553226,
+ "grad_norm": 0.02467748522758484,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.7615461349487305,
+ "step": 271
+ },
+ {
+ "epoch": 3.781849912739965,
+ "grad_norm": 0.03425183147192001,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.749448776245117,
+ "step": 272
+ },
+ {
+ "epoch": 3.7958115183246073,
+ "grad_norm": 0.046307582408189774,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.7356977462768555,
+ "step": 273
+ },
+ {
+ "epoch": 3.80977312390925,
+ "grad_norm": 0.054102856665849686,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.741869926452637,
+ "step": 274
+ },
+ {
+ "epoch": 3.823734729493892,
+ "grad_norm": 0.07026689499616623,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.721103668212891,
+ "step": 275
+ },
+ {
+ "epoch": 3.837696335078534,
+ "grad_norm": 0.0733039602637291,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.731414794921875,
+ "step": 276
+ },
+ {
+ "epoch": 3.851657940663176,
+ "grad_norm": 0.05342920497059822,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.734922885894775,
+ "step": 277
+ },
+ {
+ "epoch": 3.8656195462478182,
+ "grad_norm": 0.0504680834710598,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.732746124267578,
+ "step": 278
+ },
+ {
+ "epoch": 3.8795811518324608,
+ "grad_norm": 0.038991592824459076,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.719473838806152,
+ "step": 279
+ },
+ {
+ "epoch": 3.893542757417103,
+ "grad_norm": 0.04434404894709587,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.709481239318848,
+ "step": 280
+ },
+ {
+ "epoch": 3.9075043630017454,
+ "grad_norm": 0.047285184264183044,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.693531036376953,
+ "step": 281
+ },
+ {
+ "epoch": 3.9214659685863875,
+ "grad_norm": 0.045081138610839844,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.728959083557129,
+ "step": 282
+ },
+ {
+ "epoch": 3.9354275741710296,
+ "grad_norm": 0.025133663788437843,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.719696998596191,
+ "step": 283
+ },
+ {
+ "epoch": 3.9493891797556717,
+ "grad_norm": 0.03353596478700638,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.700235366821289,
+ "step": 284
+ },
+ {
+ "epoch": 3.9633507853403143,
+ "grad_norm": 0.035162411630153656,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.684836387634277,
+ "step": 285
+ },
+ {
+ "epoch": 3.9773123909249564,
+ "grad_norm": 0.04025450348854065,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.692097187042236,
+ "step": 286
+ },
+ {
+ "epoch": 3.9912739965095985,
+ "grad_norm": 0.04995397478342056,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.6766862869262695,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.031619708985090256,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 3.5519747734069824,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.714687705039978,
+ "eval_runtime": 59.637,
+ "eval_samples_per_second": 40.948,
+ "eval_steps_per_second": 0.654,
+ "step": 288
+ },
+ {
+ "epoch": 4.013961605584642,
+ "grad_norm": 0.036328598856925964,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.666294097900391,
+ "step": 289
+ },
+ {
+ "epoch": 4.027923211169284,
+ "grad_norm": 0.07326336950063705,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.6879167556762695,
+ "step": 290
+ },
+ {
+ "epoch": 4.041884816753926,
+ "grad_norm": 0.07947526127099991,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.674883842468262,
+ "step": 291
+ },
+ {
+ "epoch": 4.055846422338569,
+ "grad_norm": 0.04599885269999504,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.688086986541748,
+ "step": 292
+ },
+ {
+ "epoch": 4.069808027923211,
+ "grad_norm": 0.0465150885283947,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.649483680725098,
+ "step": 293
+ },
+ {
+ "epoch": 4.0837696335078535,
+ "grad_norm": 0.038542453199625015,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.653055667877197,
+ "step": 294
+ },
+ {
+ "epoch": 4.097731239092496,
+ "grad_norm": 0.03659164905548096,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.65955924987793,
+ "step": 295
+ },
+ {
+ "epoch": 4.111692844677138,
+ "grad_norm": 0.03824038803577423,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.650853157043457,
+ "step": 296
+ },
+ {
+ "epoch": 4.12565445026178,
+ "grad_norm": 0.03591204807162285,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.643499374389648,
+ "step": 297
+ },
+ {
+ "epoch": 4.139616055846423,
+ "grad_norm": 0.03269682824611664,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.657600402832031,
+ "step": 298
+ },
+ {
+ "epoch": 4.153577661431065,
+ "grad_norm": 0.03343537449836731,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.6352081298828125,
+ "step": 299
+ },
+ {
+ "epoch": 4.167539267015707,
+ "grad_norm": 0.034283801913261414,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.626757621765137,
+ "step": 300
+ },
+ {
+ "epoch": 4.181500872600349,
+ "grad_norm": 0.034614723175764084,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.6316680908203125,
+ "step": 301
+ },
+ {
+ "epoch": 4.195462478184991,
+ "grad_norm": 0.0340128131210804,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.648160934448242,
+ "step": 302
+ },
+ {
+ "epoch": 4.209424083769633,
+ "grad_norm": 0.03068280965089798,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.619738578796387,
+ "step": 303
+ },
+ {
+ "epoch": 4.223385689354275,
+ "grad_norm": 0.030269058421254158,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.609238624572754,
+ "step": 304
+ },
+ {
+ "epoch": 4.237347294938918,
+ "grad_norm": 0.03612440824508667,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.602194786071777,
+ "step": 305
+ },
+ {
+ "epoch": 4.2513089005235605,
+ "grad_norm": 0.044292811304330826,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.61159610748291,
+ "step": 306
+ },
+ {
+ "epoch": 4.265270506108203,
+ "grad_norm": 0.04449291527271271,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.6274237632751465,
+ "step": 307
+ },
+ {
+ "epoch": 4.279232111692845,
+ "grad_norm": 0.05014493316411972,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.589034080505371,
+ "step": 308
+ },
+ {
+ "epoch": 4.293193717277487,
+ "grad_norm": 0.05286206305027008,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.600770950317383,
+ "step": 309
+ },
+ {
+ "epoch": 4.307155322862129,
+ "grad_norm": 0.048459846526384354,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.5796098709106445,
+ "step": 310
+ },
+ {
+ "epoch": 4.321116928446771,
+ "grad_norm": 0.05329663306474686,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.590407371520996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335078534031414,
+ "grad_norm": 0.052595119923353195,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.583693504333496,
+ "step": 312
+ },
+ {
+ "epoch": 4.349040139616056,
+ "grad_norm": 0.05164601653814316,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.603281021118164,
+ "step": 313
+ },
+ {
+ "epoch": 4.363001745200698,
+ "grad_norm": 0.04657137766480446,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.591344833374023,
+ "step": 314
+ },
+ {
+ "epoch": 4.37696335078534,
+ "grad_norm": 0.03804599866271019,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.580578804016113,
+ "step": 315
+ },
+ {
+ "epoch": 4.390924956369982,
+ "grad_norm": 0.03835320845246315,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.597836971282959,
+ "step": 316
+ },
+ {
+ "epoch": 4.4048865619546245,
+ "grad_norm": 0.034059133380651474,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.567502021789551,
+ "step": 317
+ },
+ {
+ "epoch": 4.418848167539267,
+ "grad_norm": 0.03350386768579483,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.552022933959961,
+ "step": 318
+ },
+ {
+ "epoch": 4.43280977312391,
+ "grad_norm": 0.032574303448200226,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.559177875518799,
+ "step": 319
+ },
+ {
+ "epoch": 4.446771378708552,
+ "grad_norm": 0.030893854796886444,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.549288272857666,
+ "step": 320
+ },
+ {
+ "epoch": 4.460732984293194,
+ "grad_norm": 0.03255036100745201,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.5400390625,
+ "step": 321
+ },
+ {
+ "epoch": 4.474694589877836,
+ "grad_norm": 0.03009258769452572,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.5328474044799805,
+ "step": 322
+ },
+ {
+ "epoch": 4.488656195462478,
+ "grad_norm": 0.03178320452570915,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.5594482421875,
+ "step": 323
+ },
+ {
+ "epoch": 4.50261780104712,
+ "grad_norm": 0.03360465541481972,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.5402727127075195,
+ "step": 324
+ },
+ {
+ "epoch": 4.516579406631763,
+ "grad_norm": 0.04275049269199371,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.534474849700928,
+ "step": 325
+ },
+ {
+ "epoch": 4.530541012216405,
+ "grad_norm": 0.04140641540288925,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.5324859619140625,
+ "step": 326
+ },
+ {
+ "epoch": 4.544502617801047,
+ "grad_norm": 0.04769204929471016,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.4919352531433105,
+ "step": 327
+ },
+ {
+ "epoch": 4.558464223385689,
+ "grad_norm": 0.04402285814285278,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.520858287811279,
+ "step": 328
+ },
+ {
+ "epoch": 4.5724258289703315,
+ "grad_norm": 0.038526229560375214,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.523778438568115,
+ "step": 329
+ },
+ {
+ "epoch": 4.5863874345549736,
+ "grad_norm": 0.05422377213835716,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.517973899841309,
+ "step": 330
+ },
+ {
+ "epoch": 4.600349040139616,
+ "grad_norm": 0.07236721366643906,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.548083782196045,
+ "step": 331
+ },
+ {
+ "epoch": 4.614310645724259,
+ "grad_norm": 0.06861061602830887,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.535577774047852,
+ "step": 332
+ },
+ {
+ "epoch": 4.628272251308901,
+ "grad_norm": 0.07148340344429016,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.54155158996582,
+ "step": 333
+ },
+ {
+ "epoch": 4.642233856893543,
+ "grad_norm": 0.07838302850723267,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.552852630615234,
+ "step": 334
+ },
+ {
+ "epoch": 4.656195462478185,
+ "grad_norm": 0.10946758091449738,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.555066108703613,
+ "step": 335
+ },
+ {
+ "epoch": 4.670157068062827,
+ "grad_norm": 0.10143530368804932,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.584932804107666,
+ "step": 336
+ },
+ {
+ "epoch": 4.684118673647469,
+ "grad_norm": 0.08926606178283691,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.570462226867676,
+ "step": 337
+ },
+ {
+ "epoch": 4.698080279232112,
+ "grad_norm": 0.07060826569795609,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.5716753005981445,
+ "step": 338
+ },
+ {
+ "epoch": 4.712041884816754,
+ "grad_norm": 0.0418456569314003,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.556496620178223,
+ "step": 339
+ },
+ {
+ "epoch": 4.726003490401396,
+ "grad_norm": 0.05952411890029907,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.567801475524902,
+ "step": 340
+ },
+ {
+ "epoch": 4.739965095986038,
+ "grad_norm": 0.042999885976314545,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.537403106689453,
+ "step": 341
+ },
+ {
+ "epoch": 4.7539267015706805,
+ "grad_norm": 0.04426449537277222,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.552387237548828,
+ "step": 342
+ },
+ {
+ "epoch": 4.767888307155323,
+ "grad_norm": 0.040835145860910416,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.53051233291626,
+ "step": 343
+ },
+ {
+ "epoch": 4.781849912739965,
+ "grad_norm": 0.03819291293621063,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.531174659729004,
+ "step": 344
+ },
+ {
+ "epoch": 4.795811518324607,
+ "grad_norm": 0.030088599771261215,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.499328136444092,
+ "step": 345
+ },
+ {
+ "epoch": 4.80977312390925,
+ "grad_norm": 0.029984362423419952,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.504540920257568,
+ "step": 346
+ },
+ {
+ "epoch": 4.823734729493892,
+ "grad_norm": 0.027402915060520172,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.50899076461792,
+ "step": 347
+ },
+ {
+ "epoch": 4.837696335078534,
+ "grad_norm": 0.025096213445067406,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.480479717254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.851657940663176,
+ "grad_norm": 0.026828132569789886,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.501729965209961,
+ "step": 349
+ },
+ {
+ "epoch": 4.865619546247818,
+ "grad_norm": 0.023715168237686157,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.466860771179199,
+ "step": 350
+ },
+ {
+ "epoch": 4.879581151832461,
+ "grad_norm": 0.024503640830516815,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.491827964782715,
+ "step": 351
+ },
+ {
+ "epoch": 4.893542757417103,
+ "grad_norm": 0.020891105756163597,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.478384971618652,
+ "step": 352
+ },
+ {
+ "epoch": 4.907504363001745,
+ "grad_norm": 0.021373765543103218,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.4612135887146,
+ "step": 353
+ },
+ {
+ "epoch": 4.9214659685863875,
+ "grad_norm": 0.021773340180516243,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.4547319412231445,
+ "step": 354
+ },
+ {
+ "epoch": 4.93542757417103,
+ "grad_norm": 0.019637586548924446,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.431741714477539,
+ "step": 355
+ },
+ {
+ "epoch": 4.949389179755672,
+ "grad_norm": 0.023108696565032005,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.448612213134766,
+ "step": 356
+ },
+ {
+ "epoch": 4.963350785340314,
+ "grad_norm": 0.021935103461146355,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.447112083435059,
+ "step": 357
+ },
+ {
+ "epoch": 4.977312390924956,
+ "grad_norm": 0.023260753601789474,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.436091423034668,
+ "step": 358
+ },
+ {
+ "epoch": 4.991273996509599,
+ "grad_norm": 0.021355057135224342,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.408421993255615,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.014992325566709042,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 3.4004223346710205,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6834986805915833,
+ "eval_runtime": 60.4148,
+ "eval_samples_per_second": 40.421,
+ "eval_steps_per_second": 0.646,
+ "step": 360
+ },
+ {
+ "epoch": 5.013961605584642,
+ "grad_norm": 0.02471218630671501,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.396238803863525,
+ "step": 361
+ },
+ {
+ "epoch": 5.027923211169284,
+ "grad_norm": 0.027672411873936653,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.4048919677734375,
+ "step": 362
+ },
+ {
+ "epoch": 5.041884816753926,
+ "grad_norm": 0.031560689210891724,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.399665832519531,
+ "step": 363
+ },
+ {
+ "epoch": 5.055846422338569,
+ "grad_norm": 0.03459061309695244,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.396360397338867,
+ "step": 364
+ },
+ {
+ "epoch": 5.069808027923211,
+ "grad_norm": 0.04040960222482681,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.405309677124023,
+ "step": 365
+ },
+ {
+ "epoch": 5.0837696335078535,
+ "grad_norm": 0.05444670096039772,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.408894062042236,
+ "step": 366
+ },
+ {
+ "epoch": 5.097731239092496,
+ "grad_norm": 0.05816745012998581,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.403566360473633,
+ "step": 367
+ },
+ {
+ "epoch": 5.111692844677138,
+ "grad_norm": 0.06330069154500961,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.395961761474609,
+ "step": 368
+ },
+ {
+ "epoch": 5.12565445026178,
+ "grad_norm": 0.07458917796611786,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.400968551635742,
+ "step": 369
+ },
+ {
+ "epoch": 5.139616055846423,
+ "grad_norm": 0.06231356039643288,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.412624359130859,
+ "step": 370
+ },
+ {
+ "epoch": 5.153577661431065,
+ "grad_norm": 0.05212346464395523,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.394794464111328,
+ "step": 371
+ },
+ {
+ "epoch": 5.167539267015707,
+ "grad_norm": 0.050003375858068466,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.380125045776367,
+ "step": 372
+ },
+ {
+ "epoch": 5.181500872600349,
+ "grad_norm": 0.048393815755844116,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.390806198120117,
+ "step": 373
+ },
+ {
+ "epoch": 5.195462478184991,
+ "grad_norm": 0.04200794920325279,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.387467384338379,
+ "step": 374
+ },
+ {
+ "epoch": 5.209424083769633,
+ "grad_norm": 0.04931448772549629,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.37552547454834,
+ "step": 375
+ },
+ {
+ "epoch": 5.223385689354275,
+ "grad_norm": 0.03851857781410217,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.373126983642578,
+ "step": 376
+ },
+ {
+ "epoch": 5.237347294938918,
+ "grad_norm": 0.03458717465400696,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.346826553344727,
+ "step": 377
+ },
+ {
+ "epoch": 5.2513089005235605,
+ "grad_norm": 0.031107677146792412,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.353341102600098,
+ "step": 378
+ },
+ {
+ "epoch": 5.265270506108203,
+ "grad_norm": 0.03077058307826519,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.358076095581055,
+ "step": 379
+ },
+ {
+ "epoch": 5.279232111692845,
+ "grad_norm": 0.028170401230454445,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.354019641876221,
+ "step": 380
+ },
+ {
+ "epoch": 5.293193717277487,
+ "grad_norm": 0.02871990203857422,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.368258476257324,
+ "step": 381
+ },
+ {
+ "epoch": 5.307155322862129,
+ "grad_norm": 0.03046584501862526,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.342505931854248,
+ "step": 382
+ },
+ {
+ "epoch": 5.321116928446771,
+ "grad_norm": 0.02781563811004162,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.339411735534668,
+ "step": 383
+ },
+ {
+ "epoch": 5.335078534031414,
+ "grad_norm": 0.028825288638472557,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.341948509216309,
+ "step": 384
+ },
+ {
+ "epoch": 5.349040139616056,
+ "grad_norm": 0.026684079319238663,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.340881824493408,
+ "step": 385
+ },
+ {
+ "epoch": 5.363001745200698,
+ "grad_norm": 0.02429167740046978,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.342733383178711,
+ "step": 386
+ },
+ {
+ "epoch": 5.37696335078534,
+ "grad_norm": 0.02855239436030388,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.313878536224365,
+ "step": 387
+ },
+ {
+ "epoch": 5.390924956369982,
+ "grad_norm": 0.029651548713445663,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.312665939331055,
+ "step": 388
+ },
+ {
+ "epoch": 5.4048865619546245,
+ "grad_norm": 0.027191132307052612,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3086838722229,
+ "step": 389
+ },
+ {
+ "epoch": 5.418848167539267,
+ "grad_norm": 0.026719525456428528,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.305359840393066,
+ "step": 390
+ },
+ {
+ "epoch": 5.43280977312391,
+ "grad_norm": 0.026763100177049637,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.3164381980896,
+ "step": 391
+ },
+ {
+ "epoch": 5.446771378708552,
+ "grad_norm": 0.03125109151005745,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.3083953857421875,
+ "step": 392
+ },
+ {
+ "epoch": 5.460732984293194,
+ "grad_norm": 0.0394827201962471,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.275996208190918,
+ "step": 393
+ },
+ {
+ "epoch": 5.474694589877836,
+ "grad_norm": 0.040579475462436676,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.292995929718018,
+ "step": 394
+ },
+ {
+ "epoch": 5.488656195462478,
+ "grad_norm": 0.037223003804683685,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.298526763916016,
+ "step": 395
+ },
+ {
+ "epoch": 5.50261780104712,
+ "grad_norm": 0.04207765311002731,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.287985801696777,
+ "step": 396
+ },
+ {
+ "epoch": 5.516579406631763,
+ "grad_norm": 0.048299964517354965,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.244608402252197,
+ "step": 397
+ },
+ {
+ "epoch": 5.530541012216405,
+ "grad_norm": 0.04650307446718216,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.285453796386719,
+ "step": 398
+ },
+ {
+ "epoch": 5.544502617801047,
+ "grad_norm": 0.045339711010456085,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.312187671661377,
+ "step": 399
+ },
+ {
+ "epoch": 5.558464223385689,
+ "grad_norm": 0.041067708283662796,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.292692184448242,
+ "step": 400
+ },
+ {
+ "epoch": 5.5724258289703315,
+ "grad_norm": 0.05455026403069496,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.253364562988281,
+ "step": 401
+ },
+ {
+ "epoch": 5.5863874345549736,
+ "grad_norm": 0.06660915911197662,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.262646675109863,
+ "step": 402
+ },
+ {
+ "epoch": 5.600349040139616,
+ "grad_norm": 0.06634793430566788,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.274322986602783,
+ "step": 403
+ },
+ {
+ "epoch": 5.614310645724259,
+ "grad_norm": 0.08996330946683884,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.287951469421387,
+ "step": 404
+ },
+ {
+ "epoch": 5.628272251308901,
+ "grad_norm": 0.08526470512151718,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.3007097244262695,
+ "step": 405
+ },
+ {
+ "epoch": 5.642233856893543,
+ "grad_norm": 0.08766836673021317,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.332538604736328,
+ "step": 406
+ },
+ {
+ "epoch": 5.656195462478185,
+ "grad_norm": 0.07055048644542694,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.296169281005859,
+ "step": 407
+ },
+ {
+ "epoch": 5.670157068062827,
+ "grad_norm": 0.07520575076341629,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.309683322906494,
+ "step": 408
+ },
+ {
+ "epoch": 5.684118673647469,
+ "grad_norm": 0.06115756183862686,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.311301231384277,
+ "step": 409
+ },
+ {
+ "epoch": 5.698080279232112,
+ "grad_norm": 0.05862593650817871,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.28066349029541,
+ "step": 410
+ },
+ {
+ "epoch": 5.712041884816754,
+ "grad_norm": 0.06169601157307625,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.286840438842773,
+ "step": 411
+ },
+ {
+ "epoch": 5.726003490401396,
+ "grad_norm": 0.05270719155669212,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.302594184875488,
+ "step": 412
+ },
+ {
+ "epoch": 5.739965095986038,
+ "grad_norm": 0.03912634402513504,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.262563705444336,
+ "step": 413
+ },
+ {
+ "epoch": 5.7539267015706805,
+ "grad_norm": 0.03215811401605606,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.269310474395752,
+ "step": 414
+ },
+ {
+ "epoch": 5.767888307155323,
+ "grad_norm": 0.03236516937613487,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.268766403198242,
+ "step": 415
+ },
+ {
+ "epoch": 5.781849912739965,
+ "grad_norm": 0.027429452165961266,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.283801555633545,
+ "step": 416
+ },
+ {
+ "epoch": 5.795811518324607,
+ "grad_norm": 0.02990538813173771,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.22291374206543,
+ "step": 417
+ },
+ {
+ "epoch": 5.80977312390925,
+ "grad_norm": 0.030410146340727806,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.2281599044799805,
+ "step": 418
+ },
+ {
+ "epoch": 5.823734729493892,
+ "grad_norm": 0.030068187043070793,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.250294208526611,
+ "step": 419
+ },
+ {
+ "epoch": 5.837696335078534,
+ "grad_norm": 0.02800513058900833,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.226508617401123,
+ "step": 420
+ },
+ {
+ "epoch": 5.851657940663176,
+ "grad_norm": 0.023823332041502,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.247838020324707,
+ "step": 421
+ },
+ {
+ "epoch": 5.865619546247818,
+ "grad_norm": 0.024974249303340912,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.2392473220825195,
+ "step": 422
+ },
+ {
+ "epoch": 5.879581151832461,
+ "grad_norm": 0.023320546373724937,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.209543704986572,
+ "step": 423
+ },
+ {
+ "epoch": 5.893542757417103,
+ "grad_norm": 0.024201491847634315,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.219725608825684,
+ "step": 424
+ },
+ {
+ "epoch": 5.907504363001745,
+ "grad_norm": 0.021649666130542755,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.203141212463379,
+ "step": 425
+ },
+ {
+ "epoch": 5.9214659685863875,
+ "grad_norm": 0.0248002577573061,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.179058074951172,
+ "step": 426
+ },
+ {
+ "epoch": 5.93542757417103,
+ "grad_norm": 0.028549587354063988,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.202126502990723,
+ "step": 427
+ },
+ {
+ "epoch": 5.949389179755672,
+ "grad_norm": 0.0331057533621788,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.208107948303223,
+ "step": 428
+ },
+ {
+ "epoch": 5.963350785340314,
+ "grad_norm": 0.031843364238739014,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.184683799743652,
+ "step": 429
+ },
+ {
+ "epoch": 5.977312390924956,
+ "grad_norm": 0.024927377700805664,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.180486679077148,
+ "step": 430
+ },
+ {
+ "epoch": 5.991273996509599,
+ "grad_norm": 0.026685267686843872,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.186746597290039,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.018545351922512054,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 3.217043876647949,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.653471827507019,
+ "eval_runtime": 61.1586,
+ "eval_samples_per_second": 39.929,
+ "eval_steps_per_second": 0.638,
+ "step": 432
+ },
+ {
+ "epoch": 6.013961605584642,
+ "grad_norm": 0.025291118770837784,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.156534194946289,
+ "step": 433
+ },
+ {
+ "epoch": 6.027923211169284,
+ "grad_norm": 0.03774752840399742,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.144041538238525,
+ "step": 434
+ },
+ {
+ "epoch": 6.041884816753926,
+ "grad_norm": 0.04244692996144295,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.158862113952637,
+ "step": 435
+ },
+ {
+ "epoch": 6.055846422338569,
+ "grad_norm": 0.039782389998435974,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.166165351867676,
+ "step": 436
+ },
+ {
+ "epoch": 6.069808027923211,
+ "grad_norm": 0.05489896610379219,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.155766487121582,
+ "step": 437
+ },
+ {
+ "epoch": 6.0837696335078535,
+ "grad_norm": 0.059844326227903366,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.153733253479004,
+ "step": 438
+ },
+ {
+ "epoch": 6.097731239092496,
+ "grad_norm": 0.05539938807487488,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.130112648010254,
+ "step": 439
+ },
+ {
+ "epoch": 6.111692844677138,
+ "grad_norm": 0.05283183231949806,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.136910915374756,
+ "step": 440
+ },
+ {
+ "epoch": 6.12565445026178,
+ "grad_norm": 0.054407261312007904,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.165216445922852,
+ "step": 441
+ },
+ {
+ "epoch": 6.139616055846423,
+ "grad_norm": 0.0509788915514946,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.140888214111328,
+ "step": 442
+ },
+ {
+ "epoch": 6.153577661431065,
+ "grad_norm": 0.05018748342990875,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.13115930557251,
+ "step": 443
+ },
+ {
+ "epoch": 6.167539267015707,
+ "grad_norm": 0.047012437134981155,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.146507263183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181500872600349,
+ "grad_norm": 0.04943335801362991,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.135608673095703,
+ "step": 445
+ },
+ {
+ "epoch": 6.195462478184991,
+ "grad_norm": 0.05304260924458504,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.136394023895264,
+ "step": 446
+ },
+ {
+ "epoch": 6.209424083769633,
+ "grad_norm": 0.04554047808051109,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.124661922454834,
+ "step": 447
+ },
+ {
+ "epoch": 6.223385689354275,
+ "grad_norm": 0.040672432631254196,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.129773139953613,
+ "step": 448
+ },
+ {
+ "epoch": 6.237347294938918,
+ "grad_norm": 0.03442879393696785,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.128028392791748,
+ "step": 449
+ },
+ {
+ "epoch": 6.2513089005235605,
+ "grad_norm": 0.03927934914827347,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.092485427856445,
+ "step": 450
+ },
+ {
+ "epoch": 6.265270506108203,
+ "grad_norm": 0.029380977153778076,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.077669143676758,
+ "step": 451
+ },
+ {
+ "epoch": 6.279232111692845,
+ "grad_norm": 0.03331821784377098,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.138857841491699,
+ "step": 452
+ },
+ {
+ "epoch": 6.293193717277487,
+ "grad_norm": 0.02939118817448616,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.082937240600586,
+ "step": 453
+ },
+ {
+ "epoch": 6.307155322862129,
+ "grad_norm": 0.02730412967503071,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.093183517456055,
+ "step": 454
+ },
+ {
+ "epoch": 6.321116928446771,
+ "grad_norm": 0.02964780665934086,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.1028151512146,
+ "step": 455
+ },
+ {
+ "epoch": 6.335078534031414,
+ "grad_norm": 0.026350129395723343,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.066362380981445,
+ "step": 456
+ },
+ {
+ "epoch": 6.349040139616056,
+ "grad_norm": 0.027010295540094376,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.070037841796875,
+ "step": 457
+ },
+ {
+ "epoch": 6.363001745200698,
+ "grad_norm": 0.02906128577888012,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.107708930969238,
+ "step": 458
+ },
+ {
+ "epoch": 6.37696335078534,
+ "grad_norm": 0.028619443997740746,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.073182582855225,
+ "step": 459
+ },
+ {
+ "epoch": 6.390924956369982,
+ "grad_norm": 0.029981834813952446,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.076602935791016,
+ "step": 460
+ },
+ {
+ "epoch": 6.4048865619546245,
+ "grad_norm": 0.03255423903465271,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.063301086425781,
+ "step": 461
+ },
+ {
+ "epoch": 6.418848167539267,
+ "grad_norm": 0.03863866627216339,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.041812896728516,
+ "step": 462
+ },
+ {
+ "epoch": 6.43280977312391,
+ "grad_norm": 0.03860306739807129,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.06298828125,
+ "step": 463
+ },
+ {
+ "epoch": 6.446771378708552,
+ "grad_norm": 0.03378435596823692,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.0434370040893555,
+ "step": 464
+ },
+ {
+ "epoch": 6.460732984293194,
+ "grad_norm": 0.03276759013533592,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.036187171936035,
+ "step": 465
+ },
+ {
+ "epoch": 6.474694589877836,
+ "grad_norm": 0.04130513221025467,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.039098739624023,
+ "step": 466
+ },
+ {
+ "epoch": 6.488656195462478,
+ "grad_norm": 0.05389940366148949,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.062798976898193,
+ "step": 467
+ },
+ {
+ "epoch": 6.50261780104712,
+ "grad_norm": 0.05881704390048981,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.067028522491455,
+ "step": 468
+ },
+ {
+ "epoch": 6.516579406631763,
+ "grad_norm": 0.061897799372673035,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.056947231292725,
+ "step": 469
+ },
+ {
+ "epoch": 6.530541012216405,
+ "grad_norm": 0.06549323350191116,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.089262008666992,
+ "step": 470
+ },
+ {
+ "epoch": 6.544502617801047,
+ "grad_norm": 0.048082590103149414,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.055979251861572,
+ "step": 471
+ },
+ {
+ "epoch": 6.558464223385689,
+ "grad_norm": 0.03768027946352959,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.0660600662231445,
+ "step": 472
+ },
+ {
+ "epoch": 6.5724258289703315,
+ "grad_norm": 0.043922603130340576,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 5.036745071411133,
+ "step": 473
+ },
+ {
+ "epoch": 6.5863874345549736,
+ "grad_norm": 0.044848836958408356,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.073650360107422,
+ "step": 474
+ },
+ {
+ "epoch": 6.600349040139616,
+ "grad_norm": 0.05841026455163956,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.053045272827148,
+ "step": 475
+ },
+ {
+ "epoch": 6.614310645724259,
+ "grad_norm": 0.0650017261505127,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 5.052206039428711,
+ "step": 476
+ },
+ {
+ "epoch": 6.628272251308901,
+ "grad_norm": 0.048979002982378006,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 5.054277420043945,
+ "step": 477
+ },
+ {
+ "epoch": 6.642233856893543,
+ "grad_norm": 0.05085799843072891,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 5.063844680786133,
+ "step": 478
+ },
+ {
+ "epoch": 6.656195462478185,
+ "grad_norm": 0.04336017370223999,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 5.045474052429199,
+ "step": 479
+ },
+ {
+ "epoch": 6.670157068062827,
+ "grad_norm": 0.04113380238413811,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 5.031338214874268,
+ "step": 480
+ },
+ {
+ "epoch": 6.684118673647469,
+ "grad_norm": 0.03199538215994835,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 5.030025959014893,
+ "step": 481
+ },
+ {
+ "epoch": 6.698080279232112,
+ "grad_norm": 0.03043113648891449,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 5.033163070678711,
+ "step": 482
+ },
+ {
+ "epoch": 6.712041884816754,
+ "grad_norm": 0.031694378703832626,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 5.004739761352539,
+ "step": 483
+ },
+ {
+ "epoch": 6.726003490401396,
+ "grad_norm": 0.030243348330259323,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 5.012566089630127,
+ "step": 484
+ },
+ {
+ "epoch": 6.739965095986038,
+ "grad_norm": 0.02937416359782219,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 5.039803504943848,
+ "step": 485
+ },
+ {
+ "epoch": 6.7539267015706805,
+ "grad_norm": 0.029616443440318108,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 5.026117324829102,
+ "step": 486
+ },
+ {
+ "epoch": 6.767888307155323,
+ "grad_norm": 0.029567131772637367,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 5.006569862365723,
+ "step": 487
+ },
+ {
+ "epoch": 6.781849912739965,
+ "grad_norm": 0.029443370178341866,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.9941487312316895,
+ "step": 488
+ },
+ {
+ "epoch": 6.795811518324607,
+ "grad_norm": 0.03434352949261665,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.976431846618652,
+ "step": 489
+ },
+ {
+ "epoch": 6.80977312390925,
+ "grad_norm": 0.03845234215259552,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.993168354034424,
+ "step": 490
+ },
+ {
+ "epoch": 6.823734729493892,
+ "grad_norm": 0.05058208107948303,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.030162334442139,
+ "step": 491
+ },
+ {
+ "epoch": 6.837696335078534,
+ "grad_norm": 0.05961710214614868,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 5.007053375244141,
+ "step": 492
+ },
+ {
+ "epoch": 6.851657940663176,
+ "grad_norm": 0.056259628385305405,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 5.020356178283691,
+ "step": 493
+ },
+ {
+ "epoch": 6.865619546247818,
+ "grad_norm": 0.048628658056259155,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 5.037652969360352,
+ "step": 494
+ },
+ {
+ "epoch": 6.879581151832461,
+ "grad_norm": 0.040836453437805176,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.990390777587891,
+ "step": 495
+ },
+ {
+ "epoch": 6.893542757417103,
+ "grad_norm": 0.04112066701054573,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 5.017822742462158,
+ "step": 496
+ },
+ {
+ "epoch": 6.907504363001745,
+ "grad_norm": 0.03956829756498337,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.993971824645996,
+ "step": 497
+ },
+ {
+ "epoch": 6.9214659685863875,
+ "grad_norm": 0.03887917473912239,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.982889175415039,
+ "step": 498
+ },
+ {
+ "epoch": 6.93542757417103,
+ "grad_norm": 0.031012587249279022,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.961717128753662,
+ "step": 499
+ },
+ {
+ "epoch": 6.949389179755672,
+ "grad_norm": 0.03443489596247673,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.9904046058654785,
+ "step": 500
+ },
+ {
+ "epoch": 6.963350785340314,
+ "grad_norm": 0.035175010561943054,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.983619213104248,
+ "step": 501
+ },
+ {
+ "epoch": 6.977312390924956,
+ "grad_norm": 0.031724050641059875,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 5.000522613525391,
+ "step": 502
+ },
+ {
+ "epoch": 6.991273996509599,
+ "grad_norm": 0.032037995755672455,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.947064399719238,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.01883271522819996,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 3.112166166305542,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6296830177307129,
+ "eval_runtime": 59.5495,
+ "eval_samples_per_second": 41.008,
+ "eval_steps_per_second": 0.655,
+ "step": 504
+ },
+ {
+ "epoch": 7.013961605584642,
+ "grad_norm": 0.031091563403606415,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.919820785522461,
+ "step": 505
+ },
+ {
+ "epoch": 7.027923211169284,
+ "grad_norm": 0.04074421525001526,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.945765972137451,
+ "step": 506
+ },
+ {
+ "epoch": 7.041884816753926,
+ "grad_norm": 0.04225052148103714,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.935208797454834,
+ "step": 507
+ },
+ {
+ "epoch": 7.055846422338569,
+ "grad_norm": 0.0399051196873188,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.951314926147461,
+ "step": 508
+ },
+ {
+ "epoch": 7.069808027923211,
+ "grad_norm": 0.04785890504717827,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.92843770980835,
+ "step": 509
+ },
+ {
+ "epoch": 7.0837696335078535,
+ "grad_norm": 0.04922301694750786,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.901020050048828,
+ "step": 510
+ },
+ {
+ "epoch": 7.097731239092496,
+ "grad_norm": 0.05307295173406601,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.961151599884033,
+ "step": 511
+ },
+ {
+ "epoch": 7.111692844677138,
+ "grad_norm": 0.07003362476825714,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.9427876472473145,
+ "step": 512
+ },
+ {
+ "epoch": 7.12565445026178,
+ "grad_norm": 0.06498776376247406,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.947838306427002,
+ "step": 513
+ },
+ {
+ "epoch": 7.139616055846423,
+ "grad_norm": 0.058723628520965576,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.938859939575195,
+ "step": 514
+ },
+ {
+ "epoch": 7.153577661431065,
+ "grad_norm": 0.06590746343135834,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.962221145629883,
+ "step": 515
+ },
+ {
+ "epoch": 7.167539267015707,
+ "grad_norm": 0.052714601159095764,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.931011199951172,
+ "step": 516
+ },
+ {
+ "epoch": 7.181500872600349,
+ "grad_norm": 0.04296804964542389,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.93586540222168,
+ "step": 517
+ },
+ {
+ "epoch": 7.195462478184991,
+ "grad_norm": 0.03699173778295517,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.910763263702393,
+ "step": 518
+ },
+ {
+ "epoch": 7.209424083769633,
+ "grad_norm": 0.0317913293838501,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.9256439208984375,
+ "step": 519
+ },
+ {
+ "epoch": 7.223385689354275,
+ "grad_norm": 0.030313925817608833,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.938329696655273,
+ "step": 520
+ },
+ {
+ "epoch": 7.237347294938918,
+ "grad_norm": 0.02879229187965393,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.924966812133789,
+ "step": 521
+ },
+ {
+ "epoch": 7.2513089005235605,
+ "grad_norm": 0.02689451351761818,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.9079742431640625,
+ "step": 522
+ },
+ {
+ "epoch": 7.265270506108203,
+ "grad_norm": 0.028403930366039276,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.913238525390625,
+ "step": 523
+ },
+ {
+ "epoch": 7.279232111692845,
+ "grad_norm": 0.025912854820489883,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.896944046020508,
+ "step": 524
+ },
+ {
+ "epoch": 7.293193717277487,
+ "grad_norm": 0.024899760261178017,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.884920120239258,
+ "step": 525
+ },
+ {
+ "epoch": 7.307155322862129,
+ "grad_norm": 0.026912573724985123,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.9181060791015625,
+ "step": 526
+ },
+ {
+ "epoch": 7.321116928446771,
+ "grad_norm": 0.029539721086621284,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.914867401123047,
+ "step": 527
+ },
+ {
+ "epoch": 7.335078534031414,
+ "grad_norm": 0.028527872636914253,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.907217979431152,
+ "step": 528
+ },
+ {
+ "epoch": 7.349040139616056,
+ "grad_norm": 0.025879016146063805,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.905783176422119,
+ "step": 529
+ },
+ {
+ "epoch": 7.363001745200698,
+ "grad_norm": 0.026395199820399284,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.880505561828613,
+ "step": 530
+ },
+ {
+ "epoch": 7.37696335078534,
+ "grad_norm": 0.025072351098060608,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.870449066162109,
+ "step": 531
+ },
+ {
+ "epoch": 7.390924956369982,
+ "grad_norm": 0.02751019410789013,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.888952255249023,
+ "step": 532
+ },
+ {
+ "epoch": 7.4048865619546245,
+ "grad_norm": 0.025810658931732178,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.894552230834961,
+ "step": 533
+ },
+ {
+ "epoch": 7.418848167539267,
+ "grad_norm": 0.02399870567023754,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.8768229484558105,
+ "step": 534
+ },
+ {
+ "epoch": 7.43280977312391,
+ "grad_norm": 0.024639368057250977,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.9030914306640625,
+ "step": 535
+ },
+ {
+ "epoch": 7.446771378708552,
+ "grad_norm": 0.02543291077017784,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.86467170715332,
+ "step": 536
+ },
+ {
+ "epoch": 7.460732984293194,
+ "grad_norm": 0.02702544629573822,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.888283729553223,
+ "step": 537
+ },
+ {
+ "epoch": 7.474694589877836,
+ "grad_norm": 0.028904743492603302,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.866081237792969,
+ "step": 538
+ },
+ {
+ "epoch": 7.488656195462478,
+ "grad_norm": 0.03055848367512226,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.890739440917969,
+ "step": 539
+ },
+ {
+ "epoch": 7.50261780104712,
+ "grad_norm": 0.03410084918141365,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.862918853759766,
+ "step": 540
+ },
+ {
+ "epoch": 7.516579406631763,
+ "grad_norm": 0.03738495334982872,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.864728927612305,
+ "step": 541
+ },
+ {
+ "epoch": 7.530541012216405,
+ "grad_norm": 0.037646882236003876,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.865185737609863,
+ "step": 542
+ },
+ {
+ "epoch": 7.544502617801047,
+ "grad_norm": 0.03418618440628052,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.866114139556885,
+ "step": 543
+ },
+ {
+ "epoch": 7.558464223385689,
+ "grad_norm": 0.036924004554748535,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.883186340332031,
+ "step": 544
+ },
+ {
+ "epoch": 7.5724258289703315,
+ "grad_norm": 0.03590598329901695,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.863725662231445,
+ "step": 545
+ },
+ {
+ "epoch": 7.5863874345549736,
+ "grad_norm": 0.037363454699516296,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.841030120849609,
+ "step": 546
+ },
+ {
+ "epoch": 7.600349040139616,
+ "grad_norm": 0.046537864953279495,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.873837471008301,
+ "step": 547
+ },
+ {
+ "epoch": 7.614310645724259,
+ "grad_norm": 0.057226769626140594,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.865679740905762,
+ "step": 548
+ },
+ {
+ "epoch": 7.628272251308901,
+ "grad_norm": 0.06175961345434189,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.888645648956299,
+ "step": 549
+ },
+ {
+ "epoch": 7.642233856893543,
+ "grad_norm": 0.05063694715499878,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.877828598022461,
+ "step": 550
+ },
+ {
+ "epoch": 7.656195462478185,
+ "grad_norm": 0.03950628638267517,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.88637638092041,
+ "step": 551
+ },
+ {
+ "epoch": 7.670157068062827,
+ "grad_norm": 0.04183337092399597,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.8597917556762695,
+ "step": 552
+ },
+ {
+ "epoch": 7.684118673647469,
+ "grad_norm": 0.04642999917268753,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.880380153656006,
+ "step": 553
+ },
+ {
+ "epoch": 7.698080279232112,
+ "grad_norm": 0.05333106964826584,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.853932857513428,
+ "step": 554
+ },
+ {
+ "epoch": 7.712041884816754,
+ "grad_norm": 0.04764975979924202,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.870235443115234,
+ "step": 555
+ },
+ {
+ "epoch": 7.726003490401396,
+ "grad_norm": 0.043072231113910675,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.822697639465332,
+ "step": 556
+ },
+ {
+ "epoch": 7.739965095986038,
+ "grad_norm": 0.03913857042789459,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.849332809448242,
+ "step": 557
+ },
+ {
+ "epoch": 7.7539267015706805,
+ "grad_norm": 0.034836407750844955,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.842280387878418,
+ "step": 558
+ },
+ {
+ "epoch": 7.767888307155323,
+ "grad_norm": 0.033344294875860214,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.837811470031738,
+ "step": 559
+ },
+ {
+ "epoch": 7.781849912739965,
+ "grad_norm": 0.03342701122164726,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.876040458679199,
+ "step": 560
+ },
+ {
+ "epoch": 7.795811518324607,
+ "grad_norm": 0.034638747572898865,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.830340385437012,
+ "step": 561
+ },
+ {
+ "epoch": 7.80977312390925,
+ "grad_norm": 0.03358135372400284,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.828607559204102,
+ "step": 562
+ },
+ {
+ "epoch": 7.823734729493892,
+ "grad_norm": 0.036001041531562805,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.827905654907227,
+ "step": 563
+ },
+ {
+ "epoch": 7.837696335078534,
+ "grad_norm": 0.038079630583524704,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.866955280303955,
+ "step": 564
+ },
+ {
+ "epoch": 7.851657940663176,
+ "grad_norm": 0.03728470206260681,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.828039169311523,
+ "step": 565
+ },
+ {
+ "epoch": 7.865619546247818,
+ "grad_norm": 0.03920375928282738,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.841631889343262,
+ "step": 566
+ },
+ {
+ "epoch": 7.879581151832461,
+ "grad_norm": 0.032353926450014114,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.861110210418701,
+ "step": 567
+ },
+ {
+ "epoch": 7.893542757417103,
+ "grad_norm": 0.03241422027349472,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.818441390991211,
+ "step": 568
+ },
+ {
+ "epoch": 7.907504363001745,
+ "grad_norm": 0.03879080340266228,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.823749542236328,
+ "step": 569
+ },
+ {
+ "epoch": 7.9214659685863875,
+ "grad_norm": 0.03486498072743416,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.830697059631348,
+ "step": 570
+ },
+ {
+ "epoch": 7.93542757417103,
+ "grad_norm": 0.03279713913798332,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.815736770629883,
+ "step": 571
+ },
+ {
+ "epoch": 7.949389179755672,
+ "grad_norm": 0.03313925117254257,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.8364033699035645,
+ "step": 572
+ },
+ {
+ "epoch": 7.963350785340314,
+ "grad_norm": 0.03405062481760979,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.816010475158691,
+ "step": 573
+ },
+ {
+ "epoch": 7.977312390924956,
+ "grad_norm": 0.03259236365556717,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.813413619995117,
+ "step": 574
+ },
+ {
+ "epoch": 7.991273996509599,
+ "grad_norm": 0.030408985912799835,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.833840370178223,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.022680889815092087,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 2.99565052986145,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6126086115837097,
+ "eval_runtime": 60.2662,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 576
+ },
+ {
+ "epoch": 8.013961605584642,
+ "grad_norm": 0.029732126742601395,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.7816314697265625,
+ "step": 577
+ },
+ {
+ "epoch": 8.027923211169284,
+ "grad_norm": 0.04357453063130379,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.7575225830078125,
+ "step": 578
+ },
+ {
+ "epoch": 8.041884816753926,
+ "grad_norm": 0.051903221756219864,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.802519798278809,
+ "step": 579
+ },
+ {
+ "epoch": 8.055846422338568,
+ "grad_norm": 0.0539076030254364,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.799004554748535,
+ "step": 580
+ },
+ {
+ "epoch": 8.06980802792321,
+ "grad_norm": 0.043642807751894,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.7896528244018555,
+ "step": 581
+ },
+ {
+ "epoch": 8.083769633507853,
+ "grad_norm": 0.0469554178416729,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.765430927276611,
+ "step": 582
+ },
+ {
+ "epoch": 8.097731239092496,
+ "grad_norm": 0.042060062289237976,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.784132957458496,
+ "step": 583
+ },
+ {
+ "epoch": 8.111692844677139,
+ "grad_norm": 0.035101912915706635,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.767234802246094,
+ "step": 584
+ },
+ {
+ "epoch": 8.12565445026178,
+ "grad_norm": 0.042576879262924194,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.762052536010742,
+ "step": 585
+ },
+ {
+ "epoch": 8.139616055846423,
+ "grad_norm": 0.04231419041752815,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.752150535583496,
+ "step": 586
+ },
+ {
+ "epoch": 8.153577661431065,
+ "grad_norm": 0.046912599354982376,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.771881103515625,
+ "step": 587
+ },
+ {
+ "epoch": 8.167539267015707,
+ "grad_norm": 0.051271144300699234,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.773284912109375,
+ "step": 588
+ },
+ {
+ "epoch": 8.181500872600349,
+ "grad_norm": 0.04655958339571953,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.769917964935303,
+ "step": 589
+ },
+ {
+ "epoch": 8.195462478184991,
+ "grad_norm": 0.04463687539100647,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.77938175201416,
+ "step": 590
+ },
+ {
+ "epoch": 8.209424083769633,
+ "grad_norm": 0.047959886491298676,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.785962104797363,
+ "step": 591
+ },
+ {
+ "epoch": 8.223385689354275,
+ "grad_norm": 0.04414986073970795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.786281585693359,
+ "step": 592
+ },
+ {
+ "epoch": 8.237347294938917,
+ "grad_norm": 0.044150348752737045,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.777070045471191,
+ "step": 593
+ },
+ {
+ "epoch": 8.25130890052356,
+ "grad_norm": 0.037002332508563995,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.77547550201416,
+ "step": 594
+ },
+ {
+ "epoch": 8.265270506108202,
+ "grad_norm": 0.03417367860674858,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.782101154327393,
+ "step": 595
+ },
+ {
+ "epoch": 8.279232111692846,
+ "grad_norm": 0.03069210797548294,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.745595455169678,
+ "step": 596
+ },
+ {
+ "epoch": 8.293193717277488,
+ "grad_norm": 0.03228718787431717,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.7663655281066895,
+ "step": 597
+ },
+ {
+ "epoch": 8.30715532286213,
+ "grad_norm": 0.028603695333003998,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.767274856567383,
+ "step": 598
+ },
+ {
+ "epoch": 8.321116928446772,
+ "grad_norm": 0.03098377026617527,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.7853684425354,
+ "step": 599
+ },
+ {
+ "epoch": 8.335078534031414,
+ "grad_norm": 0.029242439195513725,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.738029956817627,
+ "step": 600
+ },
+ {
+ "epoch": 8.349040139616056,
+ "grad_norm": 0.02751081995666027,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.76801872253418,
+ "step": 601
+ },
+ {
+ "epoch": 8.363001745200698,
+ "grad_norm": 0.02838125079870224,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.743884086608887,
+ "step": 602
+ },
+ {
+ "epoch": 8.37696335078534,
+ "grad_norm": 0.029203062877058983,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.746668815612793,
+ "step": 603
+ },
+ {
+ "epoch": 8.390924956369982,
+ "grad_norm": 0.02679944597184658,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.760764122009277,
+ "step": 604
+ },
+ {
+ "epoch": 8.404886561954624,
+ "grad_norm": 0.028955848887562752,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.7353668212890625,
+ "step": 605
+ },
+ {
+ "epoch": 8.418848167539267,
+ "grad_norm": 0.02880851738154888,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.766837120056152,
+ "step": 606
+ },
+ {
+ "epoch": 8.432809773123909,
+ "grad_norm": 0.02495497092604637,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.78959321975708,
+ "step": 607
+ },
+ {
+ "epoch": 8.44677137870855,
+ "grad_norm": 0.026064753532409668,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.736433029174805,
+ "step": 608
+ },
+ {
+ "epoch": 8.460732984293193,
+ "grad_norm": 0.027566388249397278,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.782382965087891,
+ "step": 609
+ },
+ {
+ "epoch": 8.474694589877837,
+ "grad_norm": 0.02967359498143196,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.735286712646484,
+ "step": 610
+ },
+ {
+ "epoch": 8.488656195462479,
+ "grad_norm": 0.03079676441848278,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.725196838378906,
+ "step": 611
+ },
+ {
+ "epoch": 8.502617801047121,
+ "grad_norm": 0.032531846314668655,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.719333648681641,
+ "step": 612
+ },
+ {
+ "epoch": 8.516579406631763,
+ "grad_norm": 0.03120352514088154,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.723845958709717,
+ "step": 613
+ },
+ {
+ "epoch": 8.530541012216405,
+ "grad_norm": 0.03328424692153931,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.7669782638549805,
+ "step": 614
+ },
+ {
+ "epoch": 8.544502617801047,
+ "grad_norm": 0.035601649433374405,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.714925765991211,
+ "step": 615
+ },
+ {
+ "epoch": 8.55846422338569,
+ "grad_norm": 0.03675796464085579,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.715317726135254,
+ "step": 616
+ },
+ {
+ "epoch": 8.572425828970331,
+ "grad_norm": 0.03701988235116005,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.724606990814209,
+ "step": 617
+ },
+ {
+ "epoch": 8.586387434554974,
+ "grad_norm": 0.036719705909490585,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.746044635772705,
+ "step": 618
+ },
+ {
+ "epoch": 8.600349040139616,
+ "grad_norm": 0.034756604582071304,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.74885368347168,
+ "step": 619
+ },
+ {
+ "epoch": 8.614310645724258,
+ "grad_norm": 0.03237922117114067,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.721858501434326,
+ "step": 620
+ },
+ {
+ "epoch": 8.6282722513089,
+ "grad_norm": 0.028420142829418182,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.733814239501953,
+ "step": 621
+ },
+ {
+ "epoch": 8.642233856893542,
+ "grad_norm": 0.026203038170933723,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.726088523864746,
+ "step": 622
+ },
+ {
+ "epoch": 8.656195462478184,
+ "grad_norm": 0.02627231739461422,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.727700233459473,
+ "step": 623
+ },
+ {
+ "epoch": 8.670157068062828,
+ "grad_norm": 0.02920261025428772,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.731454849243164,
+ "step": 624
+ },
+ {
+ "epoch": 8.68411867364747,
+ "grad_norm": 0.02792907878756523,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.734154224395752,
+ "step": 625
+ },
+ {
+ "epoch": 8.698080279232112,
+ "grad_norm": 0.026770325377583504,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.715337753295898,
+ "step": 626
+ },
+ {
+ "epoch": 8.712041884816754,
+ "grad_norm": 0.027346540242433548,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.716512680053711,
+ "step": 627
+ },
+ {
+ "epoch": 8.726003490401396,
+ "grad_norm": 0.031965773552656174,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.7072038650512695,
+ "step": 628
+ },
+ {
+ "epoch": 8.739965095986038,
+ "grad_norm": 0.03708025813102722,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.749785900115967,
+ "step": 629
+ },
+ {
+ "epoch": 8.75392670157068,
+ "grad_norm": 0.04846490919589996,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.735586166381836,
+ "step": 630
+ },
+ {
+ "epoch": 8.767888307155323,
+ "grad_norm": 0.053352758288383484,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.720524787902832,
+ "step": 631
+ },
+ {
+ "epoch": 8.781849912739965,
+ "grad_norm": 0.040559202432632446,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.719901084899902,
+ "step": 632
+ },
+ {
+ "epoch": 8.795811518324607,
+ "grad_norm": 0.03333032503724098,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.708139419555664,
+ "step": 633
+ },
+ {
+ "epoch": 8.809773123909249,
+ "grad_norm": 0.030532941222190857,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.728184223175049,
+ "step": 634
+ },
+ {
+ "epoch": 8.823734729493891,
+ "grad_norm": 0.03337649628520012,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.726020336151123,
+ "step": 635
+ },
+ {
+ "epoch": 8.837696335078533,
+ "grad_norm": 0.03352045267820358,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.7196855545043945,
+ "step": 636
+ },
+ {
+ "epoch": 8.851657940663177,
+ "grad_norm": 0.03662872314453125,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.732823371887207,
+ "step": 637
+ },
+ {
+ "epoch": 8.86561954624782,
+ "grad_norm": 0.037085745483636856,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.728443145751953,
+ "step": 638
+ },
+ {
+ "epoch": 8.879581151832461,
+ "grad_norm": 0.036635689437389374,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.713988780975342,
+ "step": 639
+ },
+ {
+ "epoch": 8.893542757417103,
+ "grad_norm": 0.03719659522175789,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.715950965881348,
+ "step": 640
+ },
+ {
+ "epoch": 8.907504363001745,
+ "grad_norm": 0.03719159588217735,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.7124924659729,
+ "step": 641
+ },
+ {
+ "epoch": 8.921465968586388,
+ "grad_norm": 0.04275721684098244,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.708404064178467,
+ "step": 642
+ },
+ {
+ "epoch": 8.93542757417103,
+ "grad_norm": 0.04204598814249039,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.715372085571289,
+ "step": 643
+ },
+ {
+ "epoch": 8.949389179755672,
+ "grad_norm": 0.03938259929418564,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.729745864868164,
+ "step": 644
+ },
+ {
+ "epoch": 8.963350785340314,
+ "grad_norm": 0.036226850003004074,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.733749866485596,
+ "step": 645
+ },
+ {
+ "epoch": 8.977312390924956,
+ "grad_norm": 0.03542874753475189,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.706825256347656,
+ "step": 646
+ },
+ {
+ "epoch": 8.991273996509598,
+ "grad_norm": 0.029055381193757057,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.718567848205566,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.01990380510687828,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 2.943183183670044,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.601699709892273,
+ "eval_runtime": 60.0221,
+ "eval_samples_per_second": 40.685,
+ "eval_steps_per_second": 0.65,
+ "step": 648
+ },
+ {
+ "epoch": 9.013961605584642,
+ "grad_norm": 0.030904127284884453,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.660327911376953,
+ "step": 649
+ },
+ {
+ "epoch": 9.027923211169284,
+ "grad_norm": 0.0441313236951828,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.701565742492676,
+ "step": 650
+ },
+ {
+ "epoch": 9.041884816753926,
+ "grad_norm": 0.047547727823257446,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.666440486907959,
+ "step": 651
+ },
+ {
+ "epoch": 9.055846422338568,
+ "grad_norm": 0.04549463838338852,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.660548210144043,
+ "step": 652
+ },
+ {
+ "epoch": 9.06980802792321,
+ "grad_norm": 0.050346389412879944,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.640229225158691,
+ "step": 653
+ },
+ {
+ "epoch": 9.083769633507853,
+ "grad_norm": 0.058505453169345856,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.673881530761719,
+ "step": 654
+ },
+ {
+ "epoch": 9.097731239092496,
+ "grad_norm": 0.051989227533340454,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.693943977355957,
+ "step": 655
+ },
+ {
+ "epoch": 9.111692844677139,
+ "grad_norm": 0.052564945071935654,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.644315242767334,
+ "step": 656
+ },
+ {
+ "epoch": 9.12565445026178,
+ "grad_norm": 0.05647841840982437,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.681768417358398,
+ "step": 657
+ },
+ {
+ "epoch": 9.139616055846423,
+ "grad_norm": 0.058640867471694946,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.676457405090332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153577661431065,
+ "grad_norm": 0.05386025458574295,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.659588813781738,
+ "step": 659
+ },
+ {
+ "epoch": 9.167539267015707,
+ "grad_norm": 0.048699021339416504,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.680031776428223,
+ "step": 660
+ },
+ {
+ "epoch": 9.181500872600349,
+ "grad_norm": 0.04374176263809204,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.672671318054199,
+ "step": 661
+ },
+ {
+ "epoch": 9.195462478184991,
+ "grad_norm": 0.042337436228990555,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.673431873321533,
+ "step": 662
+ },
+ {
+ "epoch": 9.209424083769633,
+ "grad_norm": 0.04564661160111427,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.632595062255859,
+ "step": 663
+ },
+ {
+ "epoch": 9.223385689354275,
+ "grad_norm": 0.04404835030436516,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.678891181945801,
+ "step": 664
+ },
+ {
+ "epoch": 9.237347294938917,
+ "grad_norm": 0.040642790496349335,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.633418560028076,
+ "step": 665
+ },
+ {
+ "epoch": 9.25130890052356,
+ "grad_norm": 0.036492060869932175,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.670109748840332,
+ "step": 666
+ },
+ {
+ "epoch": 9.265270506108202,
+ "grad_norm": 0.03735879808664322,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.608570098876953,
+ "step": 667
+ },
+ {
+ "epoch": 9.279232111692846,
+ "grad_norm": 0.034563641995191574,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.632505416870117,
+ "step": 668
+ },
+ {
+ "epoch": 9.293193717277488,
+ "grad_norm": 0.03522484749555588,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.6475629806518555,
+ "step": 669
+ },
+ {
+ "epoch": 9.30715532286213,
+ "grad_norm": 0.03216472268104553,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.627873420715332,
+ "step": 670
+ },
+ {
+ "epoch": 9.321116928446772,
+ "grad_norm": 0.029155759140849113,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.652706146240234,
+ "step": 671
+ },
+ {
+ "epoch": 9.335078534031414,
+ "grad_norm": 0.031060541048645973,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.644545078277588,
+ "step": 672
+ },
+ {
+ "epoch": 9.349040139616056,
+ "grad_norm": 0.034979186952114105,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.642645359039307,
+ "step": 673
+ },
+ {
+ "epoch": 9.363001745200698,
+ "grad_norm": 0.035258810967206955,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.647383689880371,
+ "step": 674
+ },
+ {
+ "epoch": 9.37696335078534,
+ "grad_norm": 0.031552575528621674,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.667779445648193,
+ "step": 675
+ },
+ {
+ "epoch": 9.390924956369982,
+ "grad_norm": 0.03193951025605202,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.63625431060791,
+ "step": 676
+ },
+ {
+ "epoch": 9.404886561954624,
+ "grad_norm": 0.029811711981892586,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.658000946044922,
+ "step": 677
+ },
+ {
+ "epoch": 9.418848167539267,
+ "grad_norm": 0.028464138507843018,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.6545515060424805,
+ "step": 678
+ },
+ {
+ "epoch": 9.432809773123909,
+ "grad_norm": 0.0315735749900341,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.641153335571289,
+ "step": 679
+ },
+ {
+ "epoch": 9.44677137870855,
+ "grad_norm": 0.033412106335163116,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.626931667327881,
+ "step": 680
+ },
+ {
+ "epoch": 9.460732984293193,
+ "grad_norm": 0.03757626190781593,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.671254634857178,
+ "step": 681
+ },
+ {
+ "epoch": 9.474694589877837,
+ "grad_norm": 0.03910793736577034,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.651954174041748,
+ "step": 682
+ },
+ {
+ "epoch": 9.488656195462479,
+ "grad_norm": 0.04015549272298813,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.613324165344238,
+ "step": 683
+ },
+ {
+ "epoch": 9.502617801047121,
+ "grad_norm": 0.040689144283533096,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.639243125915527,
+ "step": 684
+ },
+ {
+ "epoch": 9.516579406631763,
+ "grad_norm": 0.04075361415743828,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.649250030517578,
+ "step": 685
+ },
+ {
+ "epoch": 9.530541012216405,
+ "grad_norm": 0.03914607688784599,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.66444730758667,
+ "step": 686
+ },
+ {
+ "epoch": 9.544502617801047,
+ "grad_norm": 0.033253733068704605,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.647335052490234,
+ "step": 687
+ },
+ {
+ "epoch": 9.55846422338569,
+ "grad_norm": 0.03179008886218071,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.6727800369262695,
+ "step": 688
+ },
+ {
+ "epoch": 9.572425828970331,
+ "grad_norm": 0.03477025404572487,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.655778884887695,
+ "step": 689
+ },
+ {
+ "epoch": 9.586387434554974,
+ "grad_norm": 0.030094800516963005,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.680851936340332,
+ "step": 690
+ },
+ {
+ "epoch": 9.600349040139616,
+ "grad_norm": 0.030153103172779083,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.654638290405273,
+ "step": 691
+ },
+ {
+ "epoch": 9.614310645724258,
+ "grad_norm": 0.029009278863668442,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.625916481018066,
+ "step": 692
+ },
+ {
+ "epoch": 9.6282722513089,
+ "grad_norm": 0.027248440310359,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.656370162963867,
+ "step": 693
+ },
+ {
+ "epoch": 9.642233856893542,
+ "grad_norm": 0.027727462351322174,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.631473541259766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656195462478184,
+ "grad_norm": 0.026710307225584984,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.636524200439453,
+ "step": 695
+ },
+ {
+ "epoch": 9.670157068062828,
+ "grad_norm": 0.02726181596517563,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.629701614379883,
+ "step": 696
+ },
+ {
+ "epoch": 9.68411867364747,
+ "grad_norm": 0.027146652340888977,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.640240669250488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698080279232112,
+ "grad_norm": 0.027165887877345085,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.64462947845459,
+ "step": 698
+ },
+ {
+ "epoch": 9.712041884816754,
+ "grad_norm": 0.03025982715189457,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.6271820068359375,
+ "step": 699
+ },
+ {
+ "epoch": 9.726003490401396,
+ "grad_norm": 0.02837025746703148,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.629697799682617,
+ "step": 700
+ },
+ {
+ "epoch": 9.739965095986038,
+ "grad_norm": 0.026777995750308037,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.646146774291992,
+ "step": 701
+ },
+ {
+ "epoch": 9.75392670157068,
+ "grad_norm": 0.024626802653074265,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.641671657562256,
+ "step": 702
+ },
+ {
+ "epoch": 9.767888307155323,
+ "grad_norm": 0.025329895317554474,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.626345157623291,
+ "step": 703
+ },
+ {
+ "epoch": 9.781849912739965,
+ "grad_norm": 0.027804844081401825,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.621457099914551,
+ "step": 704
+ },
+ {
+ "epoch": 9.795811518324607,
+ "grad_norm": 0.028763437643647194,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.6124114990234375,
+ "step": 705
+ },
+ {
+ "epoch": 9.809773123909249,
+ "grad_norm": 0.030769916251301765,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.604399681091309,
+ "step": 706
+ },
+ {
+ "epoch": 9.823734729493891,
+ "grad_norm": 0.03060949593782425,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.640804290771484,
+ "step": 707
+ },
+ {
+ "epoch": 9.837696335078533,
+ "grad_norm": 0.029832778498530388,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.593376159667969,
+ "step": 708
+ },
+ {
+ "epoch": 9.851657940663177,
+ "grad_norm": 0.02952776849269867,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.619533538818359,
+ "step": 709
+ },
+ {
+ "epoch": 9.86561954624782,
+ "grad_norm": 0.02894212119281292,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.640516757965088,
+ "step": 710
+ },
+ {
+ "epoch": 9.879581151832461,
+ "grad_norm": 0.026435624808073044,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.6342997550964355,
+ "step": 711
+ },
+ {
+ "epoch": 9.893542757417103,
+ "grad_norm": 0.02591666206717491,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.630884170532227,
+ "step": 712
+ },
+ {
+ "epoch": 9.907504363001745,
+ "grad_norm": 0.02702309936285019,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.632085800170898,
+ "step": 713
+ },
+ {
+ "epoch": 9.921465968586388,
+ "grad_norm": 0.026766657829284668,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.6231231689453125,
+ "step": 714
+ },
+ {
+ "epoch": 9.93542757417103,
+ "grad_norm": 0.026666559278964996,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.601619243621826,
+ "step": 715
+ },
+ {
+ "epoch": 9.949389179755672,
+ "grad_norm": 0.026801064610481262,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.641724109649658,
+ "step": 716
+ },
+ {
+ "epoch": 9.963350785340314,
+ "grad_norm": 0.03152798116207123,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.642511367797852,
+ "step": 717
+ },
+ {
+ "epoch": 9.977312390924956,
+ "grad_norm": 0.035406943410634995,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.635779857635498,
+ "step": 718
+ },
+ {
+ "epoch": 9.991273996509598,
+ "grad_norm": 0.037504684180021286,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.629463195800781,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02263174019753933,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 2.876767635345459,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.5942904949188232,
+ "eval_runtime": 60.2086,
+ "eval_samples_per_second": 40.559,
+ "eval_steps_per_second": 0.648,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03707852587103844,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.5765533447265625,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.05454680323600769,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.576363563537598,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.04832114651799202,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.586915969848633,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.050469402223825455,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.5265092849731445,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0622565895318985,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.592833518981934,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0640949010848999,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.5855560302734375,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.06502196192741394,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.591236114501953,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.06167149916291237,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.56747579574585,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.06747374683618546,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.60452938079834,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.07210271060466766,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.561588764190674,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.0733448714017868,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.610766410827637,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.0731528028845787,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.589570045471191,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.0822535902261734,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.631077289581299,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.06921929866075516,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.628293991088867,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.07443586736917496,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.627321243286133,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.07153813540935516,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.641068935394287,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.069196417927742,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.615875720977783,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06483697891235352,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.6165313720703125,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.05482150986790657,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.599890232086182,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.05364019051194191,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.61989164352417,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.053216829895973206,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.564305305480957,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05740506947040558,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.582683563232422,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.05587657541036606,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.61264705657959,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.050252825021743774,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.648050308227539,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.050955042243003845,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.615367889404297,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04903746768832207,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.617378234863281,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04252410680055618,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.609130859375,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.03801053389906883,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.625149250030518,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.03422372043132782,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.606106281280518,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.030507879331707954,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.6125993728637695,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.0340375155210495,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.5690202713012695,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.030507374554872513,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.6042985916137695,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027383938431739807,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.585570812225342,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.028792303055524826,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.566786766052246,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02866457775235176,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.566376686096191,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.028657952323555946,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.581750869750977,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.028227943927049637,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.576788902282715,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.024757765233516693,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.59373664855957,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.027442635968327522,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.570978164672852,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.02590034157037735,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.571286201477051,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.025696570053696632,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.542209625244141,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.025478003546595573,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.5900421142578125,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.02563696913421154,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.544609546661377,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.025326019152998924,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.5808515548706055,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.024502627551555634,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.560579776763916,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.02687518671154976,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.567154884338379,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.025280235335230827,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.560011863708496,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.026266315951943398,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.5758442878723145,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.026058891788125038,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.583309173583984,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.02620471641421318,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.547737121582031,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02626362442970276,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.556955337524414,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.024503158405423164,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.580735206604004,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02580532431602478,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.564000129699707,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.024437161162495613,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.545594215393066,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.026664750650525093,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.56415319442749,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.02499818615615368,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.550689697265625,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.027514580637216568,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.5548248291015625,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.026799403131008148,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.548295021057129,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.02685476280748844,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.5660600662231445,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.026104001328349113,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.556082248687744,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.027479661628603935,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.5580244064331055,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.02638247236609459,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.550073623657227,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.025853842496871948,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.578088760375977,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.02636682614684105,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.58807373046875,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.02684398740530014,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.561017036437988,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.028538964688777924,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.548762321472168,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.028327809646725655,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.548969745635986,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.025131452828645706,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.564047813415527,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.026498543098568916,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.553997039794922,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.030081963166594505,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.5787787437438965,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.03201213851571083,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.566424369812012,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.020518779754638672,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8421616554260254,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.5899484157562256,
+ "eval_runtime": 60.2671,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03692612797021866,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.492715835571289,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.050258126109838486,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.480323791503906,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.04179152473807335,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.484371185302734,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.04160204157233238,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.4905524253845215,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.03974860534071922,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.503803253173828,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.037031810730695724,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.494629859924316,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.038762833923101425,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.495111465454102,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.0412563718855381,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.490113735198975,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.041371624916791916,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.490643501281738,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04151557758450508,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.49881649017334,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04037577286362648,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.492178916931152,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.03676290065050125,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.477460861206055,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.038055796176195145,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.509150981903076,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.041152238845825195,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.47133731842041,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.03728092834353447,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.508072376251221,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.035338856279850006,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.505619049072266,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.037813492119312286,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.509489059448242,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.03789111226797104,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.508646488189697,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.03880828619003296,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.52567720413208,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.03611127287149429,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.498889446258545,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.036903638392686844,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.498601913452148,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03620576485991478,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.490840911865234,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03555903211236,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.506166458129883,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.03441924974322319,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.498001575469971,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.0334339402616024,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.491082191467285,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.031418606638908386,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.527408599853516,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03073921985924244,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.514107704162598,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.029589885845780373,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.48323917388916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.02974501997232437,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.51338529586792,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.029247552156448364,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.504205226898193,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.03132032975554466,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.505674362182617,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03088037669658661,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.477198600769043,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.031054604798555374,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.498721122741699,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03293661028146744,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.527826309204102,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.03181406110525131,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.495999336242676,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.029380083084106445,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.530351638793945,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.029296347871422768,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.492367744445801,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.02958036959171295,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.493037700653076,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03078395500779152,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.47504186630249,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03280661255121231,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.518409729003906,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.03323954716324806,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.460753917694092,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.032254140824079514,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.474285125732422,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.03018898330628872,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.5162353515625,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.02802245132625103,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.538684844970703,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.02821732498705387,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.4850287437438965,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.029189350083470345,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.50770378112793,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.02851196378469467,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.499988555908203,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.02741030789911747,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.499852180480957,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.028708625584840775,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.465211868286133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.028519680723547935,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.502464294433594,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.027846841141581535,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.503248691558838,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030456233769655228,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.490391731262207,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029952052980661392,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.50056266784668,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.02925778552889824,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.49215030670166,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.029679695144295692,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.523841857910156,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.031178541481494904,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.499871730804443,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.031264498829841614,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.50546932220459,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03482109308242798,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.494480133056641,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.03341105580329895,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.532899856567383,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028623664751648903,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.5016279220581055,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028600146993994713,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.511589050292969,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.029127901419997215,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.502058982849121,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.027898237109184265,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.493182182312012,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.028683282434940338,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.519793510437012,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.02887023612856865,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.457289695739746,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.02811296097934246,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.49398946762085,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.028485111892223358,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.478344440460205,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02914245054125786,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.487666130065918,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.028030497953295708,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.518488883972168,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.027973609045147896,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.515271186828613,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.027458522468805313,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.490856647491455,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.01909293793141842,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.822120189666748,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5866912603378296,
+ "eval_runtime": 61.3185,
+ "eval_samples_per_second": 39.825,
+ "eval_steps_per_second": 0.636,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03683992475271225,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.438363075256348,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.05042923241853714,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.443560600280762,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.055914171040058136,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.422282695770264,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.057007886469364166,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.442644119262695,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.05999710038304329,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.44780969619751,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06472347676753998,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.485074996948242,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.06923902779817581,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.457515716552734,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.0710320770740509,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.465476989746094,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.06942574679851532,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.426226615905762,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.06867868453264236,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.438092231750488,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.0688796266913414,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.446664810180664,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.0690719485282898,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.456482887268066,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.06068912148475647,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.4730730056762695,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.05411486327648163,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.4596452713012695,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.056623928248882294,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.418197154998779,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.05634097009897232,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.482892036437988,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.05056057870388031,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.457589149475098,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.04460986703634262,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.451439380645752,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.045022401958703995,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.452855110168457,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.04054102674126625,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.417323112487793,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.03682882338762283,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.465713977813721,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.03528819605708122,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.4502339363098145,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.03568802773952484,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.422399520874023,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.0363088883459568,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.447678089141846,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.03460463136434555,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.437664985656738,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.034751344472169876,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.489151954650879,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.033658601343631744,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.4471635818481445,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.03433239087462425,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.455332279205322,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.032074104994535446,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.435157775878906,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.032168325036764145,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.435118675231934,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.03007678873836994,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.42781925201416,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03094375506043434,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.433105945587158,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.032186444848775864,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.383000373840332,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.030610304325819016,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.473257064819336,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.030257435515522957,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.453041076660156,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.0301323514431715,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.396578788757324,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.030934512615203857,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.445223808288574,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.029907869175076485,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.442287445068359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03163841366767883,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.434650421142578,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.029627414420247078,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.454374313354492,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.031026875600218773,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.413784027099609,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03099491260945797,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.458438873291016,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.02981731668114662,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.438560962677002,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.02927570603787899,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.44493293762207,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.029779717326164246,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.445121765136719,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.02909068390727043,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.4736127853393555,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.029217200353741646,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.434604644775391,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.029872911050915718,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.454591751098633,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.028742222115397453,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.468222141265869,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.030313383787870407,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.45759391784668,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.03014662116765976,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.446331977844238,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029481127858161926,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.478209495544434,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.028485266491770744,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.444445610046387,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.02857700176537037,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.430604934692383,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.029038602486252785,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.43759298324585,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.029576119035482407,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.474579811096191,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.02897879108786583,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.440483093261719,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.029799485579133034,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.4839277267456055,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.030710609629750252,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.466696739196777,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.030270954594016075,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.455197334289551,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.03170182928442955,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.454664707183838,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030684730038046837,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.437619209289551,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.030972573906183243,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.431796073913574,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028914963826537132,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.460618019104004,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.029810264706611633,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.407312393188477,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.029806939885020256,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.466520309448242,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.03020886518061161,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.427349090576172,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030598826706409454,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.427891731262207,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03192012757062912,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.448113918304443,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.030492989346385002,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.450060844421387,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03188254311680794,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.43007230758667,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02081947773694992,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.782536506652832,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5860040187835693,
+ "eval_runtime": 60.6725,
+ "eval_samples_per_second": 40.249,
+ "eval_steps_per_second": 0.643,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.036721356213092804,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.358441352844238,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.056633755564689636,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.361527919769287,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.06175022944808006,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.3827924728393555,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05159056559205055,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.354856491088867,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.04885871335864067,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.359966278076172,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.04407693073153496,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.364612579345703,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.04891050234436989,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.344021320343018,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.04643657058477402,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.366066932678223,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.05241481214761734,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.3390607833862305,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.05837438255548477,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.390261650085449,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.058482322841882706,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.3856201171875,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.06507553160190582,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.414223670959473,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.0575365275144577,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.402576446533203,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05055592209100723,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.389563083648682,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.047167956829071045,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.402750015258789,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.04473405331373215,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.358315467834473,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.042440690100193024,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.392551898956299,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04465719312429428,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.373558044433594,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.04337528720498085,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.37944221496582,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.04322915896773338,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.38493013381958,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.039873406291007996,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.389793395996094,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.0413178876042366,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.427964210510254,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.04567280039191246,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.404172897338867,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04154926910996437,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.391359329223633,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.03774876892566681,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.381457328796387,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03588850423693657,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.392526626586914,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.03757261857390404,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.39273738861084,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03849692642688751,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.409083366394043,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03787294402718544,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.3776044845581055,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.034496866166591644,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.413462162017822,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.03973205015063286,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.359318733215332,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.036192040890455246,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.392844200134277,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.03436632081866264,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.421384811401367,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.035658642649650574,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.386757850646973,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.03434896469116211,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.364978313446045,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03642290085554123,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.385326385498047,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.03427736461162567,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.406352996826172,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.034845270216464996,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.370953559875488,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03390975296497345,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.3799943923950195,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.03134961053729057,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.419156551361084,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.03512260690331459,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.404660701751709,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03319067507982254,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.388535022735596,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03113507106900215,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.403522968292236,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03559132665395737,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.382323265075684,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.03579750657081604,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.37620210647583,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.033675212413072586,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.411443710327148,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.03220067173242569,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.443470001220703,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031610481441020966,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.406038284301758,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.02958764135837555,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.41604471206665,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.0318237729370594,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4253387451171875,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03260074555873871,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.370792388916016,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.03359333053231239,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.397639274597168,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032516930252313614,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.4140777587890625,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.03203396871685982,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.408872604370117,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.03201935440301895,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.402975082397461,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.03330211713910103,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.394018173217773,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.0348181314766407,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.380587100982666,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.03174913674592972,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.372378349304199,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.029864462092518806,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.42018985748291,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.033312153071165085,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.385973930358887,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03203144669532776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.395167827606201,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.031578030437231064,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.3982062339782715,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.030885562300682068,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.400731086730957,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03235667943954468,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.389830589294434,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.032664328813552856,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.401277542114258,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.030354196205735207,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.422520637512207,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.031162403523921967,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.393032073974609,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.032187070697546005,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.464164733886719,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.03445067256689072,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.368991851806641,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03398612141609192,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.418752670288086,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.029730670154094696,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.41429328918457,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022440090775489807,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7315595149993896,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5853753089904785,
+ "eval_runtime": 60.7174,
+ "eval_samples_per_second": 40.219,
+ "eval_steps_per_second": 0.642,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.03754177689552307,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.304028511047363,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.051616180688142776,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.34659481048584,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055258627980947495,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.310445785522461,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.049489911645650864,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3416619300842285,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05377533286809921,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.3143310546875,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.05443088337779045,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.320600986480713,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.050222691148519516,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.334789752960205,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.051904670894145966,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.316756248474121,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.05376426503062248,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.274116516113281,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.053313180804252625,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.321177005767822,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.058234598487615585,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.3169264793396,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05730101093649864,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.307013511657715,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.058140840381383896,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.332961082458496,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05654815956950188,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.315462112426758,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.05129384249448776,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.338476181030273,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.047652702778577805,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.321232795715332,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.050912003964185715,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.318971633911133,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.0514022521674633,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.3334150314331055,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.05457981303334236,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.325802803039551,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04678885266184807,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.307281494140625,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.05048161372542381,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.299831390380859,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.05334261804819107,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.313231468200684,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.05055771768093109,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.36859130859375,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04955561086535454,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.301530361175537,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.04548008739948273,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.339517593383789,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.046585116535425186,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.333036422729492,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.0503716766834259,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.364710807800293,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.040569406002759933,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.369089126586914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.03754054754972458,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.3349432945251465,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03775954991579056,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.333134174346924,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.04066266492009163,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.354358673095703,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.040490977466106415,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.360981464385986,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.0426657609641552,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.357726097106934,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.0415605828166008,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.363457679748535,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.04020509123802185,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.372447967529297,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.04101516678929329,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.292452812194824,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.03546427935361862,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.33892297744751,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.03661105036735535,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.329778671264648,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03565476834774017,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.351012229919434,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.03591311350464821,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.376980781555176,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.03665630891919136,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.354767799377441,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.0371236577630043,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.374916076660156,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.03464849665760994,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.3713059425354,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03345152735710144,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.350670337677002,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.03357568010687828,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.3573713302612305,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03503486514091492,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.362038612365723,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.03455005958676338,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.3263773918151855,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.03173859417438507,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.351345062255859,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03301836550235748,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.398364067077637,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.033733855932950974,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.335268020629883,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034080155193805695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.371492385864258,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.034396376460790634,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.382076740264893,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.03303401917219162,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.36623477935791,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03297369182109833,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3148393630981445,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.034819211810827255,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.358110427856445,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.033389780670404434,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.367427825927734,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.032363444566726685,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.3813676834106445,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03509577363729477,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.386520862579346,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.034685730934143066,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.33738374710083,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.034755002707242966,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.360980987548828,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03461688384413719,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.396650314331055,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03392560034990311,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.356130123138428,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.03342147544026375,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.359145164489746,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.032852694392204285,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.3796162605285645,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03472859039902687,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.361110687255859,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03277098760008812,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.361811637878418,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.03341882303357124,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.369524955749512,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.032347049564123154,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.369096755981445,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03161193057894707,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.339844703674316,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.03322301432490349,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.376675605773926,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03419092297554016,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.346864223480225,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02461826801300049,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7236275672912598,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.5859372019767761,
+ "eval_runtime": 59.0139,
+ "eval_samples_per_second": 41.38,
+ "eval_steps_per_second": 0.661,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013961605584642,
+ "grad_norm": 0.039559684693813324,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 4.289841651916504,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027923211169284,
+ "grad_norm": 0.05554218962788582,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 4.255925178527832,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041884816753926,
+ "grad_norm": 0.05330996215343475,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 4.272125244140625,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055846422338568,
+ "grad_norm": 0.053247954696416855,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 4.290765285491943,
+ "step": 1084
+ },
+ {
+ "epoch": 15.06980802792321,
+ "grad_norm": 0.05471150949597359,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 4.271744728088379,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083769633507853,
+ "grad_norm": 0.05976719409227371,
+ "learning_rate": 0.000491285979372924,
+ "loss": 4.28203821182251,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097731239092496,
+ "grad_norm": 0.05369880795478821,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 4.275313377380371,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111692844677139,
+ "grad_norm": 0.052529267966747284,
+ "learning_rate": 0.000490873338088198,
+ "loss": 4.25678825378418,
+ "step": 1088
+ },
+ {
+ "epoch": 15.12565445026178,
+ "grad_norm": 0.05184522643685341,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 4.260510444641113,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139616055846423,
+ "grad_norm": 0.051606856286525726,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 4.283233642578125,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153577661431065,
+ "grad_norm": 0.05318903923034668,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 4.286595344543457,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167539267015707,
+ "grad_norm": 0.058525051921606064,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 4.282374382019043,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181500872600349,
+ "grad_norm": 0.054776228964328766,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 4.2913126945495605,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195462478184991,
+ "grad_norm": 0.04907633364200592,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 4.297187805175781,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209424083769633,
+ "grad_norm": 0.05163136497139931,
+ "learning_rate": 0.000489424334303338,
+ "loss": 4.2968645095825195,
+ "step": 1095
+ },
+ {
+ "epoch": 15.223385689354275,
+ "grad_norm": 0.053157489746809006,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 4.285839080810547,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237347294938917,
+ "grad_norm": 0.05251535773277283,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 4.311246871948242,
+ "step": 1097
+ },
+ {
+ "epoch": 15.25130890052356,
+ "grad_norm": 0.048785097897052765,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 4.300364971160889,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265270506108202,
+ "grad_norm": 0.050393108278512955,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 4.299890995025635,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279232111692846,
+ "grad_norm": 0.04894037917256355,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 4.312449932098389,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293193717277488,
+ "grad_norm": 0.05274952948093414,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 4.2889814376831055,
+ "step": 1101
+ },
+ {
+ "epoch": 15.30715532286213,
+ "grad_norm": 0.05946570262312889,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 4.2731170654296875,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321116928446772,
+ "grad_norm": 0.054245397448539734,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 4.272463798522949,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335078534031414,
+ "grad_norm": 0.042092181742191315,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 4.299953937530518,
+ "step": 1104
+ },
+ {
+ "epoch": 15.349040139616056,
+ "grad_norm": 0.040222618728876114,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 4.292524814605713,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363001745200698,
+ "grad_norm": 0.04298696666955948,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 4.31342887878418,
+ "step": 1106
+ },
+ {
+ "epoch": 15.37696335078534,
+ "grad_norm": 0.04271050915122032,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 4.277663707733154,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390924956369982,
+ "grad_norm": 0.042364370077848434,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 4.311102390289307,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404886561954624,
+ "grad_norm": 0.04343319311738014,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 4.311610221862793,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418848167539267,
+ "grad_norm": 0.041321512311697006,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 4.29677677154541,
+ "step": 1110
+ },
+ {
+ "epoch": 15.432809773123909,
+ "grad_norm": 0.04107942432165146,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 4.3163018226623535,
+ "step": 1111
+ },
+ {
+ "epoch": 15.44677137870855,
+ "grad_norm": 0.03844515234231949,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 4.311301231384277,
+ "step": 1112
+ },
+ {
+ "epoch": 15.460732984293193,
+ "grad_norm": 0.03731007128953934,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 4.31325626373291,
+ "step": 1113
+ },
+ {
+ "epoch": 15.474694589877837,
+ "grad_norm": 0.03807826340198517,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 4.296201705932617,
+ "step": 1114
+ },
+ {
+ "epoch": 15.488656195462479,
+ "grad_norm": 0.040137600153684616,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 4.320395469665527,
+ "step": 1115
+ },
+ {
+ "epoch": 15.502617801047121,
+ "grad_norm": 0.038484394550323486,
+ "learning_rate": 0.000485033394934934,
+ "loss": 4.313858985900879,
+ "step": 1116
+ },
+ {
+ "epoch": 15.516579406631763,
+ "grad_norm": 0.03603338822722435,
+ "learning_rate": 0.000484822676912974,
+ "loss": 4.297666549682617,
+ "step": 1117
+ },
+ {
+ "epoch": 15.530541012216405,
+ "grad_norm": 0.038647182285785675,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 4.305002689361572,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544502617801047,
+ "grad_norm": 0.03907139599323273,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 4.329583644866943,
+ "step": 1119
+ },
+ {
+ "epoch": 15.55846422338569,
+ "grad_norm": 0.03621622920036316,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 4.290896415710449,
+ "step": 1120
+ },
+ {
+ "epoch": 15.572425828970331,
+ "grad_norm": 0.037496887147426605,
+ "learning_rate": 0.000483978341612154,
+ "loss": 4.33336877822876,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586387434554974,
+ "grad_norm": 0.03778311610221863,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 4.285354137420654,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600349040139616,
+ "grad_norm": 0.03782997280359268,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 4.29697847366333,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614310645724258,
+ "grad_norm": 0.03665657714009285,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 4.320103168487549,
+ "step": 1124
+ },
+ {
+ "epoch": 15.6282722513089,
+ "grad_norm": 0.03796127438545227,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 4.317153453826904,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642233856893542,
+ "grad_norm": 0.036425214260816574,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 4.293642520904541,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656195462478184,
+ "grad_norm": 0.03629784658551216,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 4.291100025177002,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670157068062828,
+ "grad_norm": 0.03674762696027756,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 4.295144557952881,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68411867364747,
+ "grad_norm": 0.036291178315877914,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 4.313133239746094,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698080279232112,
+ "grad_norm": 0.03817860782146454,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 4.294533729553223,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712041884816754,
+ "grad_norm": 0.036819685250520706,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 4.299263954162598,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726003490401396,
+ "grad_norm": 0.035103823989629745,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 4.287391185760498,
+ "step": 1132
+ },
+ {
+ "epoch": 15.739965095986038,
+ "grad_norm": 0.035701096057891846,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 4.318216323852539,
+ "step": 1133
+ },
+ {
+ "epoch": 15.75392670157068,
+ "grad_norm": 0.03639856353402138,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 4.301465034484863,
+ "step": 1134
+ },
+ {
+ "epoch": 15.767888307155323,
+ "grad_norm": 0.03593013063073158,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 4.311532974243164,
+ "step": 1135
+ },
+ {
+ "epoch": 15.781849912739965,
+ "grad_norm": 0.03564770147204399,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 4.3400726318359375,
+ "step": 1136
+ },
+ {
+ "epoch": 15.795811518324607,
+ "grad_norm": 0.03598494827747345,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 4.30976676940918,
+ "step": 1137
+ },
+ {
+ "epoch": 15.809773123909249,
+ "grad_norm": 0.034391652792692184,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 4.328464508056641,
+ "step": 1138
+ },
+ {
+ "epoch": 15.823734729493891,
+ "grad_norm": 0.03445672616362572,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 4.310579299926758,
+ "step": 1139
+ },
+ {
+ "epoch": 15.837696335078533,
+ "grad_norm": 0.03643282502889633,
+ "learning_rate": 0.000479936075927473,
+ "loss": 4.320905685424805,
+ "step": 1140
+ },
+ {
+ "epoch": 15.851657940663177,
+ "grad_norm": 0.03489850088953972,
+ "learning_rate": 0.000479721889242305,
+ "loss": 4.317229270935059,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86561954624782,
+ "grad_norm": 0.03609171137213707,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 4.309203147888184,
+ "step": 1142
+ },
+ {
+ "epoch": 15.879581151832461,
+ "grad_norm": 0.03705338388681412,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 4.287387371063232,
+ "step": 1143
+ },
+ {
+ "epoch": 15.893542757417103,
+ "grad_norm": 0.03697717934846878,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 4.302860260009766,
+ "step": 1144
+ },
+ {
+ "epoch": 15.907504363001745,
+ "grad_norm": 0.035338208079338074,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 4.3152265548706055,
+ "step": 1145
+ },
+ {
+ "epoch": 15.921465968586388,
+ "grad_norm": 0.03548634052276611,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 4.327335357666016,
+ "step": 1146
+ },
+ {
+ "epoch": 15.93542757417103,
+ "grad_norm": 0.035512328147888184,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 4.3067803382873535,
+ "step": 1147
+ },
+ {
+ "epoch": 15.949389179755672,
+ "grad_norm": 0.03799289092421532,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 4.320701599121094,
+ "step": 1148
+ },
+ {
+ "epoch": 15.963350785340314,
+ "grad_norm": 0.03866339474916458,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 4.321087837219238,
+ "step": 1149
+ },
+ {
+ "epoch": 15.977312390924956,
+ "grad_norm": 0.036002084612846375,
+ "learning_rate": 0.000477787822648007,
+ "loss": 4.287045478820801,
+ "step": 1150
+ },
+ {
+ "epoch": 15.991273996509598,
+ "grad_norm": 0.03433441370725632,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 4.2562665939331055,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.02580358274281025,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 2.67783260345459,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.5870338082313538,
+ "eval_runtime": 60.0469,
+ "eval_samples_per_second": 40.668,
+ "eval_steps_per_second": 0.649,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013961605584644,
+ "grad_norm": 0.04105810821056366,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 4.224295616149902,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027923211169284,
+ "grad_norm": 0.049538832157850266,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 4.222184181213379,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041884816753928,
+ "grad_norm": 0.05089569091796875,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 4.2067060470581055,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05584642233857,
+ "grad_norm": 0.04600589722394943,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 4.238527774810791,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069808027923212,
+ "grad_norm": 0.043311409652233124,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 4.242960453033447,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083769633507853,
+ "grad_norm": 0.047897014766931534,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 4.199069023132324,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097731239092496,
+ "grad_norm": 0.05633604899048805,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 4.211665153503418,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111692844677137,
+ "grad_norm": 0.056110311299562454,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 4.20163631439209,
+ "step": 1160
+ },
+ {
+ "epoch": 16.12565445026178,
+ "grad_norm": 0.05426819995045662,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 4.21889591217041,
+ "step": 1161
+ },
+ {
+ "epoch": 16.13961605584642,
+ "grad_norm": 0.05489835515618324,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 4.220406532287598,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153577661431065,
+ "grad_norm": 0.053941354155540466,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 4.231909275054932,
+ "step": 1163
+ },
+ {
+ "epoch": 16.167539267015705,
+ "grad_norm": 0.058548808097839355,
+ "learning_rate": 0.000474756648870413,
+ "loss": 4.237003326416016,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18150087260035,
+ "grad_norm": 0.05370597913861275,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 4.240085124969482,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195462478184993,
+ "grad_norm": 0.05684991553425789,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 4.203842639923096,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209424083769633,
+ "grad_norm": 0.060745254158973694,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 4.247998237609863,
+ "step": 1167
+ },
+ {
+ "epoch": 16.223385689354277,
+ "grad_norm": 0.058296725153923035,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 4.226161003112793,
+ "step": 1168
+ },
+ {
+ "epoch": 16.237347294938917,
+ "grad_norm": 0.05310584232211113,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 4.229870796203613,
+ "step": 1169
+ },
+ {
+ "epoch": 16.25130890052356,
+ "grad_norm": 0.04710390046238899,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 4.240084648132324,
+ "step": 1170
+ },
+ {
+ "epoch": 16.2652705061082,
+ "grad_norm": 0.05003240332007408,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 4.240972518920898,
+ "step": 1171
+ },
+ {
+ "epoch": 16.279232111692846,
+ "grad_norm": 0.05552851781249046,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 4.248024940490723,
+ "step": 1172
+ },
+ {
+ "epoch": 16.293193717277486,
+ "grad_norm": 0.049967192113399506,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 4.279229164123535,
+ "step": 1173
+ },
+ {
+ "epoch": 16.30715532286213,
+ "grad_norm": 0.04883681237697601,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 4.269807815551758,
+ "step": 1174
+ },
+ {
+ "epoch": 16.32111692844677,
+ "grad_norm": 0.04931480810046196,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 4.232419490814209,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335078534031414,
+ "grad_norm": 0.048746511340141296,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 4.267911911010742,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349040139616054,
+ "grad_norm": 0.04678189381957054,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 4.262979507446289,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363001745200698,
+ "grad_norm": 0.048893529921770096,
+ "learning_rate": 0.000471698352726896,
+ "loss": 4.254408359527588,
+ "step": 1178
+ },
+ {
+ "epoch": 16.376963350785342,
+ "grad_norm": 0.045027732849121094,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 4.2445902824401855,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390924956369982,
+ "grad_norm": 0.04324870929121971,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 4.2685136795043945,
+ "step": 1180
+ },
+ {
+ "epoch": 16.404886561954626,
+ "grad_norm": 0.0470522940158844,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 4.237505912780762,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418848167539267,
+ "grad_norm": 0.04896962642669678,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 4.2523345947265625,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43280977312391,
+ "grad_norm": 0.04562787711620331,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 4.2192792892456055,
+ "step": 1183
+ },
+ {
+ "epoch": 16.44677137870855,
+ "grad_norm": 0.04414575919508934,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 4.250238418579102,
+ "step": 1184
+ },
+ {
+ "epoch": 16.460732984293195,
+ "grad_norm": 0.04363945871591568,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 4.256250381469727,
+ "step": 1185
+ },
+ {
+ "epoch": 16.474694589877835,
+ "grad_norm": 0.04159797728061676,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 4.274903297424316,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48865619546248,
+ "grad_norm": 0.04248544201254845,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 4.278222560882568,
+ "step": 1187
+ },
+ {
+ "epoch": 16.50261780104712,
+ "grad_norm": 0.04379890486598015,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 4.267505645751953,
+ "step": 1188
+ },
+ {
+ "epoch": 16.516579406631763,
+ "grad_norm": 0.04319411888718605,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 4.2607269287109375,
+ "step": 1189
+ },
+ {
+ "epoch": 16.530541012216403,
+ "grad_norm": 0.04143810644745827,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 4.267977714538574,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544502617801047,
+ "grad_norm": 0.03914840891957283,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 4.260498523712158,
+ "step": 1191
+ },
+ {
+ "epoch": 16.55846422338569,
+ "grad_norm": 0.04107543081045151,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 4.2550482749938965,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57242582897033,
+ "grad_norm": 0.041316065937280655,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 4.3075456619262695,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586387434554975,
+ "grad_norm": 0.04069342464208603,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 4.234392166137695,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600349040139616,
+ "grad_norm": 0.03816857561469078,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 4.251038551330566,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61431064572426,
+ "grad_norm": 0.04095374792814255,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 4.284790992736816,
+ "step": 1196
+ },
+ {
+ "epoch": 16.6282722513089,
+ "grad_norm": 0.03788160905241966,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 4.255716323852539,
+ "step": 1197
+ },
+ {
+ "epoch": 16.642233856893544,
+ "grad_norm": 0.03784061595797539,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 4.269606590270996,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656195462478184,
+ "grad_norm": 0.04087168723344803,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 4.231165885925293,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670157068062828,
+ "grad_norm": 0.040546853095293045,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 4.288422584533691,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68411867364747,
+ "grad_norm": 0.03777262568473816,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 4.277492523193359,
+ "step": 1201
+ },
+ {
+ "epoch": 16.698080279232112,
+ "grad_norm": 0.039857733994722366,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 4.239383697509766,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712041884816752,
+ "grad_norm": 0.04387792944908142,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 4.300265312194824,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726003490401396,
+ "grad_norm": 0.04214833676815033,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 4.255545616149902,
+ "step": 1204
+ },
+ {
+ "epoch": 16.739965095986037,
+ "grad_norm": 0.04290056228637695,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 4.275530815124512,
+ "step": 1205
+ },
+ {
+ "epoch": 16.75392670157068,
+ "grad_norm": 0.03975554555654526,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 4.282685279846191,
+ "step": 1206
+ },
+ {
+ "epoch": 16.767888307155324,
+ "grad_norm": 0.039464011788368225,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 4.285504341125488,
+ "step": 1207
+ },
+ {
+ "epoch": 16.781849912739965,
+ "grad_norm": 0.039714012295007706,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 4.273477554321289,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79581151832461,
+ "grad_norm": 0.03936111181974411,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 4.304205894470215,
+ "step": 1209
+ },
+ {
+ "epoch": 16.80977312390925,
+ "grad_norm": 0.03782064467668533,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 4.307382583618164,
+ "step": 1210
+ },
+ {
+ "epoch": 16.823734729493893,
+ "grad_norm": 0.040340717881917953,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 4.308620452880859,
+ "step": 1211
+ },
+ {
+ "epoch": 16.837696335078533,
+ "grad_norm": 0.0403442308306694,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 4.248626708984375,
+ "step": 1212
+ },
+ {
+ "epoch": 16.851657940663177,
+ "grad_norm": 0.038351960480213165,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 4.261248588562012,
+ "step": 1213
+ },
+ {
+ "epoch": 16.865619546247817,
+ "grad_norm": 0.03965333476662636,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 4.278291702270508,
+ "step": 1214
+ },
+ {
+ "epoch": 16.87958115183246,
+ "grad_norm": 0.040909092873334885,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 4.241847991943359,
+ "step": 1215
+ },
+ {
+ "epoch": 16.8935427574171,
+ "grad_norm": 0.038684047758579254,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 4.281346321105957,
+ "step": 1216
+ },
+ {
+ "epoch": 16.907504363001745,
+ "grad_norm": 0.03698103874921799,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 4.3085432052612305,
+ "step": 1217
+ },
+ {
+ "epoch": 16.921465968586386,
+ "grad_norm": 0.03908613696694374,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 4.24220085144043,
+ "step": 1218
+ },
+ {
+ "epoch": 16.93542757417103,
+ "grad_norm": 0.03873388469219208,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 4.282486915588379,
+ "step": 1219
+ },
+ {
+ "epoch": 16.949389179755673,
+ "grad_norm": 0.03709883987903595,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 4.279009819030762,
+ "step": 1220
+ },
+ {
+ "epoch": 16.963350785340314,
+ "grad_norm": 0.03937513753771782,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 4.273211479187012,
+ "step": 1221
+ },
+ {
+ "epoch": 16.977312390924958,
+ "grad_norm": 0.038478728383779526,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 4.281154155731201,
+ "step": 1222
+ },
+ {
+ "epoch": 16.991273996509598,
+ "grad_norm": 0.03913680836558342,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 4.269218444824219,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.02641749382019043,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 2.6905035972595215,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.5879416465759277,
+ "eval_runtime": 60.2807,
+ "eval_samples_per_second": 40.511,
+ "eval_steps_per_second": 0.647,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013961605584644,
+ "grad_norm": 0.048571839928627014,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 4.2008538246154785,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027923211169284,
+ "grad_norm": 0.0640900656580925,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 4.180826663970947,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041884816753928,
+ "grad_norm": 0.06458134204149246,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 4.199562072753906,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05584642233857,
+ "grad_norm": 0.0563119500875473,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 4.171557426452637,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069808027923212,
+ "grad_norm": 0.050914082676172256,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 4.183539390563965,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083769633507853,
+ "grad_norm": 0.053884491324424744,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 4.210450172424316,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097731239092496,
+ "grad_norm": 0.05301345884799957,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 4.1553192138671875,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111692844677137,
+ "grad_norm": 0.051138825714588165,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 4.158510208129883,
+ "step": 1232
+ },
+ {
+ "epoch": 17.12565445026178,
+ "grad_norm": 0.05355675891041756,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 4.167689323425293,
+ "step": 1233
+ },
+ {
+ "epoch": 17.13961605584642,
+ "grad_norm": 0.06062275916337967,
+ "learning_rate": 0.000459203665939781,
+ "loss": 4.170838356018066,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153577661431065,
+ "grad_norm": 0.058666616678237915,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 4.187441825866699,
+ "step": 1235
+ },
+ {
+ "epoch": 17.167539267015705,
+ "grad_norm": 0.057622168213129044,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 4.190746784210205,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18150087260035,
+ "grad_norm": 0.069090336561203,
+ "learning_rate": 0.000458522886595554,
+ "loss": 4.224946022033691,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195462478184993,
+ "grad_norm": 0.06552847474813461,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 4.204067230224609,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209424083769633,
+ "grad_norm": 0.06602177768945694,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 4.189286708831787,
+ "step": 1239
+ },
+ {
+ "epoch": 17.223385689354277,
+ "grad_norm": 0.06751416623592377,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 4.208927154541016,
+ "step": 1240
+ },
+ {
+ "epoch": 17.237347294938917,
+ "grad_norm": 0.0706794261932373,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 4.212923049926758,
+ "step": 1241
+ },
+ {
+ "epoch": 17.25130890052356,
+ "grad_norm": 0.05346836522221565,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 4.2221784591674805,
+ "step": 1242
+ },
+ {
+ "epoch": 17.2652705061082,
+ "grad_norm": 0.05395734682679176,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 4.196428298950195,
+ "step": 1243
+ },
+ {
+ "epoch": 17.279232111692846,
+ "grad_norm": 0.05496550351381302,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 4.208031177520752,
+ "step": 1244
+ },
+ {
+ "epoch": 17.293193717277486,
+ "grad_norm": 0.05320895463228226,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 4.232810974121094,
+ "step": 1245
+ },
+ {
+ "epoch": 17.30715532286213,
+ "grad_norm": 0.05064759403467178,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 4.207826614379883,
+ "step": 1246
+ },
+ {
+ "epoch": 17.32111692844677,
+ "grad_norm": 0.04287239536643028,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 4.195648670196533,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335078534031414,
+ "grad_norm": 0.04527301341295242,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 4.19954776763916,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349040139616054,
+ "grad_norm": 0.04421805590391159,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 4.248671054840088,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363001745200698,
+ "grad_norm": 0.04487175494432449,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 4.2072954177856445,
+ "step": 1250
+ },
+ {
+ "epoch": 17.376963350785342,
+ "grad_norm": 0.04415794834494591,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 4.222837448120117,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390924956369982,
+ "grad_norm": 0.04474013298749924,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 4.213104248046875,
+ "step": 1252
+ },
+ {
+ "epoch": 17.404886561954626,
+ "grad_norm": 0.042631037533283234,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 4.209726333618164,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418848167539267,
+ "grad_norm": 0.04424271360039711,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 4.202699661254883,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43280977312391,
+ "grad_norm": 0.04151812940835953,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 4.190732955932617,
+ "step": 1255
+ },
+ {
+ "epoch": 17.44677137870855,
+ "grad_norm": 0.041588131338357925,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 4.222360610961914,
+ "step": 1256
+ },
+ {
+ "epoch": 17.460732984293195,
+ "grad_norm": 0.042378414422273636,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 4.198424339294434,
+ "step": 1257
+ },
+ {
+ "epoch": 17.474694589877835,
+ "grad_norm": 0.041907574981451035,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 4.224804401397705,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48865619546248,
+ "grad_norm": 0.043774355202913284,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 4.2276611328125,
+ "step": 1259
+ },
+ {
+ "epoch": 17.50261780104712,
+ "grad_norm": 0.04238573834300041,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 4.228109359741211,
+ "step": 1260
+ },
+ {
+ "epoch": 17.516579406631763,
+ "grad_norm": 0.045351386070251465,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 4.208807468414307,
+ "step": 1261
+ },
+ {
+ "epoch": 17.530541012216403,
+ "grad_norm": 0.04300729185342789,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 4.257306098937988,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544502617801047,
+ "grad_norm": 0.041442908346652985,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 4.208004951477051,
+ "step": 1263
+ },
+ {
+ "epoch": 17.55846422338569,
+ "grad_norm": 0.04337337985634804,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 4.23393440246582,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57242582897033,
+ "grad_norm": 0.04335462674498558,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 4.228087425231934,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586387434554975,
+ "grad_norm": 0.03987223282456398,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 4.213971138000488,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600349040139616,
+ "grad_norm": 0.041760094463825226,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 4.248366355895996,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61431064572426,
+ "grad_norm": 0.04420867934823036,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 4.278865814208984,
+ "step": 1268
+ },
+ {
+ "epoch": 17.6282722513089,
+ "grad_norm": 0.04244982823729515,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 4.223642349243164,
+ "step": 1269
+ },
+ {
+ "epoch": 17.642233856893544,
+ "grad_norm": 0.04280664771795273,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 4.231023788452148,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656195462478184,
+ "grad_norm": 0.04294133558869362,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 4.222380638122559,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670157068062828,
+ "grad_norm": 0.04179249703884125,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 4.215641975402832,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68411867364747,
+ "grad_norm": 0.04038798063993454,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 4.228580474853516,
+ "step": 1273
+ },
+ {
+ "epoch": 17.698080279232112,
+ "grad_norm": 0.04211528226733208,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 4.245234489440918,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712041884816752,
+ "grad_norm": 0.04102989658713341,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 4.218495845794678,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726003490401396,
+ "grad_norm": 0.04038692265748978,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 4.2466511726379395,
+ "step": 1276
+ },
+ {
+ "epoch": 17.739965095986037,
+ "grad_norm": 0.03958791121840477,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 4.237762451171875,
+ "step": 1277
+ },
+ {
+ "epoch": 17.75392670157068,
+ "grad_norm": 0.04129907861351967,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 4.238739013671875,
+ "step": 1278
+ },
+ {
+ "epoch": 17.767888307155324,
+ "grad_norm": 0.042581796646118164,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 4.235558032989502,
+ "step": 1279
+ },
+ {
+ "epoch": 17.781849912739965,
+ "grad_norm": 0.04339978098869324,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 4.226593971252441,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79581151832461,
+ "grad_norm": 0.03808023780584335,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 4.238213062286377,
+ "step": 1281
+ },
+ {
+ "epoch": 17.80977312390925,
+ "grad_norm": 0.04022229090332985,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 4.214224815368652,
+ "step": 1282
+ },
+ {
+ "epoch": 17.823734729493893,
+ "grad_norm": 0.04379156231880188,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 4.253068447113037,
+ "step": 1283
+ },
+ {
+ "epoch": 17.837696335078533,
+ "grad_norm": 0.04232964664697647,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 4.249903202056885,
+ "step": 1284
+ },
+ {
+ "epoch": 17.851657940663177,
+ "grad_norm": 0.040980905294418335,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 4.252640724182129,
+ "step": 1285
+ },
+ {
+ "epoch": 17.865619546247817,
+ "grad_norm": 0.04015526548027992,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 4.234862327575684,
+ "step": 1286
+ },
+ {
+ "epoch": 17.87958115183246,
+ "grad_norm": 0.04273828864097595,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 4.216028213500977,
+ "step": 1287
+ },
+ {
+ "epoch": 17.8935427574171,
+ "grad_norm": 0.04210563376545906,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 4.23359489440918,
+ "step": 1288
+ },
+ {
+ "epoch": 17.907504363001745,
+ "grad_norm": 0.038766685873270035,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 4.220360279083252,
+ "step": 1289
+ },
+ {
+ "epoch": 17.921465968586386,
+ "grad_norm": 0.03921609744429588,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 4.253805160522461,
+ "step": 1290
+ },
+ {
+ "epoch": 17.93542757417103,
+ "grad_norm": 0.04071544483304024,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 4.211221694946289,
+ "step": 1291
+ },
+ {
+ "epoch": 17.949389179755673,
+ "grad_norm": 0.03996877372264862,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 4.241739273071289,
+ "step": 1292
+ },
+ {
+ "epoch": 17.963350785340314,
+ "grad_norm": 0.03878853842616081,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 4.237001419067383,
+ "step": 1293
+ },
+ {
+ "epoch": 17.977312390924958,
+ "grad_norm": 0.03889666125178337,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 4.244929790496826,
+ "step": 1294
+ },
+ {
+ "epoch": 17.991273996509598,
+ "grad_norm": 0.03876441717147827,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 4.236083030700684,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.027286000549793243,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 2.636739730834961,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.5896694660186768,
+ "eval_runtime": 60.4742,
+ "eval_samples_per_second": 40.381,
+ "eval_steps_per_second": 0.645,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013961605584644,
+ "grad_norm": 0.04279542714357376,
+ "learning_rate": 0.000444677103403854,
+ "loss": 4.136401653289795,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027923211169284,
+ "grad_norm": 0.051685452461242676,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 4.159584045410156,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041884816753928,
+ "grad_norm": 0.05411457270383835,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 4.160367965698242,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05584642233857,
+ "grad_norm": 0.048574138432741165,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 4.134189128875732,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069808027923212,
+ "grad_norm": 0.04528478533029556,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 4.1022725105285645,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083769633507853,
+ "grad_norm": 0.04474090039730072,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 4.16152286529541,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097731239092496,
+ "grad_norm": 0.04882791265845299,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 4.1553144454956055,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111692844677137,
+ "grad_norm": 0.04831654578447342,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 4.147022724151611,
+ "step": 1304
+ },
+ {
+ "epoch": 18.12565445026178,
+ "grad_norm": 0.047365427017211914,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 4.139405250549316,
+ "step": 1305
+ },
+ {
+ "epoch": 18.13961605584642,
+ "grad_norm": 0.0514012910425663,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 4.115568161010742,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153577661431065,
+ "grad_norm": 0.05225161090493202,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 4.166429042816162,
+ "step": 1307
+ },
+ {
+ "epoch": 18.167539267015705,
+ "grad_norm": 0.05575249716639519,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 4.1414289474487305,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18150087260035,
+ "grad_norm": 0.05405383184552193,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 4.138680934906006,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195462478184993,
+ "grad_norm": 0.05185031518340111,
+ "learning_rate": 0.000441621628895411,
+ "loss": 4.159243583679199,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209424083769633,
+ "grad_norm": 0.05070916935801506,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 4.1844892501831055,
+ "step": 1311
+ },
+ {
+ "epoch": 18.223385689354277,
+ "grad_norm": 0.05037331581115723,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 4.132781028747559,
+ "step": 1312
+ },
+ {
+ "epoch": 18.237347294938917,
+ "grad_norm": 0.05082111805677414,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 4.143031120300293,
+ "step": 1313
+ },
+ {
+ "epoch": 18.25130890052356,
+ "grad_norm": 0.04742169380187988,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 4.137580871582031,
+ "step": 1314
+ },
+ {
+ "epoch": 18.2652705061082,
+ "grad_norm": 0.049042731523513794,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 4.159721374511719,
+ "step": 1315
+ },
+ {
+ "epoch": 18.279232111692846,
+ "grad_norm": 0.045649368315935135,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 4.144008636474609,
+ "step": 1316
+ },
+ {
+ "epoch": 18.293193717277486,
+ "grad_norm": 0.0473787859082222,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 4.164243698120117,
+ "step": 1317
+ },
+ {
+ "epoch": 18.30715532286213,
+ "grad_norm": 0.047626055777072906,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 4.164441108703613,
+ "step": 1318
+ },
+ {
+ "epoch": 18.32111692844677,
+ "grad_norm": 0.04738573357462883,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 4.180338382720947,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335078534031414,
+ "grad_norm": 0.04712498560547829,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 4.146285533905029,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349040139616054,
+ "grad_norm": 0.045505691319704056,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 4.162572860717773,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363001745200698,
+ "grad_norm": 0.0489528588950634,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 4.160312175750732,
+ "step": 1322
+ },
+ {
+ "epoch": 18.376963350785342,
+ "grad_norm": 0.047517575323581696,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 4.164361953735352,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390924956369982,
+ "grad_norm": 0.044214535504579544,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 4.182154655456543,
+ "step": 1324
+ },
+ {
+ "epoch": 18.404886561954626,
+ "grad_norm": 0.043775733560323715,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 4.130522727966309,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418848167539267,
+ "grad_norm": 0.044364530593156815,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 4.151247024536133,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43280977312391,
+ "grad_norm": 0.046320103108882904,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 4.204159736633301,
+ "step": 1327
+ },
+ {
+ "epoch": 18.44677137870855,
+ "grad_norm": 0.04802565276622772,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 4.182550430297852,
+ "step": 1328
+ },
+ {
+ "epoch": 18.460732984293195,
+ "grad_norm": 0.049340926110744476,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 4.163658142089844,
+ "step": 1329
+ },
+ {
+ "epoch": 18.474694589877835,
+ "grad_norm": 0.046701837331056595,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 4.170445442199707,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48865619546248,
+ "grad_norm": 0.04746273159980774,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 4.220232963562012,
+ "step": 1331
+ },
+ {
+ "epoch": 18.50261780104712,
+ "grad_norm": 0.04644489288330078,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 4.190027236938477,
+ "step": 1332
+ },
+ {
+ "epoch": 18.516579406631763,
+ "grad_norm": 0.04394262284040451,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 4.159570693969727,
+ "step": 1333
+ },
+ {
+ "epoch": 18.530541012216403,
+ "grad_norm": 0.04609063267707825,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 4.1657586097717285,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544502617801047,
+ "grad_norm": 0.0450211800634861,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 4.184865474700928,
+ "step": 1335
+ },
+ {
+ "epoch": 18.55846422338569,
+ "grad_norm": 0.0475253090262413,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 4.185857772827148,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57242582897033,
+ "grad_norm": 0.04332203418016434,
+ "learning_rate": 0.000435215814386134,
+ "loss": 4.160165309906006,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586387434554975,
+ "grad_norm": 0.0427570678293705,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 4.1727495193481445,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600349040139616,
+ "grad_norm": 0.04651174694299698,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 4.204797744750977,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61431064572426,
+ "grad_norm": 0.04324701428413391,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 4.186931610107422,
+ "step": 1340
+ },
+ {
+ "epoch": 18.6282722513089,
+ "grad_norm": 0.044644393026828766,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 4.209554195404053,
+ "step": 1341
+ },
+ {
+ "epoch": 18.642233856893544,
+ "grad_norm": 0.042982228100299835,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 4.205570220947266,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656195462478184,
+ "grad_norm": 0.04385090991854668,
+ "learning_rate": 0.000433781621332752,
+ "loss": 4.177936553955078,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670157068062828,
+ "grad_norm": 0.04771510139107704,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 4.1795806884765625,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68411867364747,
+ "grad_norm": 0.04551433399319649,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 4.192194938659668,
+ "step": 1345
+ },
+ {
+ "epoch": 18.698080279232112,
+ "grad_norm": 0.043501488864421844,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 4.177781105041504,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712041884816752,
+ "grad_norm": 0.04394034296274185,
+ "learning_rate": 0.000432823382695327,
+ "loss": 4.206354141235352,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726003490401396,
+ "grad_norm": 0.04911153391003609,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 4.205874919891357,
+ "step": 1348
+ },
+ {
+ "epoch": 18.739965095986037,
+ "grad_norm": 0.04602045193314552,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 4.1947021484375,
+ "step": 1349
+ },
+ {
+ "epoch": 18.75392670157068,
+ "grad_norm": 0.04393904283642769,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 4.188337802886963,
+ "step": 1350
+ },
+ {
+ "epoch": 18.767888307155324,
+ "grad_norm": 0.04526393488049507,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 4.175235748291016,
+ "step": 1351
+ },
+ {
+ "epoch": 18.781849912739965,
+ "grad_norm": 0.041930247098207474,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 4.210986137390137,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79581151832461,
+ "grad_norm": 0.044262565672397614,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 4.176301002502441,
+ "step": 1353
+ },
+ {
+ "epoch": 18.80977312390925,
+ "grad_norm": 0.044190581887960434,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 4.1888837814331055,
+ "step": 1354
+ },
+ {
+ "epoch": 18.823734729493893,
+ "grad_norm": 0.045371655374765396,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 4.182244777679443,
+ "step": 1355
+ },
+ {
+ "epoch": 18.837696335078533,
+ "grad_norm": 0.04302282631397247,
+ "learning_rate": 0.000430661245733797,
+ "loss": 4.189350128173828,
+ "step": 1356
+ },
+ {
+ "epoch": 18.851657940663177,
+ "grad_norm": 0.04402372986078262,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 4.233372211456299,
+ "step": 1357
+ },
+ {
+ "epoch": 18.865619546247817,
+ "grad_norm": 0.04397045075893402,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 4.21682071685791,
+ "step": 1358
+ },
+ {
+ "epoch": 18.87958115183246,
+ "grad_norm": 0.04319734871387482,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 4.204183578491211,
+ "step": 1359
+ },
+ {
+ "epoch": 18.8935427574171,
+ "grad_norm": 0.04600509628653526,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 4.206524848937988,
+ "step": 1360
+ },
+ {
+ "epoch": 18.907504363001745,
+ "grad_norm": 0.04364705830812454,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 4.213656425476074,
+ "step": 1361
+ },
+ {
+ "epoch": 18.921465968586386,
+ "grad_norm": 0.044004205614328384,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 4.1841583251953125,
+ "step": 1362
+ },
+ {
+ "epoch": 18.93542757417103,
+ "grad_norm": 0.04480595886707306,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 4.166214942932129,
+ "step": 1363
+ },
+ {
+ "epoch": 18.949389179755673,
+ "grad_norm": 0.04260887950658798,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 4.198726654052734,
+ "step": 1364
+ },
+ {
+ "epoch": 18.963350785340314,
+ "grad_norm": 0.04513927176594734,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 4.2122392654418945,
+ "step": 1365
+ },
+ {
+ "epoch": 18.977312390924958,
+ "grad_norm": 0.04286042973399162,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 4.192383766174316,
+ "step": 1366
+ },
+ {
+ "epoch": 18.991273996509598,
+ "grad_norm": 0.04287346079945564,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 4.221860408782959,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.030670644715428352,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 2.624823570251465,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.5910831689834595,
+ "eval_runtime": 60.5517,
+ "eval_samples_per_second": 40.329,
+ "eval_steps_per_second": 0.644,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013961605584644,
+ "grad_norm": 0.046196095645427704,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 4.104501724243164,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027923211169284,
+ "grad_norm": 0.05348595231771469,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 4.109098434448242,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041884816753928,
+ "grad_norm": 0.05256905406713486,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 4.115832328796387,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05584642233857,
+ "grad_norm": 0.05241430550813675,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 4.079610824584961,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069808027923212,
+ "grad_norm": 0.04983370006084442,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 4.095335006713867,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083769633507853,
+ "grad_norm": 0.04900007322430611,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 4.093306541442871,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097731239092496,
+ "grad_norm": 0.04995666816830635,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 4.111114501953125,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111692844677137,
+ "grad_norm": 0.04875681549310684,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 4.091944694519043,
+ "step": 1376
+ },
+ {
+ "epoch": 19.12565445026178,
+ "grad_norm": 0.05283811315894127,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 4.112549304962158,
+ "step": 1377
+ },
+ {
+ "epoch": 19.13961605584642,
+ "grad_norm": 0.04801992326974869,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 4.108818054199219,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153577661431065,
+ "grad_norm": 0.04751930385828018,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 4.089441299438477,
+ "step": 1379
+ },
+ {
+ "epoch": 19.167539267015705,
+ "grad_norm": 0.045653726905584335,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 4.07722282409668,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18150087260035,
+ "grad_norm": 0.045616209506988525,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 4.097578048706055,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195462478184993,
+ "grad_norm": 0.04929365590214729,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 4.082221984863281,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209424083769633,
+ "grad_norm": 0.04915834963321686,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 4.123330116271973,
+ "step": 1383
+ },
+ {
+ "epoch": 19.223385689354277,
+ "grad_norm": 0.0470132976770401,
+ "learning_rate": 0.000423881957237287,
+ "loss": 4.113927841186523,
+ "step": 1384
+ },
+ {
+ "epoch": 19.237347294938917,
+ "grad_norm": 0.04695259779691696,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 4.096814155578613,
+ "step": 1385
+ },
+ {
+ "epoch": 19.25130890052356,
+ "grad_norm": 0.04730398207902908,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 4.122525215148926,
+ "step": 1386
+ },
+ {
+ "epoch": 19.2652705061082,
+ "grad_norm": 0.050295207649469376,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 4.117993354797363,
+ "step": 1387
+ },
+ {
+ "epoch": 19.279232111692846,
+ "grad_norm": 0.05082492157816887,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 4.14082145690918,
+ "step": 1388
+ },
+ {
+ "epoch": 19.293193717277486,
+ "grad_norm": 0.04960232973098755,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 4.139887809753418,
+ "step": 1389
+ },
+ {
+ "epoch": 19.30715532286213,
+ "grad_norm": 0.04851173236966133,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 4.138492584228516,
+ "step": 1390
+ },
+ {
+ "epoch": 19.32111692844677,
+ "grad_norm": 0.047838468104600906,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 4.118064880371094,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335078534031414,
+ "grad_norm": 0.05148220434784889,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 4.136533737182617,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349040139616054,
+ "grad_norm": 0.053177181631326675,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 4.1125168800354,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363001745200698,
+ "grad_norm": 0.04861171543598175,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 4.1437153816223145,
+ "step": 1394
+ },
+ {
+ "epoch": 19.376963350785342,
+ "grad_norm": 0.05194924771785736,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 4.097813606262207,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390924956369982,
+ "grad_norm": 0.049609847366809845,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 4.068940162658691,
+ "step": 1396
+ },
+ {
+ "epoch": 19.404886561954626,
+ "grad_norm": 0.047042325139045715,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 4.121333122253418,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418848167539267,
+ "grad_norm": 0.047646693885326385,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 4.114943981170654,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43280977312391,
+ "grad_norm": 0.04676460474729538,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 4.139252185821533,
+ "step": 1399
+ },
+ {
+ "epoch": 19.44677137870855,
+ "grad_norm": 0.04905956983566284,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 4.135970115661621,
+ "step": 1400
+ },
+ {
+ "epoch": 19.460732984293195,
+ "grad_norm": 0.050025034695863724,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 4.15662956237793,
+ "step": 1401
+ },
+ {
+ "epoch": 19.474694589877835,
+ "grad_norm": 0.04740425944328308,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 4.140295505523682,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48865619546248,
+ "grad_norm": 0.046904709190130234,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 4.14083194732666,
+ "step": 1403
+ },
+ {
+ "epoch": 19.50261780104712,
+ "grad_norm": 0.048393651843070984,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 4.161823272705078,
+ "step": 1404
+ },
+ {
+ "epoch": 19.516579406631763,
+ "grad_norm": 0.04766222462058067,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 4.151817798614502,
+ "step": 1405
+ },
+ {
+ "epoch": 19.530541012216403,
+ "grad_norm": 0.04797709733247757,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 4.173160552978516,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544502617801047,
+ "grad_norm": 0.05012711137533188,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 4.161989688873291,
+ "step": 1407
+ },
+ {
+ "epoch": 19.55846422338569,
+ "grad_norm": 0.04726000875234604,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 4.136838436126709,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57242582897033,
+ "grad_norm": 0.046251531690359116,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 4.1526994705200195,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586387434554975,
+ "grad_norm": 0.047958049923181534,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 4.13817024230957,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600349040139616,
+ "grad_norm": 0.046333350241184235,
+ "learning_rate": 0.000417272427439646,
+ "loss": 4.160187721252441,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61431064572426,
+ "grad_norm": 0.0488678440451622,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 4.166190147399902,
+ "step": 1412
+ },
+ {
+ "epoch": 19.6282722513089,
+ "grad_norm": 0.04784323647618294,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 4.164414405822754,
+ "step": 1413
+ },
+ {
+ "epoch": 19.642233856893544,
+ "grad_norm": 0.05064673349261284,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 4.149814128875732,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656195462478184,
+ "grad_norm": 0.05072085186839104,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 4.167759895324707,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670157068062828,
+ "grad_norm": 0.0499395951628685,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 4.162792205810547,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68411867364747,
+ "grad_norm": 0.046766918152570724,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 4.1416521072387695,
+ "step": 1417
+ },
+ {
+ "epoch": 19.698080279232112,
+ "grad_norm": 0.04523273929953575,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 4.161751747131348,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712041884816752,
+ "grad_norm": 0.046106480062007904,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 4.166394233703613,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726003490401396,
+ "grad_norm": 0.04666793718934059,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 4.124961853027344,
+ "step": 1420
+ },
+ {
+ "epoch": 19.739965095986037,
+ "grad_norm": 0.047407012432813644,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 4.1794328689575195,
+ "step": 1421
+ },
+ {
+ "epoch": 19.75392670157068,
+ "grad_norm": 0.04690873995423317,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 4.160114765167236,
+ "step": 1422
+ },
+ {
+ "epoch": 19.767888307155324,
+ "grad_norm": 0.04589436575770378,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 4.137969970703125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.781849912739965,
+ "grad_norm": 0.04641961678862572,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 4.161533355712891,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79581151832461,
+ "grad_norm": 0.045799050480127335,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 4.187039375305176,
+ "step": 1425
+ },
+ {
+ "epoch": 19.80977312390925,
+ "grad_norm": 0.04881894588470459,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 4.158107757568359,
+ "step": 1426
+ },
+ {
+ "epoch": 19.823734729493893,
+ "grad_norm": 0.04506978392601013,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 4.143360137939453,
+ "step": 1427
+ },
+ {
+ "epoch": 19.837696335078533,
+ "grad_norm": 0.04504861310124397,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 4.174564361572266,
+ "step": 1428
+ },
+ {
+ "epoch": 19.851657940663177,
+ "grad_norm": 0.04639774188399315,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 4.162899971008301,
+ "step": 1429
+ },
+ {
+ "epoch": 19.865619546247817,
+ "grad_norm": 0.047444432973861694,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 4.143390655517578,
+ "step": 1430
+ },
+ {
+ "epoch": 19.87958115183246,
+ "grad_norm": 0.04461613669991493,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 4.171284198760986,
+ "step": 1431
+ },
+ {
+ "epoch": 19.8935427574171,
+ "grad_norm": 0.046383798122406006,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 4.14809513092041,
+ "step": 1432
+ },
+ {
+ "epoch": 19.907504363001745,
+ "grad_norm": 0.04642366245388985,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 4.178903579711914,
+ "step": 1433
+ },
+ {
+ "epoch": 19.921465968586386,
+ "grad_norm": 0.0451771542429924,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 4.171825408935547,
+ "step": 1434
+ },
+ {
+ "epoch": 19.93542757417103,
+ "grad_norm": 0.04767893627285957,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 4.169270038604736,
+ "step": 1435
+ },
+ {
+ "epoch": 19.949389179755673,
+ "grad_norm": 0.04673412814736366,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 4.175506591796875,
+ "step": 1436
+ },
+ {
+ "epoch": 19.963350785340314,
+ "grad_norm": 0.04465145245194435,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 4.149707794189453,
+ "step": 1437
+ },
+ {
+ "epoch": 19.977312390924958,
+ "grad_norm": 0.04734131321310997,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 4.166044235229492,
+ "step": 1438
+ },
+ {
+ "epoch": 19.991273996509598,
+ "grad_norm": 0.04706178605556488,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 4.152081489562988,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.031556785106658936,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 2.60109543800354,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013961605584644,
+ "grad_norm": 0.05152537673711777,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 4.084830284118652,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027923211169284,
+ "grad_norm": 0.05918274074792862,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 4.045302391052246,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041884816753928,
+ "grad_norm": 0.05093073844909668,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 4.070842742919922,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05584642233857,
+ "grad_norm": 0.05095985531806946,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 4.083273410797119,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069808027923212,
+ "grad_norm": 0.05502687394618988,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 4.072626113891602,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083769633507853,
+ "grad_norm": 0.05037364363670349,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 4.075478553771973,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097731239092496,
+ "grad_norm": 0.04983789846301079,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 4.066207408905029,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111692844677137,
+ "grad_norm": 0.049478255212306976,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 4.069969654083252,
+ "step": 1448
+ },
+ {
+ "epoch": 20.12565445026178,
+ "grad_norm": 0.051651593297719955,
+ "learning_rate": 0.000407857329622967,
+ "loss": 4.050713062286377,
+ "step": 1449
+ },
+ {
+ "epoch": 20.13961605584642,
+ "grad_norm": 0.05436555668711662,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 4.069849967956543,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153577661431065,
+ "grad_norm": 0.055190350860357285,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 4.081624984741211,
+ "step": 1451
+ },
+ {
+ "epoch": 20.167539267015705,
+ "grad_norm": 0.053246211260557175,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 4.076076507568359,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18150087260035,
+ "grad_norm": 0.051757775247097015,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 4.049139022827148,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195462478184993,
+ "grad_norm": 0.05064203217625618,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 4.068076133728027,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209424083769633,
+ "grad_norm": 0.05114085227251053,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 4.092901229858398,
+ "step": 1455
+ },
+ {
+ "epoch": 20.223385689354277,
+ "grad_norm": 0.047997843474149704,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 4.077084541320801,
+ "step": 1456
+ },
+ {
+ "epoch": 20.237347294938917,
+ "grad_norm": 0.04880451038479805,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 4.059187889099121,
+ "step": 1457
+ },
+ {
+ "epoch": 20.25130890052356,
+ "grad_norm": 0.05036284029483795,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 4.102115154266357,
+ "step": 1458
+ },
+ {
+ "epoch": 20.2652705061082,
+ "grad_norm": 0.04916979745030403,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 4.098356246948242,
+ "step": 1459
+ },
+ {
+ "epoch": 20.279232111692846,
+ "grad_norm": 0.04713984206318855,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 4.057218074798584,
+ "step": 1460
+ },
+ {
+ "epoch": 20.293193717277486,
+ "grad_norm": 0.04982447996735573,
+ "learning_rate": 0.000404858275823277,
+ "loss": 4.070363998413086,
+ "step": 1461
+ },
+ {
+ "epoch": 20.30715532286213,
+ "grad_norm": 0.0475616492331028,
+ "learning_rate": 0.000404607816444578,
+ "loss": 4.074314117431641,
+ "step": 1462
+ },
+ {
+ "epoch": 20.32111692844677,
+ "grad_norm": 0.046537093818187714,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 4.079501152038574,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335078534031414,
+ "grad_norm": 0.04889644682407379,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 4.0731425285339355,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349040139616054,
+ "grad_norm": 0.05095166340470314,
+ "learning_rate": 0.000403855947934478,
+ "loss": 4.08765983581543,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363001745200698,
+ "grad_norm": 0.049677830189466476,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 4.128052711486816,
+ "step": 1466
+ },
+ {
+ "epoch": 20.376963350785342,
+ "grad_norm": 0.05221942439675331,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 4.094407081604004,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390924956369982,
+ "grad_norm": 0.052042894065380096,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 4.088282585144043,
+ "step": 1468
+ },
+ {
+ "epoch": 20.404886561954626,
+ "grad_norm": 0.04822041094303131,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 4.1070990562438965,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418848167539267,
+ "grad_norm": 0.047207772731781006,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 4.102565765380859,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43280977312391,
+ "grad_norm": 0.049633774906396866,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 4.097216606140137,
+ "step": 1471
+ },
+ {
+ "epoch": 20.44677137870855,
+ "grad_norm": 0.05135185644030571,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 4.095193862915039,
+ "step": 1472
+ },
+ {
+ "epoch": 20.460732984293195,
+ "grad_norm": 0.05215826258063316,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 4.085207939147949,
+ "step": 1473
+ },
+ {
+ "epoch": 20.474694589877835,
+ "grad_norm": 0.050653353333473206,
+ "learning_rate": 0.000401595976318565,
+ "loss": 4.085142612457275,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48865619546248,
+ "grad_norm": 0.04795852303504944,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 4.1158270835876465,
+ "step": 1475
+ },
+ {
+ "epoch": 20.50261780104712,
+ "grad_norm": 0.047416940331459045,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 4.096989154815674,
+ "step": 1476
+ },
+ {
+ "epoch": 20.516579406631763,
+ "grad_norm": 0.049753669649362564,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 4.118251800537109,
+ "step": 1477
+ },
+ {
+ "epoch": 20.530541012216403,
+ "grad_norm": 0.04889055714011192,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 4.081287384033203,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544502617801047,
+ "grad_norm": 0.04857790097594261,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 4.084896564483643,
+ "step": 1479
+ },
+ {
+ "epoch": 20.55846422338569,
+ "grad_norm": 0.047250498086214066,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 4.113656997680664,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57242582897033,
+ "grad_norm": 0.048608481884002686,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 4.113809108734131,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586387434554975,
+ "grad_norm": 0.048189081251621246,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 4.113949775695801,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600349040139616,
+ "grad_norm": 0.049929242581129074,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 4.104221820831299,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61431064572426,
+ "grad_norm": 0.048482850193977356,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 4.091039180755615,
+ "step": 1484
+ },
+ {
+ "epoch": 20.6282722513089,
+ "grad_norm": 0.04810645803809166,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 4.130088806152344,
+ "step": 1485
+ },
+ {
+ "epoch": 20.642233856893544,
+ "grad_norm": 0.04745674133300781,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 4.106161117553711,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656195462478184,
+ "grad_norm": 0.048513609915971756,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 4.119222164154053,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670157068062828,
+ "grad_norm": 0.04506325349211693,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 4.115543842315674,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68411867364747,
+ "grad_norm": 0.04726741835474968,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 4.084345817565918,
+ "step": 1489
+ },
+ {
+ "epoch": 20.698080279232112,
+ "grad_norm": 0.04873545467853546,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 4.116580009460449,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712041884816752,
+ "grad_norm": 0.051252465695142746,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 4.129733085632324,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726003490401396,
+ "grad_norm": 0.0489567406475544,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 4.103477954864502,
+ "step": 1492
+ },
+ {
+ "epoch": 20.739965095986037,
+ "grad_norm": 0.04539056867361069,
+ "learning_rate": 0.000396803974909638,
+ "loss": 4.1327667236328125,
+ "step": 1493
+ },
+ {
+ "epoch": 20.75392670157068,
+ "grad_norm": 0.04954095557332039,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 4.084133625030518,
+ "step": 1494
+ },
+ {
+ "epoch": 20.767888307155324,
+ "grad_norm": 0.05041591823101044,
+ "learning_rate": 0.000396297942393265,
+ "loss": 4.12288761138916,
+ "step": 1495
+ },
+ {
+ "epoch": 20.781849912739965,
+ "grad_norm": 0.046549517661333084,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 4.1259918212890625,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79581151832461,
+ "grad_norm": 0.04704999923706055,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 4.102563858032227,
+ "step": 1497
+ },
+ {
+ "epoch": 20.80977312390925,
+ "grad_norm": 0.05025320500135422,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 4.135492324829102,
+ "step": 1498
+ },
+ {
+ "epoch": 20.823734729493893,
+ "grad_norm": 0.04559441655874252,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 4.118548393249512,
+ "step": 1499
+ },
+ {
+ "epoch": 20.837696335078533,
+ "grad_norm": 0.04913092032074928,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 4.129302024841309,
+ "step": 1500
+ },
+ {
+ "epoch": 20.851657940663177,
+ "grad_norm": 0.04846089705824852,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 4.166059970855713,
+ "step": 1501
+ },
+ {
+ "epoch": 20.865619546247817,
+ "grad_norm": 0.04710553213953972,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 4.104712009429932,
+ "step": 1502
+ },
+ {
+ "epoch": 20.87958115183246,
+ "grad_norm": 0.04896444454789162,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 4.140282154083252,
+ "step": 1503
+ },
+ {
+ "epoch": 20.8935427574171,
+ "grad_norm": 0.0469740591943264,
+ "learning_rate": 0.000394017098798633,
+ "loss": 4.125306129455566,
+ "step": 1504
+ },
+ {
+ "epoch": 20.907504363001745,
+ "grad_norm": 0.04638523980975151,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 4.133161544799805,
+ "step": 1505
+ },
+ {
+ "epoch": 20.921465968586386,
+ "grad_norm": 0.04814442619681358,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 4.093092918395996,
+ "step": 1506
+ },
+ {
+ "epoch": 20.93542757417103,
+ "grad_norm": 0.049911532551050186,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 4.1001715660095215,
+ "step": 1507
+ },
+ {
+ "epoch": 20.949389179755673,
+ "grad_norm": 0.0446670837700367,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 4.140350341796875,
+ "step": 1508
+ },
+ {
+ "epoch": 20.963350785340314,
+ "grad_norm": 0.04789397493004799,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 4.163155555725098,
+ "step": 1509
+ },
+ {
+ "epoch": 20.977312390924958,
+ "grad_norm": 0.050148315727710724,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 4.118380069732666,
+ "step": 1510
+ },
+ {
+ "epoch": 20.991273996509598,
+ "grad_norm": 0.047446198761463165,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 4.116507530212402,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.03415544703602791,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 2.5738282203674316,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.5933796167373657,
+ "eval_runtime": 60.4011,
+ "eval_samples_per_second": 40.43,
+ "eval_steps_per_second": 0.646,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013961605584644,
+ "grad_norm": 0.052276045083999634,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 4.0301923751831055,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027923211169284,
+ "grad_norm": 0.06514342874288559,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 4.031280517578125,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041884816753928,
+ "grad_norm": 0.05948134884238243,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 4.060680389404297,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05584642233857,
+ "grad_norm": 0.05251640826463699,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 4.0218400955200195,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069808027923212,
+ "grad_norm": 0.05426032841205597,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 4.026187896728516,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083769633507853,
+ "grad_norm": 0.052583567798137665,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 4.016767501831055,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097731239092496,
+ "grad_norm": 0.04966135695576668,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 4.043508529663086,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111692844677137,
+ "grad_norm": 0.053838398307561874,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 4.054832458496094,
+ "step": 1520
+ },
+ {
+ "epoch": 21.12565445026178,
+ "grad_norm": 0.05304184556007385,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 4.03280782699585,
+ "step": 1521
+ },
+ {
+ "epoch": 21.13961605584642,
+ "grad_norm": 0.05252664163708687,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 4.039841175079346,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153577661431065,
+ "grad_norm": 0.054593686014413834,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 4.048308372497559,
+ "step": 1523
+ },
+ {
+ "epoch": 21.167539267015705,
+ "grad_norm": 0.050722211599349976,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 4.051860332489014,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18150087260035,
+ "grad_norm": 0.05248095095157623,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 4.056846618652344,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195462478184993,
+ "grad_norm": 0.05166107788681984,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 4.045178413391113,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209424083769633,
+ "grad_norm": 0.04987792298197746,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 4.014827728271484,
+ "step": 1527
+ },
+ {
+ "epoch": 21.223385689354277,
+ "grad_norm": 0.05316167324781418,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 4.033143997192383,
+ "step": 1528
+ },
+ {
+ "epoch": 21.237347294938917,
+ "grad_norm": 0.05228257551789284,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 4.033842086791992,
+ "step": 1529
+ },
+ {
+ "epoch": 21.25130890052356,
+ "grad_norm": 0.049349382519721985,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 4.025982856750488,
+ "step": 1530
+ },
+ {
+ "epoch": 21.2652705061082,
+ "grad_norm": 0.055104926228523254,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 4.059151649475098,
+ "step": 1531
+ },
+ {
+ "epoch": 21.279232111692846,
+ "grad_norm": 0.052629340440034866,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 4.032320976257324,
+ "step": 1532
+ },
+ {
+ "epoch": 21.293193717277486,
+ "grad_norm": 0.05396415665745735,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 4.0425567626953125,
+ "step": 1533
+ },
+ {
+ "epoch": 21.30715532286213,
+ "grad_norm": 0.054417286068201065,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 4.03759241104126,
+ "step": 1534
+ },
+ {
+ "epoch": 21.32111692844677,
+ "grad_norm": 0.061248697340488434,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 4.0441083908081055,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335078534031414,
+ "grad_norm": 0.05934970825910568,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 4.050718307495117,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349040139616054,
+ "grad_norm": 0.052805181592702866,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 4.069460868835449,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363001745200698,
+ "grad_norm": 0.052820611745119095,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 4.063657283782959,
+ "step": 1538
+ },
+ {
+ "epoch": 21.376963350785342,
+ "grad_norm": 0.05043052136898041,
+ "learning_rate": 0.00038509205478744,
+ "loss": 4.029552459716797,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390924956369982,
+ "grad_norm": 0.04828709363937378,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 4.05656623840332,
+ "step": 1540
+ },
+ {
+ "epoch": 21.404886561954626,
+ "grad_norm": 0.049760788679122925,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 4.066338539123535,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418848167539267,
+ "grad_norm": 0.05037159472703934,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 4.06976842880249,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43280977312391,
+ "grad_norm": 0.05276317894458771,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 4.062975883483887,
+ "step": 1543
+ },
+ {
+ "epoch": 21.44677137870855,
+ "grad_norm": 0.05571872368454933,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 4.065622329711914,
+ "step": 1544
+ },
+ {
+ "epoch": 21.460732984293195,
+ "grad_norm": 0.05438469722867012,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 4.074963569641113,
+ "step": 1545
+ },
+ {
+ "epoch": 21.474694589877835,
+ "grad_norm": 0.053415946662425995,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 4.080055236816406,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48865619546248,
+ "grad_norm": 0.05216296389698982,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 4.091355800628662,
+ "step": 1547
+ },
+ {
+ "epoch": 21.50261780104712,
+ "grad_norm": 0.051212459802627563,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 4.058528900146484,
+ "step": 1548
+ },
+ {
+ "epoch": 21.516579406631763,
+ "grad_norm": 0.05033959075808525,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 4.064615726470947,
+ "step": 1549
+ },
+ {
+ "epoch": 21.530541012216403,
+ "grad_norm": 0.04665188118815422,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 4.059698104858398,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544502617801047,
+ "grad_norm": 0.050221361219882965,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 4.068614959716797,
+ "step": 1551
+ },
+ {
+ "epoch": 21.55846422338569,
+ "grad_norm": 0.05089551955461502,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 4.0833892822265625,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57242582897033,
+ "grad_norm": 0.051966603845357895,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 4.083354949951172,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586387434554975,
+ "grad_norm": 0.04981927201151848,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 4.083340644836426,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600349040139616,
+ "grad_norm": 0.05141765624284744,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 4.066590309143066,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61431064572426,
+ "grad_norm": 0.048985838890075684,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 4.088237762451172,
+ "step": 1556
+ },
+ {
+ "epoch": 21.6282722513089,
+ "grad_norm": 0.05013054236769676,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 4.081377983093262,
+ "step": 1557
+ },
+ {
+ "epoch": 21.642233856893544,
+ "grad_norm": 0.05170975625514984,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 4.054303169250488,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656195462478184,
+ "grad_norm": 0.048257142305374146,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 4.103672504425049,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670157068062828,
+ "grad_norm": 0.049942679703235626,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 4.029951572418213,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68411867364747,
+ "grad_norm": 0.05083315819501877,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 4.048798084259033,
+ "step": 1561
+ },
+ {
+ "epoch": 21.698080279232112,
+ "grad_norm": 0.04984128847718239,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 4.079655647277832,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712041884816752,
+ "grad_norm": 0.04985307902097702,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 4.071877956390381,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726003490401396,
+ "grad_norm": 0.051137227565050125,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 4.076539993286133,
+ "step": 1564
+ },
+ {
+ "epoch": 21.739965095986037,
+ "grad_norm": 0.04968905821442604,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 4.084498882293701,
+ "step": 1565
+ },
+ {
+ "epoch": 21.75392670157068,
+ "grad_norm": 0.049119509756565094,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 4.066773414611816,
+ "step": 1566
+ },
+ {
+ "epoch": 21.767888307155324,
+ "grad_norm": 0.05107425898313522,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 4.065309047698975,
+ "step": 1567
+ },
+ {
+ "epoch": 21.781849912739965,
+ "grad_norm": 0.049012504518032074,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 4.062110900878906,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79581151832461,
+ "grad_norm": 0.04972011595964432,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 4.0630903244018555,
+ "step": 1569
+ },
+ {
+ "epoch": 21.80977312390925,
+ "grad_norm": 0.04822700098156929,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 4.106127738952637,
+ "step": 1570
+ },
+ {
+ "epoch": 21.823734729493893,
+ "grad_norm": 0.04808267205953598,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 4.085200309753418,
+ "step": 1571
+ },
+ {
+ "epoch": 21.837696335078533,
+ "grad_norm": 0.048644762486219406,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 4.0856170654296875,
+ "step": 1572
+ },
+ {
+ "epoch": 21.851657940663177,
+ "grad_norm": 0.05232283100485802,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 4.126387119293213,
+ "step": 1573
+ },
+ {
+ "epoch": 21.865619546247817,
+ "grad_norm": 0.05095643922686577,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 4.108254432678223,
+ "step": 1574
+ },
+ {
+ "epoch": 21.87958115183246,
+ "grad_norm": 0.045972101390361786,
+ "learning_rate": 0.000375827577564042,
+ "loss": 4.058515548706055,
+ "step": 1575
+ },
+ {
+ "epoch": 21.8935427574171,
+ "grad_norm": 0.04974910616874695,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 4.089197158813477,
+ "step": 1576
+ },
+ {
+ "epoch": 21.907504363001745,
+ "grad_norm": 0.04890159144997597,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 4.1078410148620605,
+ "step": 1577
+ },
+ {
+ "epoch": 21.921465968586386,
+ "grad_norm": 0.04738525673747063,
+ "learning_rate": 0.000375051971546236,
+ "loss": 4.1219282150268555,
+ "step": 1578
+ },
+ {
+ "epoch": 21.93542757417103,
+ "grad_norm": 0.04713096842169762,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 4.10295295715332,
+ "step": 1579
+ },
+ {
+ "epoch": 21.949389179755673,
+ "grad_norm": 0.04834369197487831,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 4.102747917175293,
+ "step": 1580
+ },
+ {
+ "epoch": 21.963350785340314,
+ "grad_norm": 0.047073688358068466,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 4.08024787902832,
+ "step": 1581
+ },
+ {
+ "epoch": 21.977312390924958,
+ "grad_norm": 0.04568164795637131,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 4.078930854797363,
+ "step": 1582
+ },
+ {
+ "epoch": 21.991273996509598,
+ "grad_norm": 0.04822581633925438,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 4.089601516723633,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.03507270663976669,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 2.5586814880371094,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.595272958278656,
+ "eval_runtime": 60.7797,
+ "eval_samples_per_second": 40.178,
+ "eval_steps_per_second": 0.642,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013961605584644,
+ "grad_norm": 0.05232533439993858,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 4.011311054229736,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027923211169284,
+ "grad_norm": 0.058515798300504684,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 4.0049638748168945,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041884816753928,
+ "grad_norm": 0.055082619190216064,
+ "learning_rate": 0.000372722041257003,
+ "loss": 3.9904184341430664,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05584642233857,
+ "grad_norm": 0.05213981494307518,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 4.003454685211182,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069808027923212,
+ "grad_norm": 0.051619648933410645,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 3.9900388717651367,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083769633507853,
+ "grad_norm": 0.055212799459695816,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 3.973494052886963,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097731239092496,
+ "grad_norm": 0.05262092128396034,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 4.00933837890625,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111692844677137,
+ "grad_norm": 0.05012505501508713,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 4.01861047744751,
+ "step": 1592
+ },
+ {
+ "epoch": 22.12565445026178,
+ "grad_norm": 0.049657098948955536,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 4.011143684387207,
+ "step": 1593
+ },
+ {
+ "epoch": 22.13961605584642,
+ "grad_norm": 0.05187768116593361,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 3.9922969341278076,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153577661431065,
+ "grad_norm": 0.05053112655878067,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 4.007021903991699,
+ "step": 1595
+ },
+ {
+ "epoch": 22.167539267015705,
+ "grad_norm": 0.048542529344558716,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 3.9966416358947754,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18150087260035,
+ "grad_norm": 0.050972841680049896,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 3.9803225994110107,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195462478184993,
+ "grad_norm": 0.05001165345311165,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 4.0140509605407715,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209424083769633,
+ "grad_norm": 0.05100536718964577,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 3.9906058311462402,
+ "step": 1599
+ },
+ {
+ "epoch": 22.223385689354277,
+ "grad_norm": 0.052977193146944046,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 4.012383460998535,
+ "step": 1600
+ },
+ {
+ "epoch": 22.237347294938917,
+ "grad_norm": 0.051219623535871506,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 3.982931613922119,
+ "step": 1601
+ },
+ {
+ "epoch": 22.25130890052356,
+ "grad_norm": 0.05275029316544533,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 4.032669544219971,
+ "step": 1602
+ },
+ {
+ "epoch": 22.2652705061082,
+ "grad_norm": 0.04781945049762726,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 3.9933102130889893,
+ "step": 1603
+ },
+ {
+ "epoch": 22.279232111692846,
+ "grad_norm": 0.054087597876787186,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 4.0075201988220215,
+ "step": 1604
+ },
+ {
+ "epoch": 22.293193717277486,
+ "grad_norm": 0.05485915020108223,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 4.004299640655518,
+ "step": 1605
+ },
+ {
+ "epoch": 22.30715532286213,
+ "grad_norm": 0.05148882791399956,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 4.044161796569824,
+ "step": 1606
+ },
+ {
+ "epoch": 22.32111692844677,
+ "grad_norm": 0.051912397146224976,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 3.9777069091796875,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335078534031414,
+ "grad_norm": 0.052630942314863205,
+ "learning_rate": 0.000367268105959126,
+ "loss": 4.029580116271973,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349040139616054,
+ "grad_norm": 0.0512809231877327,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 4.022092342376709,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363001745200698,
+ "grad_norm": 0.05094102770090103,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 4.018102169036865,
+ "step": 1610
+ },
+ {
+ "epoch": 22.376963350785342,
+ "grad_norm": 0.04972759634256363,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 4.027868270874023,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390924956369982,
+ "grad_norm": 0.05072534456849098,
+ "learning_rate": 0.000366226612810024,
+ "loss": 3.995988368988037,
+ "step": 1612
+ },
+ {
+ "epoch": 22.404886561954626,
+ "grad_norm": 0.050623781979084015,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 4.019711494445801,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418848167539267,
+ "grad_norm": 0.05035007372498512,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 4.007514953613281,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43280977312391,
+ "grad_norm": 0.054426416754722595,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 4.0167131423950195,
+ "step": 1615
+ },
+ {
+ "epoch": 22.44677137870855,
+ "grad_norm": 0.05365690588951111,
+ "learning_rate": 0.000365184304613104,
+ "loss": 4.0313920974731445,
+ "step": 1616
+ },
+ {
+ "epoch": 22.460732984293195,
+ "grad_norm": 0.0509161539375782,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 4.022063255310059,
+ "step": 1617
+ },
+ {
+ "epoch": 22.474694589877835,
+ "grad_norm": 0.05309660732746124,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 4.017317771911621,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48865619546248,
+ "grad_norm": 0.0488961897790432,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 4.028288841247559,
+ "step": 1619
+ },
+ {
+ "epoch": 22.50261780104712,
+ "grad_norm": 0.04982095956802368,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 4.007115840911865,
+ "step": 1620
+ },
+ {
+ "epoch": 22.516579406631763,
+ "grad_norm": 0.05136589705944061,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 4.028022766113281,
+ "step": 1621
+ },
+ {
+ "epoch": 22.530541012216403,
+ "grad_norm": 0.04865279048681259,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 4.015928268432617,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544502617801047,
+ "grad_norm": 0.05131891369819641,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 4.041997909545898,
+ "step": 1623
+ },
+ {
+ "epoch": 22.55846422338569,
+ "grad_norm": 0.05063068866729736,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 4.042117118835449,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57242582897033,
+ "grad_norm": 0.049348752945661545,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 4.012609481811523,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586387434554975,
+ "grad_norm": 0.05123874917626381,
+ "learning_rate": 0.000362575056487632,
+ "loss": 4.03358793258667,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600349040139616,
+ "grad_norm": 0.04921206086874008,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 4.039857864379883,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61431064572426,
+ "grad_norm": 0.05011922866106033,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 4.015835762023926,
+ "step": 1628
+ },
+ {
+ "epoch": 22.6282722513089,
+ "grad_norm": 0.049950532615184784,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 4.068299293518066,
+ "step": 1629
+ },
+ {
+ "epoch": 22.642233856893544,
+ "grad_norm": 0.04962172359228134,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 4.03205680847168,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656195462478184,
+ "grad_norm": 0.05059773847460747,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 4.065027236938477,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670157068062828,
+ "grad_norm": 0.04986768960952759,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 4.01540470123291,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68411867364747,
+ "grad_norm": 0.050884976983070374,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 4.060809135437012,
+ "step": 1633
+ },
+ {
+ "epoch": 22.698080279232112,
+ "grad_norm": 0.052478861063718796,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 4.088409423828125,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712041884816752,
+ "grad_norm": 0.05118529871106148,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 4.071645736694336,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726003490401396,
+ "grad_norm": 0.05059075728058815,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 4.028438568115234,
+ "step": 1636
+ },
+ {
+ "epoch": 22.739965095986037,
+ "grad_norm": 0.049222562462091446,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 4.02166223526001,
+ "step": 1637
+ },
+ {
+ "epoch": 22.75392670157068,
+ "grad_norm": 0.049673549830913544,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 4.0511860847473145,
+ "step": 1638
+ },
+ {
+ "epoch": 22.767888307155324,
+ "grad_norm": 0.04954935610294342,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 4.036063194274902,
+ "step": 1639
+ },
+ {
+ "epoch": 22.781849912739965,
+ "grad_norm": 0.05110877379775047,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 4.081101417541504,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79581151832461,
+ "grad_norm": 0.049009695649147034,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 4.042314529418945,
+ "step": 1641
+ },
+ {
+ "epoch": 22.80977312390925,
+ "grad_norm": 0.0496281161904335,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 4.028315544128418,
+ "step": 1642
+ },
+ {
+ "epoch": 22.823734729493893,
+ "grad_norm": 0.04931195452809334,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 4.018457412719727,
+ "step": 1643
+ },
+ {
+ "epoch": 22.837696335078533,
+ "grad_norm": 0.05006880313158035,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 4.0709991455078125,
+ "step": 1644
+ },
+ {
+ "epoch": 22.851657940663177,
+ "grad_norm": 0.04956946149468422,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 4.065585136413574,
+ "step": 1645
+ },
+ {
+ "epoch": 22.865619546247817,
+ "grad_norm": 0.048182398080825806,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 4.0522589683532715,
+ "step": 1646
+ },
+ {
+ "epoch": 22.87958115183246,
+ "grad_norm": 0.04964807257056236,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 4.06989860534668,
+ "step": 1647
+ },
+ {
+ "epoch": 22.8935427574171,
+ "grad_norm": 0.049299128353595734,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 4.082858085632324,
+ "step": 1648
+ },
+ {
+ "epoch": 22.907504363001745,
+ "grad_norm": 0.04893320053815842,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 4.047144889831543,
+ "step": 1649
+ },
+ {
+ "epoch": 22.921465968586386,
+ "grad_norm": 0.04772866889834404,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 4.043060302734375,
+ "step": 1650
+ },
+ {
+ "epoch": 22.93542757417103,
+ "grad_norm": 0.049417730420827866,
+ "learning_rate": 0.000356031397755195,
+ "loss": 4.031186580657959,
+ "step": 1651
+ },
+ {
+ "epoch": 22.949389179755673,
+ "grad_norm": 0.049944665282964706,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 4.074260711669922,
+ "step": 1652
+ },
+ {
+ "epoch": 22.963350785340314,
+ "grad_norm": 0.04856271669268608,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 4.054103851318359,
+ "step": 1653
+ },
+ {
+ "epoch": 22.977312390924958,
+ "grad_norm": 0.0470513179898262,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 4.041568756103516,
+ "step": 1654
+ },
+ {
+ "epoch": 22.991273996509598,
+ "grad_norm": 0.049018729478120804,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 4.066161632537842,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.034226156771183014,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 2.5147223472595215,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.5970026850700378,
+ "eval_runtime": 60.0787,
+ "eval_samples_per_second": 40.647,
+ "eval_steps_per_second": 0.649,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013961605584644,
+ "grad_norm": 0.05091085657477379,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 3.95741868019104,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027923211169284,
+ "grad_norm": 0.05874285474419594,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 3.966327667236328,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041884816753928,
+ "grad_norm": 0.05394266918301582,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.937232732772827,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05584642233857,
+ "grad_norm": 0.05120540037751198,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 3.9637370109558105,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069808027923212,
+ "grad_norm": 0.05310065299272537,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 3.998631477355957,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083769633507853,
+ "grad_norm": 0.05127466097474098,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.949101448059082,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097731239092496,
+ "grad_norm": 0.05201594531536102,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 3.986267566680908,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111692844677137,
+ "grad_norm": 0.051927369087934494,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 3.9604103565216064,
+ "step": 1664
+ },
+ {
+ "epoch": 23.12565445026178,
+ "grad_norm": 0.05125141143798828,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.9339547157287598,
+ "step": 1665
+ },
+ {
+ "epoch": 23.13961605584642,
+ "grad_norm": 0.05241088196635246,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 3.971261978149414,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153577661431065,
+ "grad_norm": 0.050563592463731766,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.9450082778930664,
+ "step": 1667
+ },
+ {
+ "epoch": 23.167539267015705,
+ "grad_norm": 0.05204763635993004,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 3.9696707725524902,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18150087260035,
+ "grad_norm": 0.05518610402941704,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 3.9816057682037354,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195462478184993,
+ "grad_norm": 0.05336211621761322,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.932252883911133,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209424083769633,
+ "grad_norm": 0.052204642444849014,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 3.9818131923675537,
+ "step": 1671
+ },
+ {
+ "epoch": 23.223385689354277,
+ "grad_norm": 0.05299545079469681,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.9488978385925293,
+ "step": 1672
+ },
+ {
+ "epoch": 23.237347294938917,
+ "grad_norm": 0.05143733322620392,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 4.00959587097168,
+ "step": 1673
+ },
+ {
+ "epoch": 23.25130890052356,
+ "grad_norm": 0.05186324194073677,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.9529542922973633,
+ "step": 1674
+ },
+ {
+ "epoch": 23.2652705061082,
+ "grad_norm": 0.05334410071372986,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 3.96693754196167,
+ "step": 1675
+ },
+ {
+ "epoch": 23.279232111692846,
+ "grad_norm": 0.05345796048641205,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 3.9674174785614014,
+ "step": 1676
+ },
+ {
+ "epoch": 23.293193717277486,
+ "grad_norm": 0.050988901406526566,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 3.9698638916015625,
+ "step": 1677
+ },
+ {
+ "epoch": 23.30715532286213,
+ "grad_norm": 0.056467678397893906,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 3.9645955562591553,
+ "step": 1678
+ },
+ {
+ "epoch": 23.32111692844677,
+ "grad_norm": 0.05554647371172905,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 4.0036091804504395,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335078534031414,
+ "grad_norm": 0.05607818067073822,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 3.973773717880249,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349040139616054,
+ "grad_norm": 0.05734477937221527,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 3.987692356109619,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363001745200698,
+ "grad_norm": 0.052175361663103104,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 3.9761803150177,
+ "step": 1682
+ },
+ {
+ "epoch": 23.376963350785342,
+ "grad_norm": 0.053368814289569855,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 3.965020179748535,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390924956369982,
+ "grad_norm": 0.05505364388227463,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 3.987494468688965,
+ "step": 1684
+ },
+ {
+ "epoch": 23.404886561954626,
+ "grad_norm": 0.0524514764547348,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 3.9624319076538086,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418848167539267,
+ "grad_norm": 0.05681568384170532,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 4.008373737335205,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43280977312391,
+ "grad_norm": 0.05344909429550171,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 3.9840354919433594,
+ "step": 1687
+ },
+ {
+ "epoch": 23.44677137870855,
+ "grad_norm": 0.052657466381788254,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 3.992262363433838,
+ "step": 1688
+ },
+ {
+ "epoch": 23.460732984293195,
+ "grad_norm": 0.05526747554540634,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 3.97529935836792,
+ "step": 1689
+ },
+ {
+ "epoch": 23.474694589877835,
+ "grad_norm": 0.05492551624774933,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 4.004708766937256,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48865619546248,
+ "grad_norm": 0.053742099553346634,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 3.994016647338867,
+ "step": 1691
+ },
+ {
+ "epoch": 23.50261780104712,
+ "grad_norm": 0.05363722890615463,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 3.9942378997802734,
+ "step": 1692
+ },
+ {
+ "epoch": 23.516579406631763,
+ "grad_norm": 0.05167972669005394,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 3.998253345489502,
+ "step": 1693
+ },
+ {
+ "epoch": 23.530541012216403,
+ "grad_norm": 0.0518590584397316,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 3.9850645065307617,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544502617801047,
+ "grad_norm": 0.05521136894822121,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 4.02669620513916,
+ "step": 1695
+ },
+ {
+ "epoch": 23.55846422338569,
+ "grad_norm": 0.054986029863357544,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 4.022308826446533,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57242582897033,
+ "grad_norm": 0.05405082181096077,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 4.021487236022949,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586387434554975,
+ "grad_norm": 0.051257841289043427,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 3.9724135398864746,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600349040139616,
+ "grad_norm": 0.05109119415283203,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 3.978914499282837,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61431064572426,
+ "grad_norm": 0.05303420498967171,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 4.027509689331055,
+ "step": 1700
+ },
+ {
+ "epoch": 23.6282722513089,
+ "grad_norm": 0.055543459951877594,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 3.9750852584838867,
+ "step": 1701
+ },
+ {
+ "epoch": 23.642233856893544,
+ "grad_norm": 0.05068540200591087,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 4.029600143432617,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656195462478184,
+ "grad_norm": 0.04991193488240242,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 3.9901509284973145,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670157068062828,
+ "grad_norm": 0.05403713881969452,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 4.017679214477539,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68411867364747,
+ "grad_norm": 0.05203867331147194,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 4.008406639099121,
+ "step": 1705
+ },
+ {
+ "epoch": 23.698080279232112,
+ "grad_norm": 0.05106597766280174,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 4.009983062744141,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712041884816752,
+ "grad_norm": 0.04976251348853111,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 4.007230281829834,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726003490401396,
+ "grad_norm": 0.05199650675058365,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 3.9902141094207764,
+ "step": 1708
+ },
+ {
+ "epoch": 23.739965095986037,
+ "grad_norm": 0.05285118520259857,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 3.9966492652893066,
+ "step": 1709
+ },
+ {
+ "epoch": 23.75392670157068,
+ "grad_norm": 0.051937371492385864,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 3.962397336959839,
+ "step": 1710
+ },
+ {
+ "epoch": 23.767888307155324,
+ "grad_norm": 0.05179062858223915,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 4.027141571044922,
+ "step": 1711
+ },
+ {
+ "epoch": 23.781849912739965,
+ "grad_norm": 0.049667056649923325,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 4.024991512298584,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79581151832461,
+ "grad_norm": 0.05204658955335617,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 4.002647876739502,
+ "step": 1713
+ },
+ {
+ "epoch": 23.80977312390925,
+ "grad_norm": 0.05069981515407562,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 4.035422325134277,
+ "step": 1714
+ },
+ {
+ "epoch": 23.823734729493893,
+ "grad_norm": 0.05044785887002945,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 4.042465686798096,
+ "step": 1715
+ },
+ {
+ "epoch": 23.837696335078533,
+ "grad_norm": 0.053949885070323944,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 4.002717971801758,
+ "step": 1716
+ },
+ {
+ "epoch": 23.851657940663177,
+ "grad_norm": 0.05139258876442909,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 4.026609897613525,
+ "step": 1717
+ },
+ {
+ "epoch": 23.865619546247817,
+ "grad_norm": 0.05316372960805893,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 4.019758224487305,
+ "step": 1718
+ },
+ {
+ "epoch": 23.87958115183246,
+ "grad_norm": 0.050791382789611816,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 3.9859633445739746,
+ "step": 1719
+ },
+ {
+ "epoch": 23.8935427574171,
+ "grad_norm": 0.05124847590923309,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 4.038055419921875,
+ "step": 1720
+ },
+ {
+ "epoch": 23.907504363001745,
+ "grad_norm": 0.0495096817612648,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 4.02764368057251,
+ "step": 1721
+ },
+ {
+ "epoch": 23.921465968586386,
+ "grad_norm": 0.05205567181110382,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 4.003867149353027,
+ "step": 1722
+ },
+ {
+ "epoch": 23.93542757417103,
+ "grad_norm": 0.05233347788453102,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 4.024157524108887,
+ "step": 1723
+ },
+ {
+ "epoch": 23.949389179755673,
+ "grad_norm": 0.05126349255442619,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 4.05349063873291,
+ "step": 1724
+ },
+ {
+ "epoch": 23.963350785340314,
+ "grad_norm": 0.050006967037916183,
+ "learning_rate": 0.000336518346307424,
+ "loss": 4.024704933166504,
+ "step": 1725
+ },
+ {
+ "epoch": 23.977312390924958,
+ "grad_norm": 0.052343279123306274,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 4.008549213409424,
+ "step": 1726
+ },
+ {
+ "epoch": 23.991273996509598,
+ "grad_norm": 0.05126652494072914,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 4.002162933349609,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.03560106083750725,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 2.511049270629883,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.5985628962516785,
+ "eval_runtime": 59.0712,
+ "eval_samples_per_second": 41.34,
+ "eval_steps_per_second": 0.66,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013961605584644,
+ "grad_norm": 0.05205906555056572,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.936516284942627,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027923211169284,
+ "grad_norm": 0.056368257850408554,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.95200777053833,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041884816753928,
+ "grad_norm": 0.05075085908174515,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.940810441970825,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05584642233857,
+ "grad_norm": 0.05119713768362999,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.90130615234375,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069808027923212,
+ "grad_norm": 0.05254847928881645,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.9491095542907715,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083769633507853,
+ "grad_norm": 0.053025662899017334,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.9371917247772217,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097731239092496,
+ "grad_norm": 0.05402665585279465,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 3.9651317596435547,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111692844677137,
+ "grad_norm": 0.0519571453332901,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.9189062118530273,
+ "step": 1736
+ },
+ {
+ "epoch": 24.12565445026178,
+ "grad_norm": 0.05436019226908684,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.948184013366699,
+ "step": 1737
+ },
+ {
+ "epoch": 24.13961605584642,
+ "grad_norm": 0.056572459638118744,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.9076738357543945,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153577661431065,
+ "grad_norm": 0.05553183704614639,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.933659076690674,
+ "step": 1739
+ },
+ {
+ "epoch": 24.167539267015705,
+ "grad_norm": 0.05554374307394028,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.941751480102539,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18150087260035,
+ "grad_norm": 0.05451614409685135,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.928593635559082,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195462478184993,
+ "grad_norm": 0.055458780378103256,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 3.961820602416992,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209424083769633,
+ "grad_norm": 0.054980456829071045,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 3.9736528396606445,
+ "step": 1743
+ },
+ {
+ "epoch": 24.223385689354277,
+ "grad_norm": 0.05538604408502579,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.935093641281128,
+ "step": 1744
+ },
+ {
+ "epoch": 24.237347294938917,
+ "grad_norm": 0.05580678582191467,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.9240689277648926,
+ "step": 1745
+ },
+ {
+ "epoch": 24.25130890052356,
+ "grad_norm": 0.052294690161943436,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.937718152999878,
+ "step": 1746
+ },
+ {
+ "epoch": 24.2652705061082,
+ "grad_norm": 0.052018795162439346,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.942049503326416,
+ "step": 1747
+ },
+ {
+ "epoch": 24.279232111692846,
+ "grad_norm": 0.0550091527402401,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.9215564727783203,
+ "step": 1748
+ },
+ {
+ "epoch": 24.293193717277486,
+ "grad_norm": 0.05447716638445854,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.9154953956604004,
+ "step": 1749
+ },
+ {
+ "epoch": 24.30715532286213,
+ "grad_norm": 0.05300029739737511,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 3.9642622470855713,
+ "step": 1750
+ },
+ {
+ "epoch": 24.32111692844677,
+ "grad_norm": 0.05312013998627663,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.938897132873535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335078534031414,
+ "grad_norm": 0.053737837821245193,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.94972562789917,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349040139616054,
+ "grad_norm": 0.05087469890713692,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.9301164150238037,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363001745200698,
+ "grad_norm": 0.05179893225431442,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.9393205642700195,
+ "step": 1754
+ },
+ {
+ "epoch": 24.376963350785342,
+ "grad_norm": 0.05277438461780548,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 3.960968494415283,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390924956369982,
+ "grad_norm": 0.05090474337339401,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 3.9516637325286865,
+ "step": 1756
+ },
+ {
+ "epoch": 24.404886561954626,
+ "grad_norm": 0.052786506712436676,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.9332518577575684,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418848167539267,
+ "grad_norm": 0.052095942199230194,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.9108662605285645,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43280977312391,
+ "grad_norm": 0.05310450494289398,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 3.9746041297912598,
+ "step": 1759
+ },
+ {
+ "epoch": 24.44677137870855,
+ "grad_norm": 0.05078566074371338,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.93316650390625,
+ "step": 1760
+ },
+ {
+ "epoch": 24.460732984293195,
+ "grad_norm": 0.0519949309527874,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.9489870071411133,
+ "step": 1761
+ },
+ {
+ "epoch": 24.474694589877835,
+ "grad_norm": 0.05239349976181984,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 3.955118417739868,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48865619546248,
+ "grad_norm": 0.05193360522389412,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 3.9707555770874023,
+ "step": 1763
+ },
+ {
+ "epoch": 24.50261780104712,
+ "grad_norm": 0.05397362262010574,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 3.973053455352783,
+ "step": 1764
+ },
+ {
+ "epoch": 24.516579406631763,
+ "grad_norm": 0.05379117652773857,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 3.957289695739746,
+ "step": 1765
+ },
+ {
+ "epoch": 24.530541012216403,
+ "grad_norm": 0.0526256300508976,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 3.9595577716827393,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544502617801047,
+ "grad_norm": 0.05377693101763725,
+ "learning_rate": 0.000325372061241796,
+ "loss": 3.97849178314209,
+ "step": 1767
+ },
+ {
+ "epoch": 24.55846422338569,
+ "grad_norm": 0.05182809755206108,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 3.9714043140411377,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57242582897033,
+ "grad_norm": 0.05364793911576271,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 3.9582200050354004,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586387434554975,
+ "grad_norm": 0.051558300852775574,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 3.9747936725616455,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600349040139616,
+ "grad_norm": 0.053487639874219894,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 3.9556500911712646,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61431064572426,
+ "grad_norm": 0.05146970972418785,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.94858455657959,
+ "step": 1772
+ },
+ {
+ "epoch": 24.6282722513089,
+ "grad_norm": 0.05099836364388466,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 3.9716954231262207,
+ "step": 1773
+ },
+ {
+ "epoch": 24.642233856893544,
+ "grad_norm": 0.054634734988212585,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 3.974881887435913,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656195462478184,
+ "grad_norm": 0.054601747542619705,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 3.9746241569519043,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670157068062828,
+ "grad_norm": 0.052188221365213394,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 3.957106113433838,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68411867364747,
+ "grad_norm": 0.055673446506261826,
+ "learning_rate": 0.000322712903929477,
+ "loss": 3.9803967475891113,
+ "step": 1777
+ },
+ {
+ "epoch": 24.698080279232112,
+ "grad_norm": 0.05397328734397888,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 3.967238426208496,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712041884816752,
+ "grad_norm": 0.05384642630815506,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 3.949392795562744,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726003490401396,
+ "grad_norm": 0.05452995374798775,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.9432482719421387,
+ "step": 1780
+ },
+ {
+ "epoch": 24.739965095986037,
+ "grad_norm": 0.052430473268032074,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 3.9944653511047363,
+ "step": 1781
+ },
+ {
+ "epoch": 24.75392670157068,
+ "grad_norm": 0.05514240264892578,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 3.97713565826416,
+ "step": 1782
+ },
+ {
+ "epoch": 24.767888307155324,
+ "grad_norm": 0.05189919099211693,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 3.983150005340576,
+ "step": 1783
+ },
+ {
+ "epoch": 24.781849912739965,
+ "grad_norm": 0.05590776726603508,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 3.9548301696777344,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79581151832461,
+ "grad_norm": 0.05564190074801445,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 3.9725170135498047,
+ "step": 1785
+ },
+ {
+ "epoch": 24.80977312390925,
+ "grad_norm": 0.05121326446533203,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 3.9722347259521484,
+ "step": 1786
+ },
+ {
+ "epoch": 24.823734729493893,
+ "grad_norm": 0.05444202199578285,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 3.9634029865264893,
+ "step": 1787
+ },
+ {
+ "epoch": 24.837696335078533,
+ "grad_norm": 0.05242442339658737,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 3.973270893096924,
+ "step": 1788
+ },
+ {
+ "epoch": 24.851657940663177,
+ "grad_norm": 0.05497463420033455,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 3.9851887226104736,
+ "step": 1789
+ },
+ {
+ "epoch": 24.865619546247817,
+ "grad_norm": 0.05482025071978569,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 4.001110076904297,
+ "step": 1790
+ },
+ {
+ "epoch": 24.87958115183246,
+ "grad_norm": 0.054101862013339996,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 3.9782891273498535,
+ "step": 1791
+ },
+ {
+ "epoch": 24.8935427574171,
+ "grad_norm": 0.05538944527506828,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 3.973541498184204,
+ "step": 1792
+ },
+ {
+ "epoch": 24.907504363001745,
+ "grad_norm": 0.0502893440425396,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 3.9875426292419434,
+ "step": 1793
+ },
+ {
+ "epoch": 24.921465968586386,
+ "grad_norm": 0.05576152354478836,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.947035551071167,
+ "step": 1794
+ },
+ {
+ "epoch": 24.93542757417103,
+ "grad_norm": 0.05356278270483017,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 3.9683666229248047,
+ "step": 1795
+ },
+ {
+ "epoch": 24.949389179755673,
+ "grad_norm": 0.05377693101763725,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 3.9973440170288086,
+ "step": 1796
+ },
+ {
+ "epoch": 24.963350785340314,
+ "grad_norm": 0.05334382876753807,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 3.9843404293060303,
+ "step": 1797
+ },
+ {
+ "epoch": 24.977312390924958,
+ "grad_norm": 0.05514024198055267,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 3.9920568466186523,
+ "step": 1798
+ },
+ {
+ "epoch": 24.991273996509598,
+ "grad_norm": 0.05631054937839508,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 4.005859375,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.03653077408671379,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 2.503864049911499,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 0.5995591878890991,
+ "eval_runtime": 59.5213,
+ "eval_samples_per_second": 41.027,
+ "eval_steps_per_second": 0.655,
+ "step": 1800
+ },
+ {
+ "epoch": 25.013961605584644,
+ "grad_norm": 0.0576021745800972,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 3.894834518432617,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027923211169284,
+ "grad_norm": 0.05842140316963196,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 3.8879270553588867,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041884816753928,
+ "grad_norm": 0.051821205765008926,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 3.9011340141296387,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05584642233857,
+ "grad_norm": 0.053872160613536835,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 3.877183437347412,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069808027923212,
+ "grad_norm": 0.05487574636936188,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 3.9126815795898438,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083769633507853,
+ "grad_norm": 0.054469428956508636,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 3.894768238067627,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097731239092496,
+ "grad_norm": 0.05124446377158165,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 3.916809320449829,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111692844677137,
+ "grad_norm": 0.05193263292312622,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 3.917140483856201,
+ "step": 1808
+ },
+ {
+ "epoch": 25.12565445026178,
+ "grad_norm": 0.0519125871360302,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 3.903704881668091,
+ "step": 1809
+ },
+ {
+ "epoch": 25.13961605584642,
+ "grad_norm": 0.052205223590135574,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 3.892639636993408,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153577661431065,
+ "grad_norm": 0.052036914974451065,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 3.895559787750244,
+ "step": 1811
+ },
+ {
+ "epoch": 25.167539267015705,
+ "grad_norm": 0.054616156965494156,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 3.8976142406463623,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18150087260035,
+ "grad_norm": 0.05603514611721039,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 3.914311170578003,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195462478184993,
+ "grad_norm": 0.05170242115855217,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 3.9141643047332764,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209424083769633,
+ "grad_norm": 0.05476640164852142,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 3.8952436447143555,
+ "step": 1815
+ },
+ {
+ "epoch": 25.223385689354277,
+ "grad_norm": 0.055114105343818665,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 3.872978687286377,
+ "step": 1816
+ },
+ {
+ "epoch": 25.237347294938917,
+ "grad_norm": 0.051593828946352005,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 3.893911600112915,
+ "step": 1817
+ },
+ {
+ "epoch": 25.25130890052356,
+ "grad_norm": 0.05363611876964569,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 3.9062986373901367,
+ "step": 1818
+ },
+ {
+ "epoch": 25.2652705061082,
+ "grad_norm": 0.054656051099300385,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 3.9145383834838867,
+ "step": 1819
+ },
+ {
+ "epoch": 25.279232111692846,
+ "grad_norm": 0.05251947045326233,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 3.912965774536133,
+ "step": 1820
+ },
+ {
+ "epoch": 25.293193717277486,
+ "grad_norm": 0.05575991049408913,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 3.901510238647461,
+ "step": 1821
+ },
+ {
+ "epoch": 25.30715532286213,
+ "grad_norm": 0.05540676787495613,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 3.8659887313842773,
+ "step": 1822
+ },
+ {
+ "epoch": 25.32111692844677,
+ "grad_norm": 0.0577131025493145,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 3.884077548980713,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335078534031414,
+ "grad_norm": 0.055947501212358475,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 3.9274377822875977,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349040139616054,
+ "grad_norm": 0.05328478664159775,
+ "learning_rate": 0.000309930621834911,
+ "loss": 3.8783111572265625,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363001745200698,
+ "grad_norm": 0.057696059346199036,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 3.903721570968628,
+ "step": 1826
+ },
+ {
+ "epoch": 25.376963350785342,
+ "grad_norm": 0.05717543140053749,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 3.9259400367736816,
+ "step": 1827
+ },
+ {
+ "epoch": 25.390924956369982,
+ "grad_norm": 0.05328287556767464,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 3.9450511932373047,
+ "step": 1828
+ },
+ {
+ "epoch": 25.404886561954626,
+ "grad_norm": 0.05414140224456787,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 3.9231765270233154,
+ "step": 1829
+ },
+ {
+ "epoch": 25.418848167539267,
+ "grad_norm": 0.05535200983285904,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 3.8928630352020264,
+ "step": 1830
+ },
+ {
+ "epoch": 25.43280977312391,
+ "grad_norm": 0.05468215420842171,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 3.933401584625244,
+ "step": 1831
+ },
+ {
+ "epoch": 25.44677137870855,
+ "grad_norm": 0.056624677032232285,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 3.922624111175537,
+ "step": 1832
+ },
+ {
+ "epoch": 25.460732984293195,
+ "grad_norm": 0.0545354001224041,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 3.916515827178955,
+ "step": 1833
+ },
+ {
+ "epoch": 25.474694589877835,
+ "grad_norm": 0.053979165852069855,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 3.891812324523926,
+ "step": 1834
+ },
+ {
+ "epoch": 25.48865619546248,
+ "grad_norm": 0.05619875714182854,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 3.959524631500244,
+ "step": 1835
+ },
+ {
+ "epoch": 25.50261780104712,
+ "grad_norm": 0.053561773151159286,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 3.9109320640563965,
+ "step": 1836
+ },
+ {
+ "epoch": 25.516579406631763,
+ "grad_norm": 0.05566735938191414,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 3.910492420196533,
+ "step": 1837
+ },
+ {
+ "epoch": 25.530541012216403,
+ "grad_norm": 0.05757030472159386,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 3.9716763496398926,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544502617801047,
+ "grad_norm": 0.05121897906064987,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 3.9114646911621094,
+ "step": 1839
+ },
+ {
+ "epoch": 25.55846422338569,
+ "grad_norm": 0.05664774775505066,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 3.9465560913085938,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57242582897033,
+ "grad_norm": 0.0569215826690197,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 3.908374309539795,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586387434554975,
+ "grad_norm": 0.05345458909869194,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 3.9336414337158203,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600349040139616,
+ "grad_norm": 0.0565933920443058,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 3.9478759765625,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61431064572426,
+ "grad_norm": 0.05887870490550995,
+ "learning_rate": 0.000304866093757771,
+ "loss": 3.9018349647521973,
+ "step": 1844
+ },
+ {
+ "epoch": 25.6282722513089,
+ "grad_norm": 0.05515308678150177,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 3.9446825981140137,
+ "step": 1845
+ },
+ {
+ "epoch": 25.642233856893544,
+ "grad_norm": 0.058602213859558105,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 3.9497885704040527,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656195462478184,
+ "grad_norm": 0.053806282579898834,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 3.9247193336486816,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670157068062828,
+ "grad_norm": 0.05495952442288399,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 3.912296772003174,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68411867364747,
+ "grad_norm": 0.05479473993182182,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 3.9471306800842285,
+ "step": 1849
+ },
+ {
+ "epoch": 25.698080279232112,
+ "grad_norm": 0.053319111466407776,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 3.955942153930664,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712041884816752,
+ "grad_norm": 0.05832696333527565,
+ "learning_rate": 0.000303,
+ "loss": 3.959169626235962,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726003490401396,
+ "grad_norm": 0.057019371539354324,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 3.9682188034057617,
+ "step": 1852
+ },
+ {
+ "epoch": 25.739965095986037,
+ "grad_norm": 0.05099530518054962,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 3.921377658843994,
+ "step": 1853
+ },
+ {
+ "epoch": 25.75392670157068,
+ "grad_norm": 0.05466439574956894,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 3.946341037750244,
+ "step": 1854
+ },
+ {
+ "epoch": 25.767888307155324,
+ "grad_norm": 0.05742308124899864,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 3.948031425476074,
+ "step": 1855
+ },
+ {
+ "epoch": 25.781849912739965,
+ "grad_norm": 0.057581741362810135,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 3.91162109375,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79581151832461,
+ "grad_norm": 0.05520833283662796,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 3.937495231628418,
+ "step": 1857
+ },
+ {
+ "epoch": 25.80977312390925,
+ "grad_norm": 0.05763327330350876,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 3.9565534591674805,
+ "step": 1858
+ },
+ {
+ "epoch": 25.823734729493893,
+ "grad_norm": 0.0544707253575325,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 3.965190887451172,
+ "step": 1859
+ },
+ {
+ "epoch": 25.837696335078533,
+ "grad_norm": 0.05445104092359543,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 3.938063859939575,
+ "step": 1860
+ },
+ {
+ "epoch": 25.851657940663177,
+ "grad_norm": 0.05505113676190376,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 3.928392171859741,
+ "step": 1861
+ },
+ {
+ "epoch": 25.865619546247817,
+ "grad_norm": 0.05574212223291397,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 3.959397077560425,
+ "step": 1862
+ },
+ {
+ "epoch": 25.87958115183246,
+ "grad_norm": 0.053950466215610504,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 3.9532437324523926,
+ "step": 1863
+ },
+ {
+ "epoch": 25.8935427574171,
+ "grad_norm": 0.05495546758174896,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 3.946746826171875,
+ "step": 1864
+ },
+ {
+ "epoch": 25.907504363001745,
+ "grad_norm": 0.05328625068068504,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 3.9494776725769043,
+ "step": 1865
+ },
+ {
+ "epoch": 25.921465968586386,
+ "grad_norm": 0.05593104660511017,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 3.959989547729492,
+ "step": 1866
+ },
+ {
+ "epoch": 25.93542757417103,
+ "grad_norm": 0.05442969873547554,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 3.952371597290039,
+ "step": 1867
+ },
+ {
+ "epoch": 25.949389179755673,
+ "grad_norm": 0.05446415767073631,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 3.9538888931274414,
+ "step": 1868
+ },
+ {
+ "epoch": 25.963350785340314,
+ "grad_norm": 0.054761797189712524,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 3.9386215209960938,
+ "step": 1869
+ },
+ {
+ "epoch": 25.977312390924958,
+ "grad_norm": 0.054201845079660416,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 3.9252963066101074,
+ "step": 1870
+ },
+ {
+ "epoch": 25.991273996509598,
+ "grad_norm": 0.05364517495036125,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 3.9733197689056396,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.038914378732442856,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 2.4261250495910645,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 0.6009622812271118,
+ "eval_runtime": 60.6864,
+ "eval_samples_per_second": 40.24,
+ "eval_steps_per_second": 0.643,
+ "step": 1872
+ },
+ {
+ "epoch": 26.013961605584644,
+ "grad_norm": 0.05345558747649193,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 3.8902673721313477,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027923211169284,
+ "grad_norm": 0.05672457069158554,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 3.8676247596740723,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041884816753928,
+ "grad_norm": 0.051981374621391296,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 3.872048854827881,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05584642233857,
+ "grad_norm": 0.05262405797839165,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 3.8582329750061035,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069808027923212,
+ "grad_norm": 0.054473262280225754,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 3.8352112770080566,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083769633507853,
+ "grad_norm": 0.053167302161455154,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 3.881760835647583,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097731239092496,
+ "grad_norm": 0.053720805794000626,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 3.891775608062744,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111692844677137,
+ "grad_norm": 0.055102936923503876,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 3.8872156143188477,
+ "step": 1880
+ },
+ {
+ "epoch": 26.12565445026178,
+ "grad_norm": 0.05108853057026863,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 3.8625407218933105,
+ "step": 1881
+ },
+ {
+ "epoch": 26.13961605584642,
+ "grad_norm": 0.05526706948876381,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 3.886127233505249,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153577661431065,
+ "grad_norm": 0.05458226427435875,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 3.8722105026245117,
+ "step": 1883
+ },
+ {
+ "epoch": 26.167539267015705,
+ "grad_norm": 0.05342113599181175,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 3.853677749633789,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18150087260035,
+ "grad_norm": 0.053665641695261,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 3.849315881729126,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195462478184993,
+ "grad_norm": 0.054922688752412796,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 3.8599777221679688,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209424083769633,
+ "grad_norm": 0.05412537604570389,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 3.8767571449279785,
+ "step": 1887
+ },
+ {
+ "epoch": 26.223385689354277,
+ "grad_norm": 0.05467838793992996,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 3.8436965942382812,
+ "step": 1888
+ },
+ {
+ "epoch": 26.237347294938917,
+ "grad_norm": 0.054886411875486374,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 3.8796300888061523,
+ "step": 1889
+ },
+ {
+ "epoch": 26.25130890052356,
+ "grad_norm": 0.05437029153108597,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 3.858133316040039,
+ "step": 1890
+ },
+ {
+ "epoch": 26.2652705061082,
+ "grad_norm": 0.05273836478590965,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 3.839015483856201,
+ "step": 1891
+ },
+ {
+ "epoch": 26.279232111692846,
+ "grad_norm": 0.0546451173722744,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 3.8723373413085938,
+ "step": 1892
+ },
+ {
+ "epoch": 26.293193717277486,
+ "grad_norm": 0.05240490287542343,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 3.877549886703491,
+ "step": 1893
+ },
+ {
+ "epoch": 26.30715532286213,
+ "grad_norm": 0.05420667678117752,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 3.8643171787261963,
+ "step": 1894
+ },
+ {
+ "epoch": 26.32111692844677,
+ "grad_norm": 0.053440142422914505,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 3.8681447505950928,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335078534031414,
+ "grad_norm": 0.052396632730960846,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 3.8691887855529785,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349040139616054,
+ "grad_norm": 0.05318925902247429,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 3.897003650665283,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363001745200698,
+ "grad_norm": 0.05328533053398132,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 3.9079437255859375,
+ "step": 1898
+ },
+ {
+ "epoch": 26.376963350785342,
+ "grad_norm": 0.05196928232908249,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 3.8392016887664795,
+ "step": 1899
+ },
+ {
+ "epoch": 26.390924956369982,
+ "grad_norm": 0.053326766937971115,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 3.8930301666259766,
+ "step": 1900
+ },
+ {
+ "epoch": 26.404886561954626,
+ "grad_norm": 0.052926234900951385,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 3.866997718811035,
+ "step": 1901
+ },
+ {
+ "epoch": 26.418848167539267,
+ "grad_norm": 0.05289287120103836,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 3.896690607070923,
+ "step": 1902
+ },
+ {
+ "epoch": 26.43280977312391,
+ "grad_norm": 0.052162300795316696,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 3.8642401695251465,
+ "step": 1903
+ },
+ {
+ "epoch": 26.44677137870855,
+ "grad_norm": 0.05495629832148552,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 3.8982696533203125,
+ "step": 1904
+ },
+ {
+ "epoch": 26.460732984293195,
+ "grad_norm": 0.055102866142988205,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 3.8725972175598145,
+ "step": 1905
+ },
+ {
+ "epoch": 26.474694589877835,
+ "grad_norm": 0.05537770688533783,
+ "learning_rate": 0.000288343693342466,
+ "loss": 3.9050111770629883,
+ "step": 1906
+ },
+ {
+ "epoch": 26.48865619546248,
+ "grad_norm": 0.05490661785006523,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 3.90586256980896,
+ "step": 1907
+ },
+ {
+ "epoch": 26.50261780104712,
+ "grad_norm": 0.054393451660871506,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 3.8841705322265625,
+ "step": 1908
+ },
+ {
+ "epoch": 26.516579406631763,
+ "grad_norm": 0.0553058385848999,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 3.883500576019287,
+ "step": 1909
+ },
+ {
+ "epoch": 26.530541012216403,
+ "grad_norm": 0.053976885974407196,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 3.894948959350586,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544502617801047,
+ "grad_norm": 0.05290692299604416,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 3.8880579471588135,
+ "step": 1911
+ },
+ {
+ "epoch": 26.55846422338569,
+ "grad_norm": 0.05570613220334053,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 3.9035487174987793,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57242582897033,
+ "grad_norm": 0.054136309772729874,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 3.9137563705444336,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586387434554975,
+ "grad_norm": 0.05328850820660591,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 3.87838077545166,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600349040139616,
+ "grad_norm": 0.05169905349612236,
+ "learning_rate": 0.000285947841605349,
+ "loss": 3.882963180541992,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61431064572426,
+ "grad_norm": 0.05304642394185066,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 3.9250073432922363,
+ "step": 1916
+ },
+ {
+ "epoch": 26.6282722513089,
+ "grad_norm": 0.051766350865364075,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 3.8915140628814697,
+ "step": 1917
+ },
+ {
+ "epoch": 26.642233856893544,
+ "grad_norm": 0.054485417902469635,
+ "learning_rate": 0.000285149463934261,
+ "loss": 3.8973851203918457,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656195462478184,
+ "grad_norm": 0.05437815189361572,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 3.9054274559020996,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670157068062828,
+ "grad_norm": 0.05402572825551033,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 3.9060683250427246,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68411867364747,
+ "grad_norm": 0.05332810431718826,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 3.9053144454956055,
+ "step": 1921
+ },
+ {
+ "epoch": 26.698080279232112,
+ "grad_norm": 0.0557575598359108,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 3.9026694297790527,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712041884816752,
+ "grad_norm": 0.05650033429265022,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 3.8795413970947266,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726003490401396,
+ "grad_norm": 0.054456304758787155,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 3.932950496673584,
+ "step": 1924
+ },
+ {
+ "epoch": 26.739965095986037,
+ "grad_norm": 0.05676668882369995,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 3.878704071044922,
+ "step": 1925
+ },
+ {
+ "epoch": 26.75392670157068,
+ "grad_norm": 0.05875959247350693,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 3.920717239379883,
+ "step": 1926
+ },
+ {
+ "epoch": 26.767888307155324,
+ "grad_norm": 0.054490379989147186,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 3.912592887878418,
+ "step": 1927
+ },
+ {
+ "epoch": 26.781849912739965,
+ "grad_norm": 0.055268481373786926,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 3.900418758392334,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79581151832461,
+ "grad_norm": 0.05617086961865425,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 3.899095296859741,
+ "step": 1929
+ },
+ {
+ "epoch": 26.80977312390925,
+ "grad_norm": 0.05310523882508278,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 3.917501449584961,
+ "step": 1930
+ },
+ {
+ "epoch": 26.823734729493893,
+ "grad_norm": 0.05476578325033188,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 3.867847442626953,
+ "step": 1931
+ },
+ {
+ "epoch": 26.837696335078533,
+ "grad_norm": 0.05456431582570076,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 3.8760366439819336,
+ "step": 1932
+ },
+ {
+ "epoch": 26.851657940663177,
+ "grad_norm": 0.055269334465265274,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 3.910797119140625,
+ "step": 1933
+ },
+ {
+ "epoch": 26.865619546247817,
+ "grad_norm": 0.054999858140945435,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 3.946951389312744,
+ "step": 1934
+ },
+ {
+ "epoch": 26.87958115183246,
+ "grad_norm": 0.05520344525575638,
+ "learning_rate": 0.000280627938758204,
+ "loss": 3.9161229133605957,
+ "step": 1935
+ },
+ {
+ "epoch": 26.8935427574171,
+ "grad_norm": 0.05451727285981178,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 3.908015727996826,
+ "step": 1936
+ },
+ {
+ "epoch": 26.907504363001745,
+ "grad_norm": 0.05561735853552818,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 3.944608449935913,
+ "step": 1937
+ },
+ {
+ "epoch": 26.921465968586386,
+ "grad_norm": 0.054743457585573196,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 3.906559467315674,
+ "step": 1938
+ },
+ {
+ "epoch": 26.93542757417103,
+ "grad_norm": 0.05497661978006363,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 3.900981903076172,
+ "step": 1939
+ },
+ {
+ "epoch": 26.949389179755673,
+ "grad_norm": 0.05452290549874306,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 3.910825252532959,
+ "step": 1940
+ },
+ {
+ "epoch": 26.963350785340314,
+ "grad_norm": 0.05409001186490059,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 3.921509265899658,
+ "step": 1941
+ },
+ {
+ "epoch": 26.977312390924958,
+ "grad_norm": 0.05467002093791962,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 3.9039864540100098,
+ "step": 1942
+ },
+ {
+ "epoch": 26.991273996509598,
+ "grad_norm": 0.05442725867033005,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 3.9153316020965576,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.039271868765354156,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 2.451029062271118,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 0.6021912097930908,
+ "eval_runtime": 61.3232,
+ "eval_samples_per_second": 39.822,
+ "eval_steps_per_second": 0.636,
+ "step": 1944
+ },
+ {
+ "epoch": 27.013961605584644,
+ "grad_norm": 0.05251244455575943,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 3.8143270015716553,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027923211169284,
+ "grad_norm": 0.056699808686971664,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 3.8331453800201416,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041884816753928,
+ "grad_norm": 0.055446673184633255,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 3.833841323852539,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05584642233857,
+ "grad_norm": 0.053186994045972824,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 3.82643985748291,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069808027923212,
+ "grad_norm": 0.05483868718147278,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 3.8209316730499268,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083769633507853,
+ "grad_norm": 0.05387664586305618,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 3.8402364253997803,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097731239092496,
+ "grad_norm": 0.05422234162688255,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 3.8125908374786377,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111692844677137,
+ "grad_norm": 0.05522442236542702,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 3.830599784851074,
+ "step": 1952
+ },
+ {
+ "epoch": 27.12565445026178,
+ "grad_norm": 0.05163128301501274,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 3.833247184753418,
+ "step": 1953
+ },
+ {
+ "epoch": 27.13961605584642,
+ "grad_norm": 0.053523529320955276,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 3.812887191772461,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153577661431065,
+ "grad_norm": 0.05362863466143608,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 3.8325610160827637,
+ "step": 1955
+ },
+ {
+ "epoch": 27.167539267015705,
+ "grad_norm": 0.054255858063697815,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 3.807152271270752,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18150087260035,
+ "grad_norm": 0.05259740352630615,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 3.843714475631714,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195462478184993,
+ "grad_norm": 0.05457160621881485,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 3.847249984741211,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209424083769633,
+ "grad_norm": 0.05517619848251343,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 3.864114284515381,
+ "step": 1959
+ },
+ {
+ "epoch": 27.223385689354277,
+ "grad_norm": 0.052946750074625015,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 3.8219656944274902,
+ "step": 1960
+ },
+ {
+ "epoch": 27.237347294938917,
+ "grad_norm": 0.053545452654361725,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 3.8502726554870605,
+ "step": 1961
+ },
+ {
+ "epoch": 27.25130890052356,
+ "grad_norm": 0.054657287895679474,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 3.837273120880127,
+ "step": 1962
+ },
+ {
+ "epoch": 27.2652705061082,
+ "grad_norm": 0.05385086312890053,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 3.837573528289795,
+ "step": 1963
+ },
+ {
+ "epoch": 27.279232111692846,
+ "grad_norm": 0.05512048304080963,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 3.842432737350464,
+ "step": 1964
+ },
+ {
+ "epoch": 27.293193717277486,
+ "grad_norm": 0.05286296457052231,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 3.845123767852783,
+ "step": 1965
+ },
+ {
+ "epoch": 27.30715532286213,
+ "grad_norm": 0.05396866053342819,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 3.836874485015869,
+ "step": 1966
+ },
+ {
+ "epoch": 27.32111692844677,
+ "grad_norm": 0.053341615945100784,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 3.8860960006713867,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335078534031414,
+ "grad_norm": 0.0531761609017849,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 3.828686237335205,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349040139616054,
+ "grad_norm": 0.054875023663043976,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 3.83782696723938,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363001745200698,
+ "grad_norm": 0.057710494846105576,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 3.8841819763183594,
+ "step": 1970
+ },
+ {
+ "epoch": 27.376963350785342,
+ "grad_norm": 0.05443769320845604,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 3.856564521789551,
+ "step": 1971
+ },
+ {
+ "epoch": 27.390924956369982,
+ "grad_norm": 0.05571551248431206,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 3.847541570663452,
+ "step": 1972
+ },
+ {
+ "epoch": 27.404886561954626,
+ "grad_norm": 0.05770037695765495,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 3.843390941619873,
+ "step": 1973
+ },
+ {
+ "epoch": 27.418848167539267,
+ "grad_norm": 0.05414504557847977,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 3.8786306381225586,
+ "step": 1974
+ },
+ {
+ "epoch": 27.43280977312391,
+ "grad_norm": 0.058058444410562515,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 3.8680999279022217,
+ "step": 1975
+ },
+ {
+ "epoch": 27.44677137870855,
+ "grad_norm": 0.056312717497348785,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 3.858879566192627,
+ "step": 1976
+ },
+ {
+ "epoch": 27.460732984293195,
+ "grad_norm": 0.055812690407037735,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 3.870741844177246,
+ "step": 1977
+ },
+ {
+ "epoch": 27.474694589877835,
+ "grad_norm": 0.057005349546670914,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 3.8850440979003906,
+ "step": 1978
+ },
+ {
+ "epoch": 27.48865619546248,
+ "grad_norm": 0.055214181542396545,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 3.8485918045043945,
+ "step": 1979
+ },
+ {
+ "epoch": 27.50261780104712,
+ "grad_norm": 0.054300058633089066,
+ "learning_rate": 0.00026868712586269,
+ "loss": 3.8759729862213135,
+ "step": 1980
+ },
+ {
+ "epoch": 27.516579406631763,
+ "grad_norm": 0.05370737239718437,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 3.8498623371124268,
+ "step": 1981
+ },
+ {
+ "epoch": 27.530541012216403,
+ "grad_norm": 0.05481775104999542,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 3.8408432006835938,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544502617801047,
+ "grad_norm": 0.05360998213291168,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 3.8800129890441895,
+ "step": 1983
+ },
+ {
+ "epoch": 27.55846422338569,
+ "grad_norm": 0.05510537698864937,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 3.8329358100891113,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57242582897033,
+ "grad_norm": 0.057950302958488464,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 3.870562791824341,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586387434554975,
+ "grad_norm": 0.0546487458050251,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 3.85231351852417,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600349040139616,
+ "grad_norm": 0.056359753012657166,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 3.8782615661621094,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61431064572426,
+ "grad_norm": 0.05409953370690346,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 3.8209736347198486,
+ "step": 1988
+ },
+ {
+ "epoch": 27.6282722513089,
+ "grad_norm": 0.05549091845750809,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 3.9041504859924316,
+ "step": 1989
+ },
+ {
+ "epoch": 27.642233856893544,
+ "grad_norm": 0.05437365546822548,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 3.8530654907226562,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656195462478184,
+ "grad_norm": 0.055456943809986115,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 3.862886428833008,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670157068062828,
+ "grad_norm": 0.05614335089921951,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 3.867595672607422,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68411867364747,
+ "grad_norm": 0.05374857038259506,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 3.8731391429901123,
+ "step": 1993
+ },
+ {
+ "epoch": 27.698080279232112,
+ "grad_norm": 0.05544714257121086,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 3.837608575820923,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712041884816752,
+ "grad_norm": 0.05495234578847885,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 3.8587915897369385,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726003490401396,
+ "grad_norm": 0.05354582890868187,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 3.8618052005767822,
+ "step": 1996
+ },
+ {
+ "epoch": 27.739965095986037,
+ "grad_norm": 0.05556262657046318,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 3.8630597591400146,
+ "step": 1997
+ },
+ {
+ "epoch": 27.75392670157068,
+ "grad_norm": 0.05576452240347862,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 3.8570914268493652,
+ "step": 1998
+ },
+ {
+ "epoch": 27.767888307155324,
+ "grad_norm": 0.053394027054309845,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 3.868783950805664,
+ "step": 1999
+ },
+ {
+ "epoch": 27.781849912739965,
+ "grad_norm": 0.05506644397974014,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 3.871555805206299,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79581151832461,
+ "grad_norm": 0.054367583245038986,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 3.8487513065338135,
+ "step": 2001
+ },
+ {
+ "epoch": 27.80977312390925,
+ "grad_norm": 0.05482787266373634,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 3.88199520111084,
+ "step": 2002
+ },
+ {
+ "epoch": 27.823734729493893,
+ "grad_norm": 0.05416601523756981,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 3.8876938819885254,
+ "step": 2003
+ },
+ {
+ "epoch": 27.837696335078533,
+ "grad_norm": 0.05612047761678696,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 3.8772940635681152,
+ "step": 2004
+ },
+ {
+ "epoch": 27.851657940663177,
+ "grad_norm": 0.05662519112229347,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 3.8429012298583984,
+ "step": 2005
+ },
+ {
+ "epoch": 27.865619546247817,
+ "grad_norm": 0.055557794868946075,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 3.8712010383605957,
+ "step": 2006
+ },
+ {
+ "epoch": 27.87958115183246,
+ "grad_norm": 0.054365914314985275,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 3.884854793548584,
+ "step": 2007
+ },
+ {
+ "epoch": 27.8935427574171,
+ "grad_norm": 0.05622334033250809,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 3.853909492492676,
+ "step": 2008
+ },
+ {
+ "epoch": 27.907504363001745,
+ "grad_norm": 0.05513857677578926,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 3.888460159301758,
+ "step": 2009
+ },
+ {
+ "epoch": 27.921465968586386,
+ "grad_norm": 0.05537945404648781,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 3.891343593597412,
+ "step": 2010
+ },
+ {
+ "epoch": 27.93542757417103,
+ "grad_norm": 0.05649508908390999,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 3.8807168006896973,
+ "step": 2011
+ },
+ {
+ "epoch": 27.949389179755673,
+ "grad_norm": 0.054261550307273865,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 3.8827662467956543,
+ "step": 2012
+ },
+ {
+ "epoch": 27.963350785340314,
+ "grad_norm": 0.05615224689245224,
+ "learning_rate": 0.00025996500713765,
+ "loss": 3.8720929622650146,
+ "step": 2013
+ },
+ {
+ "epoch": 27.977312390924958,
+ "grad_norm": 0.05571882426738739,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 3.8978285789489746,
+ "step": 2014
+ },
+ {
+ "epoch": 27.991273996509598,
+ "grad_norm": 0.05477411672472954,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 3.875284194946289,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.03992476686835289,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 2.404466152191162,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 0.6039577126502991,
+ "eval_runtime": 59.8223,
+ "eval_samples_per_second": 40.821,
+ "eval_steps_per_second": 0.652,
+ "step": 2016
+ },
+ {
+ "epoch": 28.013961605584644,
+ "grad_norm": 0.05395985022187233,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 3.78012752532959,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027923211169284,
+ "grad_norm": 0.05445432290434837,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 3.794297218322754,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041884816753928,
+ "grad_norm": 0.05048619955778122,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 3.8059635162353516,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05584642233857,
+ "grad_norm": 0.054649483412504196,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 3.8080105781555176,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069808027923212,
+ "grad_norm": 0.054571982473134995,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 3.8171639442443848,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083769633507853,
+ "grad_norm": 0.05476690083742142,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 3.7937488555908203,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097731239092496,
+ "grad_norm": 0.05586248263716698,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 3.804347515106201,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111692844677137,
+ "grad_norm": 0.05466577783226967,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 3.8114442825317383,
+ "step": 2024
+ },
+ {
+ "epoch": 28.12565445026178,
+ "grad_norm": 0.05342315137386322,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 3.8107991218566895,
+ "step": 2025
+ },
+ {
+ "epoch": 28.13961605584642,
+ "grad_norm": 0.05318617820739746,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 3.8390064239501953,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153577661431065,
+ "grad_norm": 0.05371803045272827,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 3.8021721839904785,
+ "step": 2027
+ },
+ {
+ "epoch": 28.167539267015705,
+ "grad_norm": 0.05319817736744881,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 3.844275712966919,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18150087260035,
+ "grad_norm": 0.05202893167734146,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 3.8008670806884766,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195462478184993,
+ "grad_norm": 0.054026514291763306,
+ "learning_rate": 0.000255486047794226,
+ "loss": 3.81990909576416,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209424083769633,
+ "grad_norm": 0.05195610597729683,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 3.8142457008361816,
+ "step": 2031
+ },
+ {
+ "epoch": 28.223385689354277,
+ "grad_norm": 0.05212428793311119,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 3.8139138221740723,
+ "step": 2032
+ },
+ {
+ "epoch": 28.237347294938917,
+ "grad_norm": 0.05503816530108452,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 3.8072667121887207,
+ "step": 2033
+ },
+ {
+ "epoch": 28.25130890052356,
+ "grad_norm": 0.05305452272295952,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 3.8146510124206543,
+ "step": 2034
+ },
+ {
+ "epoch": 28.2652705061082,
+ "grad_norm": 0.05301898717880249,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 3.7936806678771973,
+ "step": 2035
+ },
+ {
+ "epoch": 28.279232111692846,
+ "grad_norm": 0.0542520172894001,
+ "learning_rate": 0.000253907826333243,
+ "loss": 3.7920427322387695,
+ "step": 2036
+ },
+ {
+ "epoch": 28.293193717277486,
+ "grad_norm": 0.05349896103143692,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 3.819732189178467,
+ "step": 2037
+ },
+ {
+ "epoch": 28.30715532286213,
+ "grad_norm": 0.05463109537959099,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 3.804321765899658,
+ "step": 2038
+ },
+ {
+ "epoch": 28.32111692844677,
+ "grad_norm": 0.05445708706974983,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 3.8128511905670166,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335078534031414,
+ "grad_norm": 0.05655837804079056,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 3.798435688018799,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349040139616054,
+ "grad_norm": 0.05543001368641853,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 3.779137134552002,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363001745200698,
+ "grad_norm": 0.05458657816052437,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 3.833167552947998,
+ "step": 2042
+ },
+ {
+ "epoch": 28.376963350785342,
+ "grad_norm": 0.056820500642061234,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 3.8207151889801025,
+ "step": 2043
+ },
+ {
+ "epoch": 28.390924956369982,
+ "grad_norm": 0.05464114993810654,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 3.788215160369873,
+ "step": 2044
+ },
+ {
+ "epoch": 28.404886561954626,
+ "grad_norm": 0.053662411868572235,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 3.8059287071228027,
+ "step": 2045
+ },
+ {
+ "epoch": 28.418848167539267,
+ "grad_norm": 0.055799756199121475,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 3.8050360679626465,
+ "step": 2046
+ },
+ {
+ "epoch": 28.43280977312391,
+ "grad_norm": 0.05412062630057335,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 3.8110790252685547,
+ "step": 2047
+ },
+ {
+ "epoch": 28.44677137870855,
+ "grad_norm": 0.05672532692551613,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 3.841888904571533,
+ "step": 2048
+ },
+ {
+ "epoch": 28.460732984293195,
+ "grad_norm": 0.05622367933392525,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 3.8200669288635254,
+ "step": 2049
+ },
+ {
+ "epoch": 28.474694589877835,
+ "grad_norm": 0.0544997975230217,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 3.8031818866729736,
+ "step": 2050
+ },
+ {
+ "epoch": 28.48865619546248,
+ "grad_norm": 0.058867089450359344,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 3.856843948364258,
+ "step": 2051
+ },
+ {
+ "epoch": 28.50261780104712,
+ "grad_norm": 0.055324845016002655,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 3.807135820388794,
+ "step": 2052
+ },
+ {
+ "epoch": 28.516579406631763,
+ "grad_norm": 0.055414266884326935,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 3.8380215167999268,
+ "step": 2053
+ },
+ {
+ "epoch": 28.530541012216403,
+ "grad_norm": 0.05787660554051399,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 3.831721782684326,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544502617801047,
+ "grad_norm": 0.05551844462752342,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 3.8572306632995605,
+ "step": 2055
+ },
+ {
+ "epoch": 28.55846422338569,
+ "grad_norm": 0.05452537164092064,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 3.840963840484619,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57242582897033,
+ "grad_norm": 0.05530945956707001,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 3.840442419052124,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586387434554975,
+ "grad_norm": 0.05631886050105095,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 3.7851438522338867,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600349040139616,
+ "grad_norm": 0.05489915981888771,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 3.7870893478393555,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61431064572426,
+ "grad_norm": 0.055331867188215256,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 3.8441572189331055,
+ "step": 2060
+ },
+ {
+ "epoch": 28.6282722513089,
+ "grad_norm": 0.05580870062112808,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 3.834148645401001,
+ "step": 2061
+ },
+ {
+ "epoch": 28.642233856893544,
+ "grad_norm": 0.05681769549846649,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 3.8349759578704834,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656195462478184,
+ "grad_norm": 0.05484442040324211,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 3.798792600631714,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670157068062828,
+ "grad_norm": 0.05657925084233284,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 3.853915214538574,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68411867364747,
+ "grad_norm": 0.05419609695672989,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 3.8423995971679688,
+ "step": 2065
+ },
+ {
+ "epoch": 28.698080279232112,
+ "grad_norm": 0.05700303241610527,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 3.822547674179077,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712041884816752,
+ "grad_norm": 0.055846985429525375,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 3.830044984817505,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726003490401396,
+ "grad_norm": 0.05754954740405083,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 3.868316173553467,
+ "step": 2068
+ },
+ {
+ "epoch": 28.739965095986037,
+ "grad_norm": 0.056265417486429214,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 3.8297243118286133,
+ "step": 2069
+ },
+ {
+ "epoch": 28.75392670157068,
+ "grad_norm": 0.05429232493042946,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 3.872836112976074,
+ "step": 2070
+ },
+ {
+ "epoch": 28.767888307155324,
+ "grad_norm": 0.05589764565229416,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 3.838311195373535,
+ "step": 2071
+ },
+ {
+ "epoch": 28.781849912739965,
+ "grad_norm": 0.05460003390908241,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 3.8316597938537598,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79581151832461,
+ "grad_norm": 0.054542504251003265,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 3.816542625427246,
+ "step": 2073
+ },
+ {
+ "epoch": 28.80977312390925,
+ "grad_norm": 0.0549996979534626,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 3.796245574951172,
+ "step": 2074
+ },
+ {
+ "epoch": 28.823734729493893,
+ "grad_norm": 0.05521940439939499,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 3.8112473487854004,
+ "step": 2075
+ },
+ {
+ "epoch": 28.837696335078533,
+ "grad_norm": 0.05690402165055275,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 3.8242125511169434,
+ "step": 2076
+ },
+ {
+ "epoch": 28.851657940663177,
+ "grad_norm": 0.055383820086717606,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 3.84938383102417,
+ "step": 2077
+ },
+ {
+ "epoch": 28.865619546247817,
+ "grad_norm": 0.05612736567854881,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 3.8383970260620117,
+ "step": 2078
+ },
+ {
+ "epoch": 28.87958115183246,
+ "grad_norm": 0.05723752826452255,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 3.828171730041504,
+ "step": 2079
+ },
+ {
+ "epoch": 28.8935427574171,
+ "grad_norm": 0.0576927475631237,
+ "learning_rate": 0.000242380656502223,
+ "loss": 3.8756160736083984,
+ "step": 2080
+ },
+ {
+ "epoch": 28.907504363001745,
+ "grad_norm": 0.05609026178717613,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 3.8703267574310303,
+ "step": 2081
+ },
+ {
+ "epoch": 28.921465968586386,
+ "grad_norm": 0.057038433849811554,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 3.8566956520080566,
+ "step": 2082
+ },
+ {
+ "epoch": 28.93542757417103,
+ "grad_norm": 0.05683749541640282,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 3.8409035205841064,
+ "step": 2083
+ },
+ {
+ "epoch": 28.949389179755673,
+ "grad_norm": 0.05682830885052681,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 3.8435707092285156,
+ "step": 2084
+ },
+ {
+ "epoch": 28.963350785340314,
+ "grad_norm": 0.05645477771759033,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 3.8142364025115967,
+ "step": 2085
+ },
+ {
+ "epoch": 28.977312390924958,
+ "grad_norm": 0.05519694834947586,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 3.830415725708008,
+ "step": 2086
+ },
+ {
+ "epoch": 28.991273996509598,
+ "grad_norm": 0.05785970389842987,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 3.8413705825805664,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.04018386825919151,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 2.3790979385375977,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 0.6047304272651672,
+ "eval_runtime": 61.1049,
+ "eval_samples_per_second": 39.964,
+ "eval_steps_per_second": 0.638,
+ "step": 2088
+ },
+ {
+ "epoch": 29.013961605584644,
+ "grad_norm": 0.058019861578941345,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 3.793783664703369,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027923211169284,
+ "grad_norm": 0.058003105223178864,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 3.783708095550537,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041884816753928,
+ "grad_norm": 0.05485067516565323,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 3.7829298973083496,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05584642233857,
+ "grad_norm": 0.056009069085121155,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 3.7747392654418945,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069808027923212,
+ "grad_norm": 0.05480542406439781,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 3.774707317352295,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083769633507853,
+ "grad_norm": 0.05384726822376251,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 3.75443959236145,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097731239092496,
+ "grad_norm": 0.053818102926015854,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 3.7790050506591797,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111692844677137,
+ "grad_norm": 0.05595344305038452,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 3.74898624420166,
+ "step": 2096
+ },
+ {
+ "epoch": 29.12565445026178,
+ "grad_norm": 0.05394545570015907,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 3.7751331329345703,
+ "step": 2097
+ },
+ {
+ "epoch": 29.13961605584642,
+ "grad_norm": 0.05346408113837242,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 3.7671167850494385,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153577661431065,
+ "grad_norm": 0.0541236512362957,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 3.749098062515259,
+ "step": 2099
+ },
+ {
+ "epoch": 29.167539267015705,
+ "grad_norm": 0.05369208753108978,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 3.7778515815734863,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18150087260035,
+ "grad_norm": 0.055362965911626816,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 3.747697591781616,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195462478184993,
+ "grad_norm": 0.05389911308884621,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 3.7759058475494385,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209424083769633,
+ "grad_norm": 0.05315527692437172,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 3.7832705974578857,
+ "step": 2103
+ },
+ {
+ "epoch": 29.223385689354277,
+ "grad_norm": 0.05377589911222458,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 3.7743210792541504,
+ "step": 2104
+ },
+ {
+ "epoch": 29.237347294938917,
+ "grad_norm": 0.05307800695300102,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 3.7654497623443604,
+ "step": 2105
+ },
+ {
+ "epoch": 29.25130890052356,
+ "grad_norm": 0.056719157844781876,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 3.794179916381836,
+ "step": 2106
+ },
+ {
+ "epoch": 29.2652705061082,
+ "grad_norm": 0.053281862288713455,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 3.7620739936828613,
+ "step": 2107
+ },
+ {
+ "epoch": 29.279232111692846,
+ "grad_norm": 0.05470574274659157,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 3.788421869277954,
+ "step": 2108
+ },
+ {
+ "epoch": 29.293193717277486,
+ "grad_norm": 0.05293853580951691,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 3.7655067443847656,
+ "step": 2109
+ },
+ {
+ "epoch": 29.30715532286213,
+ "grad_norm": 0.05524440482258797,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 3.7701616287231445,
+ "step": 2110
+ },
+ {
+ "epoch": 29.32111692844677,
+ "grad_norm": 0.05508396402001381,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 3.7993319034576416,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335078534031414,
+ "grad_norm": 0.05559603124856949,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 3.7799668312072754,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349040139616054,
+ "grad_norm": 0.05530122295022011,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 3.751032829284668,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363001745200698,
+ "grad_norm": 0.05408896878361702,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 3.7761611938476562,
+ "step": 2114
+ },
+ {
+ "epoch": 29.376963350785342,
+ "grad_norm": 0.05530737340450287,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 3.745713233947754,
+ "step": 2115
+ },
+ {
+ "epoch": 29.390924956369982,
+ "grad_norm": 0.05424542352557182,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 3.8041834831237793,
+ "step": 2116
+ },
+ {
+ "epoch": 29.404886561954626,
+ "grad_norm": 0.054800380021333694,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 3.7995352745056152,
+ "step": 2117
+ },
+ {
+ "epoch": 29.418848167539267,
+ "grad_norm": 0.05685900151729584,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 3.7970776557922363,
+ "step": 2118
+ },
+ {
+ "epoch": 29.43280977312391,
+ "grad_norm": 0.05555475875735283,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 3.7950801849365234,
+ "step": 2119
+ },
+ {
+ "epoch": 29.44677137870855,
+ "grad_norm": 0.05670571327209473,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 3.7967746257781982,
+ "step": 2120
+ },
+ {
+ "epoch": 29.460732984293195,
+ "grad_norm": 0.05598503723740578,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 3.809349298477173,
+ "step": 2121
+ },
+ {
+ "epoch": 29.474694589877835,
+ "grad_norm": 0.05501085892319679,
+ "learning_rate": 0.000231465389734324,
+ "loss": 3.803968906402588,
+ "step": 2122
+ },
+ {
+ "epoch": 29.48865619546248,
+ "grad_norm": 0.05519728362560272,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 3.7615163326263428,
+ "step": 2123
+ },
+ {
+ "epoch": 29.50261780104712,
+ "grad_norm": 0.055866893380880356,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 3.7939207553863525,
+ "step": 2124
+ },
+ {
+ "epoch": 29.516579406631763,
+ "grad_norm": 0.05460986867547035,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 3.7740321159362793,
+ "step": 2125
+ },
+ {
+ "epoch": 29.530541012216403,
+ "grad_norm": 0.055752526968717575,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 3.82424259185791,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544502617801047,
+ "grad_norm": 0.056792002171278,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 3.8006062507629395,
+ "step": 2127
+ },
+ {
+ "epoch": 29.55846422338569,
+ "grad_norm": 0.054986681789159775,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 3.7713897228240967,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57242582897033,
+ "grad_norm": 0.054215673357248306,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 3.8005690574645996,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586387434554975,
+ "grad_norm": 0.05700439214706421,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 3.825161933898926,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600349040139616,
+ "grad_norm": 0.055078551173210144,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 3.786716938018799,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61431064572426,
+ "grad_norm": 0.055569086223840714,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 3.8020172119140625,
+ "step": 2132
+ },
+ {
+ "epoch": 29.6282722513089,
+ "grad_norm": 0.0548536442220211,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 3.8038570880889893,
+ "step": 2133
+ },
+ {
+ "epoch": 29.642233856893544,
+ "grad_norm": 0.05604325234889984,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 3.796006441116333,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656195462478184,
+ "grad_norm": 0.05614005774259567,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 3.7938876152038574,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670157068062828,
+ "grad_norm": 0.05761241167783737,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 3.813568592071533,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68411867364747,
+ "grad_norm": 0.05579913780093193,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 3.826862335205078,
+ "step": 2137
+ },
+ {
+ "epoch": 29.698080279232112,
+ "grad_norm": 0.05784283205866814,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 3.808281421661377,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712041884816752,
+ "grad_norm": 0.05596453696489334,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 3.8119921684265137,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726003490401396,
+ "grad_norm": 0.05699361115694046,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 3.822702407836914,
+ "step": 2140
+ },
+ {
+ "epoch": 29.739965095986037,
+ "grad_norm": 0.05375700071454048,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 3.78023099899292,
+ "step": 2141
+ },
+ {
+ "epoch": 29.75392670157068,
+ "grad_norm": 0.05588206648826599,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 3.791456699371338,
+ "step": 2142
+ },
+ {
+ "epoch": 29.767888307155324,
+ "grad_norm": 0.05579312890768051,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 3.799713134765625,
+ "step": 2143
+ },
+ {
+ "epoch": 29.781849912739965,
+ "grad_norm": 0.05713854730129242,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 3.8211703300476074,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79581151832461,
+ "grad_norm": 0.05645165592432022,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 3.7947046756744385,
+ "step": 2145
+ },
+ {
+ "epoch": 29.80977312390925,
+ "grad_norm": 0.056201402097940445,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 3.811009407043457,
+ "step": 2146
+ },
+ {
+ "epoch": 29.823734729493893,
+ "grad_norm": 0.05659479275345802,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 3.8009982109069824,
+ "step": 2147
+ },
+ {
+ "epoch": 29.837696335078533,
+ "grad_norm": 0.05797034874558449,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 3.7994565963745117,
+ "step": 2148
+ },
+ {
+ "epoch": 29.851657940663177,
+ "grad_norm": 0.05734949931502342,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 3.7696728706359863,
+ "step": 2149
+ },
+ {
+ "epoch": 29.865619546247817,
+ "grad_norm": 0.060396358370780945,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 3.8269689083099365,
+ "step": 2150
+ },
+ {
+ "epoch": 29.87958115183246,
+ "grad_norm": 0.057348065078258514,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 3.7925050258636475,
+ "step": 2151
+ },
+ {
+ "epoch": 29.8935427574171,
+ "grad_norm": 0.058563247323036194,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 3.82887864112854,
+ "step": 2152
+ },
+ {
+ "epoch": 29.907504363001745,
+ "grad_norm": 0.06011585891246796,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 3.8061296939849854,
+ "step": 2153
+ },
+ {
+ "epoch": 29.921465968586386,
+ "grad_norm": 0.058440592139959335,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 3.792586326599121,
+ "step": 2154
+ },
+ {
+ "epoch": 29.93542757417103,
+ "grad_norm": 0.057050157338380814,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 3.7885212898254395,
+ "step": 2155
+ },
+ {
+ "epoch": 29.949389179755673,
+ "grad_norm": 0.05920996889472008,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 3.813566207885742,
+ "step": 2156
+ },
+ {
+ "epoch": 29.963350785340314,
+ "grad_norm": 0.057791322469711304,
+ "learning_rate": 0.00022244633283095,
+ "loss": 3.8399810791015625,
+ "step": 2157
+ },
+ {
+ "epoch": 29.977312390924958,
+ "grad_norm": 0.05644001066684723,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 3.79464054107666,
+ "step": 2158
+ },
+ {
+ "epoch": 29.991273996509598,
+ "grad_norm": 0.06021320819854736,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 3.804086208343506,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.040621865540742874,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 2.386565923690796,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 0.6063579320907593,
+ "eval_runtime": 59.5345,
+ "eval_samples_per_second": 41.018,
+ "eval_steps_per_second": 0.655,
+ "step": 2160
+ },
+ {
+ "epoch": 30.013961605584644,
+ "grad_norm": 0.05961514264345169,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 3.756401777267456,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027923211169284,
+ "grad_norm": 0.057169776409864426,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 3.7327475547790527,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041884816753928,
+ "grad_norm": 0.056084226816892624,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 3.7365498542785645,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05584642233857,
+ "grad_norm": 0.059099674224853516,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 3.7383875846862793,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069808027923212,
+ "grad_norm": 0.05632825568318367,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 3.7642762660980225,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083769633507853,
+ "grad_norm": 0.056937988847494125,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 3.7614428997039795,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097731239092496,
+ "grad_norm": 0.05722609534859657,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 3.7394845485687256,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111692844677137,
+ "grad_norm": 0.056204989552497864,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 3.7473866939544678,
+ "step": 2168
+ },
+ {
+ "epoch": 30.12565445026178,
+ "grad_norm": 0.05640130862593651,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 3.7504067420959473,
+ "step": 2169
+ },
+ {
+ "epoch": 30.13961605584642,
+ "grad_norm": 0.05625637248158455,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 3.754812717437744,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153577661431065,
+ "grad_norm": 0.05596087500452995,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 3.741854429244995,
+ "step": 2171
+ },
+ {
+ "epoch": 30.167539267015705,
+ "grad_norm": 0.054819896817207336,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 3.724705934524536,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18150087260035,
+ "grad_norm": 0.05644344910979271,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 3.7609610557556152,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195462478184993,
+ "grad_norm": 0.05366845801472664,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 3.7557997703552246,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209424083769633,
+ "grad_norm": 0.05517663434147835,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 3.7494120597839355,
+ "step": 2175
+ },
+ {
+ "epoch": 30.223385689354277,
+ "grad_norm": 0.05463553965091705,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 3.7434799671173096,
+ "step": 2176
+ },
+ {
+ "epoch": 30.237347294938917,
+ "grad_norm": 0.05352703109383583,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 3.7435178756713867,
+ "step": 2177
+ },
+ {
+ "epoch": 30.25130890052356,
+ "grad_norm": 0.053156204521656036,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 3.7648849487304688,
+ "step": 2178
+ },
+ {
+ "epoch": 30.2652705061082,
+ "grad_norm": 0.05557673051953316,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 3.746366024017334,
+ "step": 2179
+ },
+ {
+ "epoch": 30.279232111692846,
+ "grad_norm": 0.052989061921834946,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 3.7453460693359375,
+ "step": 2180
+ },
+ {
+ "epoch": 30.293193717277486,
+ "grad_norm": 0.055707018822431564,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 3.800962209701538,
+ "step": 2181
+ },
+ {
+ "epoch": 30.30715532286213,
+ "grad_norm": 0.05560754984617233,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 3.7416906356811523,
+ "step": 2182
+ },
+ {
+ "epoch": 30.32111692844677,
+ "grad_norm": 0.055344413965940475,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 3.7672080993652344,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335078534031414,
+ "grad_norm": 0.05490704998373985,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 3.7664341926574707,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349040139616054,
+ "grad_norm": 0.0578596331179142,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 3.7632641792297363,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363001745200698,
+ "grad_norm": 0.055299609899520874,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 3.7502024173736572,
+ "step": 2186
+ },
+ {
+ "epoch": 30.376963350785342,
+ "grad_norm": 0.055016499012708664,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 3.732128143310547,
+ "step": 2187
+ },
+ {
+ "epoch": 30.390924956369982,
+ "grad_norm": 0.057352419942617416,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 3.723008871078491,
+ "step": 2188
+ },
+ {
+ "epoch": 30.404886561954626,
+ "grad_norm": 0.05596115067601204,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 3.786574363708496,
+ "step": 2189
+ },
+ {
+ "epoch": 30.418848167539267,
+ "grad_norm": 0.05596112087368965,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 3.770382881164551,
+ "step": 2190
+ },
+ {
+ "epoch": 30.43280977312391,
+ "grad_norm": 0.0562058687210083,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 3.7605714797973633,
+ "step": 2191
+ },
+ {
+ "epoch": 30.44677137870855,
+ "grad_norm": 0.056407757103443146,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 3.7485427856445312,
+ "step": 2192
+ },
+ {
+ "epoch": 30.460732984293195,
+ "grad_norm": 0.05596652254462242,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 3.7630670070648193,
+ "step": 2193
+ },
+ {
+ "epoch": 30.474694589877835,
+ "grad_norm": 0.056551866233348846,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 3.7981796264648438,
+ "step": 2194
+ },
+ {
+ "epoch": 30.48865619546248,
+ "grad_norm": 0.057354725897312164,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 3.7544898986816406,
+ "step": 2195
+ },
+ {
+ "epoch": 30.50261780104712,
+ "grad_norm": 0.05791718140244484,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 3.7241907119750977,
+ "step": 2196
+ },
+ {
+ "epoch": 30.516579406631763,
+ "grad_norm": 0.05675847455859184,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 3.794506549835205,
+ "step": 2197
+ },
+ {
+ "epoch": 30.530541012216403,
+ "grad_norm": 0.060349952429533005,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 3.75946044921875,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544502617801047,
+ "grad_norm": 0.05547260120511055,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 3.7278571128845215,
+ "step": 2199
+ },
+ {
+ "epoch": 30.55846422338569,
+ "grad_norm": 0.058400996029376984,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 3.756891965866089,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57242582897033,
+ "grad_norm": 0.057871121913194656,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 3.783863067626953,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586387434554975,
+ "grad_norm": 0.05899912118911743,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 3.7787115573883057,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600349040139616,
+ "grad_norm": 0.056164853274822235,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 3.7599334716796875,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61431064572426,
+ "grad_norm": 0.05775163322687149,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 3.7457070350646973,
+ "step": 2204
+ },
+ {
+ "epoch": 30.6282722513089,
+ "grad_norm": 0.057221125811338425,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 3.7663235664367676,
+ "step": 2205
+ },
+ {
+ "epoch": 30.642233856893544,
+ "grad_norm": 0.05788435414433479,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 3.7676830291748047,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656195462478184,
+ "grad_norm": 0.057100843638181686,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 3.7586870193481445,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670157068062828,
+ "grad_norm": 0.05703200399875641,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 3.7561769485473633,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68411867364747,
+ "grad_norm": 0.05664575845003128,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 3.761331081390381,
+ "step": 2209
+ },
+ {
+ "epoch": 30.698080279232112,
+ "grad_norm": 0.056622982025146484,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 3.737915515899658,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712041884816752,
+ "grad_norm": 0.056481845676898956,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 3.7413077354431152,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726003490401396,
+ "grad_norm": 0.057718146592378616,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 3.751267433166504,
+ "step": 2212
+ },
+ {
+ "epoch": 30.739965095986037,
+ "grad_norm": 0.05772862955927849,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 3.753678321838379,
+ "step": 2213
+ },
+ {
+ "epoch": 30.75392670157068,
+ "grad_norm": 0.056571926921606064,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 3.751033306121826,
+ "step": 2214
+ },
+ {
+ "epoch": 30.767888307155324,
+ "grad_norm": 0.05682666599750519,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 3.763153553009033,
+ "step": 2215
+ },
+ {
+ "epoch": 30.781849912739965,
+ "grad_norm": 0.05993897467851639,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 3.787890911102295,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79581151832461,
+ "grad_norm": 0.05659458413720131,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 3.7655086517333984,
+ "step": 2217
+ },
+ {
+ "epoch": 30.80977312390925,
+ "grad_norm": 0.057996224611997604,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 3.774881601333618,
+ "step": 2218
+ },
+ {
+ "epoch": 30.823734729493893,
+ "grad_norm": 0.05677317455410957,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 3.7600479125976562,
+ "step": 2219
+ },
+ {
+ "epoch": 30.837696335078533,
+ "grad_norm": 0.05664737895131111,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 3.787637233734131,
+ "step": 2220
+ },
+ {
+ "epoch": 30.851657940663177,
+ "grad_norm": 0.05808616429567337,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 3.7529220581054688,
+ "step": 2221
+ },
+ {
+ "epoch": 30.865619546247817,
+ "grad_norm": 0.05666997656226158,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 3.7790794372558594,
+ "step": 2222
+ },
+ {
+ "epoch": 30.87958115183246,
+ "grad_norm": 0.05963682755827904,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 3.762615203857422,
+ "step": 2223
+ },
+ {
+ "epoch": 30.8935427574171,
+ "grad_norm": 0.05558215081691742,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 3.7871851921081543,
+ "step": 2224
+ },
+ {
+ "epoch": 30.907504363001745,
+ "grad_norm": 0.05762261897325516,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 3.7605397701263428,
+ "step": 2225
+ },
+ {
+ "epoch": 30.921465968586386,
+ "grad_norm": 0.0565858818590641,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 3.7693653106689453,
+ "step": 2226
+ },
+ {
+ "epoch": 30.93542757417103,
+ "grad_norm": 0.05821581557393074,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 3.766758918762207,
+ "step": 2227
+ },
+ {
+ "epoch": 30.949389179755673,
+ "grad_norm": 0.055186714977025986,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 3.79498553276062,
+ "step": 2228
+ },
+ {
+ "epoch": 30.963350785340314,
+ "grad_norm": 0.05850847810506821,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 3.799769163131714,
+ "step": 2229
+ },
+ {
+ "epoch": 30.977312390924958,
+ "grad_norm": 0.05700397863984108,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 3.7555625438690186,
+ "step": 2230
+ },
+ {
+ "epoch": 30.991273996509598,
+ "grad_norm": 0.057813212275505066,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 3.777778148651123,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.041730619966983795,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 2.3564491271972656,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 0.6080329418182373,
+ "eval_runtime": 60.4607,
+ "eval_samples_per_second": 40.39,
+ "eval_steps_per_second": 0.645,
+ "step": 2232
+ },
+ {
+ "epoch": 31.013961605584644,
+ "grad_norm": 0.0563117116689682,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 3.719982624053955,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027923211169284,
+ "grad_norm": 0.05564197897911072,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 3.719735622406006,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041884816753928,
+ "grad_norm": 0.05562874674797058,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 3.739713430404663,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05584642233857,
+ "grad_norm": 0.057146716862916946,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 3.736401081085205,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069808027923212,
+ "grad_norm": 0.053163520991802216,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 3.7049455642700195,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083769633507853,
+ "grad_norm": 0.057067256420850754,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 3.723662853240967,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097731239092496,
+ "grad_norm": 0.05276816710829735,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 3.710062026977539,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111692844677137,
+ "grad_norm": 0.055126190185546875,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 3.696173667907715,
+ "step": 2240
+ },
+ {
+ "epoch": 31.12565445026178,
+ "grad_norm": 0.05533793941140175,
+ "learning_rate": 0.000201141724176723,
+ "loss": 3.710416793823242,
+ "step": 2241
+ },
+ {
+ "epoch": 31.13961605584642,
+ "grad_norm": 0.055686868727207184,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 3.7188220024108887,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153577661431065,
+ "grad_norm": 0.055188920348882675,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 3.7026562690734863,
+ "step": 2243
+ },
+ {
+ "epoch": 31.167539267015705,
+ "grad_norm": 0.053627416491508484,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 3.7224509716033936,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18150087260035,
+ "grad_norm": 0.05535290017724037,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 3.710698366165161,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195462478184993,
+ "grad_norm": 0.0528891459107399,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 3.6979012489318848,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209424083769633,
+ "grad_norm": 0.055201176553964615,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 3.7324817180633545,
+ "step": 2247
+ },
+ {
+ "epoch": 31.223385689354277,
+ "grad_norm": 0.05338778346776962,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 3.722564697265625,
+ "step": 2248
+ },
+ {
+ "epoch": 31.237347294938917,
+ "grad_norm": 0.05368843302130699,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 3.7118425369262695,
+ "step": 2249
+ },
+ {
+ "epoch": 31.25130890052356,
+ "grad_norm": 0.052788376808166504,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 3.6875901222229004,
+ "step": 2250
+ },
+ {
+ "epoch": 31.2652705061082,
+ "grad_norm": 0.05427199602127075,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 3.7042768001556396,
+ "step": 2251
+ },
+ {
+ "epoch": 31.279232111692846,
+ "grad_norm": 0.05434224382042885,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 3.7138795852661133,
+ "step": 2252
+ },
+ {
+ "epoch": 31.293193717277486,
+ "grad_norm": 0.054364923387765884,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 3.735142946243286,
+ "step": 2253
+ },
+ {
+ "epoch": 31.30715532286213,
+ "grad_norm": 0.05535430088639259,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 3.7198615074157715,
+ "step": 2254
+ },
+ {
+ "epoch": 31.32111692844677,
+ "grad_norm": 0.05463665351271629,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 3.733370542526245,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335078534031414,
+ "grad_norm": 0.0553361251950264,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 3.702805519104004,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349040139616054,
+ "grad_norm": 0.05423365905880928,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 3.7315120697021484,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363001745200698,
+ "grad_norm": 0.05499394237995148,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 3.731891632080078,
+ "step": 2258
+ },
+ {
+ "epoch": 31.376963350785342,
+ "grad_norm": 0.054968126118183136,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 3.7355899810791016,
+ "step": 2259
+ },
+ {
+ "epoch": 31.390924956369982,
+ "grad_norm": 0.05478367581963539,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 3.7143867015838623,
+ "step": 2260
+ },
+ {
+ "epoch": 31.404886561954626,
+ "grad_norm": 0.05554179847240448,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 3.7227137088775635,
+ "step": 2261
+ },
+ {
+ "epoch": 31.418848167539267,
+ "grad_norm": 0.05399509146809578,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 3.750650405883789,
+ "step": 2262
+ },
+ {
+ "epoch": 31.43280977312391,
+ "grad_norm": 0.056201111525297165,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 3.7478344440460205,
+ "step": 2263
+ },
+ {
+ "epoch": 31.44677137870855,
+ "grad_norm": 0.05437091737985611,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 3.7283718585968018,
+ "step": 2264
+ },
+ {
+ "epoch": 31.460732984293195,
+ "grad_norm": 0.057144906371831894,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 3.7482407093048096,
+ "step": 2265
+ },
+ {
+ "epoch": 31.474694589877835,
+ "grad_norm": 0.05441201478242874,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 3.726378917694092,
+ "step": 2266
+ },
+ {
+ "epoch": 31.48865619546248,
+ "grad_norm": 0.056569576263427734,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 3.7211437225341797,
+ "step": 2267
+ },
+ {
+ "epoch": 31.50261780104712,
+ "grad_norm": 0.05561831593513489,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 3.7278850078582764,
+ "step": 2268
+ },
+ {
+ "epoch": 31.516579406631763,
+ "grad_norm": 0.0569809265434742,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 3.713228702545166,
+ "step": 2269
+ },
+ {
+ "epoch": 31.530541012216403,
+ "grad_norm": 0.05654345825314522,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 3.7354423999786377,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544502617801047,
+ "grad_norm": 0.05482077598571777,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 3.7043380737304688,
+ "step": 2271
+ },
+ {
+ "epoch": 31.55846422338569,
+ "grad_norm": 0.055417660623788834,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 3.7199621200561523,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57242582897033,
+ "grad_norm": 0.054656337946653366,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 3.734508514404297,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586387434554975,
+ "grad_norm": 0.05609512701630592,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 3.7341227531433105,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600349040139616,
+ "grad_norm": 0.05602060630917549,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 3.719949960708618,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61431064572426,
+ "grad_norm": 0.05535270646214485,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 3.7124414443969727,
+ "step": 2276
+ },
+ {
+ "epoch": 31.6282722513089,
+ "grad_norm": 0.05577492713928223,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 3.706301212310791,
+ "step": 2277
+ },
+ {
+ "epoch": 31.642233856893544,
+ "grad_norm": 0.057382550090551376,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 3.7411818504333496,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656195462478184,
+ "grad_norm": 0.056383028626441956,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 3.751235008239746,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670157068062828,
+ "grad_norm": 0.05667940899729729,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 3.7379674911499023,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68411867364747,
+ "grad_norm": 0.05635855346918106,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 3.728001594543457,
+ "step": 2281
+ },
+ {
+ "epoch": 31.698080279232112,
+ "grad_norm": 0.05649873986840248,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 3.740757465362549,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712041884816752,
+ "grad_norm": 0.055871494114398956,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 3.7373616695404053,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726003490401396,
+ "grad_norm": 0.057106468826532364,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 3.7572178840637207,
+ "step": 2284
+ },
+ {
+ "epoch": 31.739965095986037,
+ "grad_norm": 0.05788225308060646,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 3.7474923133850098,
+ "step": 2285
+ },
+ {
+ "epoch": 31.75392670157068,
+ "grad_norm": 0.057152729481458664,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 3.7224936485290527,
+ "step": 2286
+ },
+ {
+ "epoch": 31.767888307155324,
+ "grad_norm": 0.0560687892138958,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 3.7336764335632324,
+ "step": 2287
+ },
+ {
+ "epoch": 31.781849912739965,
+ "grad_norm": 0.05890286713838577,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 3.754326343536377,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79581151832461,
+ "grad_norm": 0.058598410338163376,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 3.7265028953552246,
+ "step": 2289
+ },
+ {
+ "epoch": 31.80977312390925,
+ "grad_norm": 0.05728591978549957,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 3.715481758117676,
+ "step": 2290
+ },
+ {
+ "epoch": 31.823734729493893,
+ "grad_norm": 0.05874952673912048,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 3.7139101028442383,
+ "step": 2291
+ },
+ {
+ "epoch": 31.837696335078533,
+ "grad_norm": 0.056882575154304504,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 3.728553533554077,
+ "step": 2292
+ },
+ {
+ "epoch": 31.851657940663177,
+ "grad_norm": 0.057810861617326736,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 3.757908582687378,
+ "step": 2293
+ },
+ {
+ "epoch": 31.865619546247817,
+ "grad_norm": 0.057366516441106796,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 3.7265639305114746,
+ "step": 2294
+ },
+ {
+ "epoch": 31.87958115183246,
+ "grad_norm": 0.059405453503131866,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 3.740511417388916,
+ "step": 2295
+ },
+ {
+ "epoch": 31.8935427574171,
+ "grad_norm": 0.05881514027714729,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 3.751497268676758,
+ "step": 2296
+ },
+ {
+ "epoch": 31.907504363001745,
+ "grad_norm": 0.06030813232064247,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 3.75439715385437,
+ "step": 2297
+ },
+ {
+ "epoch": 31.921465968586386,
+ "grad_norm": 0.06062643602490425,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 3.7215723991394043,
+ "step": 2298
+ },
+ {
+ "epoch": 31.93542757417103,
+ "grad_norm": 0.05953266844153404,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 3.7635340690612793,
+ "step": 2299
+ },
+ {
+ "epoch": 31.949389179755673,
+ "grad_norm": 0.06019049137830734,
+ "learning_rate": 0.000186516746349841,
+ "loss": 3.7750091552734375,
+ "step": 2300
+ },
+ {
+ "epoch": 31.963350785340314,
+ "grad_norm": 0.058755744248628616,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 3.7646124362945557,
+ "step": 2301
+ },
+ {
+ "epoch": 31.977312390924958,
+ "grad_norm": 0.05986425653100014,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 3.7566630840301514,
+ "step": 2302
+ },
+ {
+ "epoch": 31.991273996509598,
+ "grad_norm": 0.057212650775909424,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 3.714637279510498,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.04401736706495285,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 2.33223032951355,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 0.6091130375862122,
+ "eval_runtime": 60.7334,
+ "eval_samples_per_second": 40.209,
+ "eval_steps_per_second": 0.642,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01396160558464,
+ "grad_norm": 0.05650010704994202,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 3.702322244644165,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02792321116929,
+ "grad_norm": 0.054223258048295975,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 3.699272632598877,
+ "step": 2306
+ },
+ {
+ "epoch": 32.04188481675393,
+ "grad_norm": 0.05616949126124382,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 3.6740047931671143,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05584642233857,
+ "grad_norm": 0.05479196831583977,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 3.6942710876464844,
+ "step": 2308
+ },
+ {
+ "epoch": 32.06980802792321,
+ "grad_norm": 0.05735877528786659,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 3.6724581718444824,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083769633507856,
+ "grad_norm": 0.056006450206041336,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 3.6736817359924316,
+ "step": 2310
+ },
+ {
+ "epoch": 32.0977312390925,
+ "grad_norm": 0.055535659193992615,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 3.710303783416748,
+ "step": 2311
+ },
+ {
+ "epoch": 32.11169284467714,
+ "grad_norm": 0.05510629341006279,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 3.669536590576172,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12565445026178,
+ "grad_norm": 0.05426035821437836,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 3.670957088470459,
+ "step": 2313
+ },
+ {
+ "epoch": 32.139616055846425,
+ "grad_norm": 0.05547798424959183,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 3.7050318717956543,
+ "step": 2314
+ },
+ {
+ "epoch": 32.153577661431065,
+ "grad_norm": 0.05412966012954712,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 3.6849310398101807,
+ "step": 2315
+ },
+ {
+ "epoch": 32.167539267015705,
+ "grad_norm": 0.0554192028939724,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 3.702563762664795,
+ "step": 2316
+ },
+ {
+ "epoch": 32.181500872600346,
+ "grad_norm": 0.054564788937568665,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 3.707547187805176,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19546247818499,
+ "grad_norm": 0.05424094945192337,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 3.650768280029297,
+ "step": 2318
+ },
+ {
+ "epoch": 32.20942408376963,
+ "grad_norm": 0.05562194436788559,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 3.6763458251953125,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223385689354274,
+ "grad_norm": 0.053762275725603104,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 3.690493583679199,
+ "step": 2320
+ },
+ {
+ "epoch": 32.23734729493892,
+ "grad_norm": 0.05513778328895569,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 3.6990034580230713,
+ "step": 2321
+ },
+ {
+ "epoch": 32.25130890052356,
+ "grad_norm": 0.05438566580414772,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 3.702357292175293,
+ "step": 2322
+ },
+ {
+ "epoch": 32.2652705061082,
+ "grad_norm": 0.054521508514881134,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 3.701066255569458,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27923211169284,
+ "grad_norm": 0.05416509136557579,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 3.673935651779175,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29319371727749,
+ "grad_norm": 0.05350432172417641,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 3.6740450859069824,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30715532286213,
+ "grad_norm": 0.054736558347940445,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 3.710815668106079,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32111692844677,
+ "grad_norm": 0.055077534168958664,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 3.6732773780822754,
+ "step": 2327
+ },
+ {
+ "epoch": 32.33507853403141,
+ "grad_norm": 0.054749298840761185,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 3.6978728771209717,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34904013961606,
+ "grad_norm": 0.05378924310207367,
+ "learning_rate": 0.000179445406945268,
+ "loss": 3.6783430576324463,
+ "step": 2329
+ },
+ {
+ "epoch": 32.3630017452007,
+ "grad_norm": 0.0553874596953392,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 3.6876044273376465,
+ "step": 2330
+ },
+ {
+ "epoch": 32.37696335078534,
+ "grad_norm": 0.053926583379507065,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 3.712519884109497,
+ "step": 2331
+ },
+ {
+ "epoch": 32.390924956369986,
+ "grad_norm": 0.055778294801712036,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 3.686643123626709,
+ "step": 2332
+ },
+ {
+ "epoch": 32.404886561954626,
+ "grad_norm": 0.05676745995879173,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 3.685737371444702,
+ "step": 2333
+ },
+ {
+ "epoch": 32.41884816753927,
+ "grad_norm": 0.054717857390642166,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 3.6684136390686035,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43280977312391,
+ "grad_norm": 0.0552208311855793,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 3.696580410003662,
+ "step": 2335
+ },
+ {
+ "epoch": 32.446771378708554,
+ "grad_norm": 0.0549275167286396,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 3.7011008262634277,
+ "step": 2336
+ },
+ {
+ "epoch": 32.460732984293195,
+ "grad_norm": 0.05518808960914612,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 3.7204928398132324,
+ "step": 2337
+ },
+ {
+ "epoch": 32.474694589877835,
+ "grad_norm": 0.05621159076690674,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 3.68092679977417,
+ "step": 2338
+ },
+ {
+ "epoch": 32.488656195462475,
+ "grad_norm": 0.054848797619342804,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 3.677091598510742,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50261780104712,
+ "grad_norm": 0.0562039278447628,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 3.7283501625061035,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51657940663176,
+ "grad_norm": 0.05434577167034149,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 3.710252285003662,
+ "step": 2341
+ },
+ {
+ "epoch": 32.5305410122164,
+ "grad_norm": 0.05737415328621864,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 3.7027816772460938,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544502617801044,
+ "grad_norm": 0.057093650102615356,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 3.7027580738067627,
+ "step": 2343
+ },
+ {
+ "epoch": 32.55846422338569,
+ "grad_norm": 0.0550924576818943,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 3.68161678314209,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57242582897033,
+ "grad_norm": 0.057579416781663895,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 3.67604923248291,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58638743455497,
+ "grad_norm": 0.05515696108341217,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 3.693655490875244,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60034904013962,
+ "grad_norm": 0.05788129195570946,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 3.725567102432251,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61431064572426,
+ "grad_norm": 0.05617021024227142,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 3.7027430534362793,
+ "step": 2348
+ },
+ {
+ "epoch": 32.6282722513089,
+ "grad_norm": 0.05649486556649208,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 3.717970371246338,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64223385689354,
+ "grad_norm": 0.05582305043935776,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 3.6795706748962402,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65619546247819,
+ "grad_norm": 0.058177400380373,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 3.6926321983337402,
+ "step": 2351
+ },
+ {
+ "epoch": 32.67015706806283,
+ "grad_norm": 0.05563648045063019,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 3.7227001190185547,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68411867364747,
+ "grad_norm": 0.058799829334020615,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 3.706979513168335,
+ "step": 2353
+ },
+ {
+ "epoch": 32.69808027923211,
+ "grad_norm": 0.05645359680056572,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 3.6872620582580566,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712041884816756,
+ "grad_norm": 0.058064281940460205,
+ "learning_rate": 0.000173176617304673,
+ "loss": 3.695530414581299,
+ "step": 2355
+ },
+ {
+ "epoch": 32.726003490401396,
+ "grad_norm": 0.05756758525967598,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 3.7229819297790527,
+ "step": 2356
+ },
+ {
+ "epoch": 32.73996509598604,
+ "grad_norm": 0.057592105120420456,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 3.7098472118377686,
+ "step": 2357
+ },
+ {
+ "epoch": 32.753926701570684,
+ "grad_norm": 0.05855923518538475,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 3.704477310180664,
+ "step": 2358
+ },
+ {
+ "epoch": 32.767888307155324,
+ "grad_norm": 0.05809807404875755,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 3.7231266498565674,
+ "step": 2359
+ },
+ {
+ "epoch": 32.781849912739965,
+ "grad_norm": 0.05663153529167175,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 3.708165168762207,
+ "step": 2360
+ },
+ {
+ "epoch": 32.795811518324605,
+ "grad_norm": 0.05867684260010719,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 3.70230770111084,
+ "step": 2361
+ },
+ {
+ "epoch": 32.80977312390925,
+ "grad_norm": 0.05850773677229881,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 3.7339577674865723,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82373472949389,
+ "grad_norm": 0.05651238560676575,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 3.7185606956481934,
+ "step": 2363
+ },
+ {
+ "epoch": 32.83769633507853,
+ "grad_norm": 0.05758710578083992,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 3.677966356277466,
+ "step": 2364
+ },
+ {
+ "epoch": 32.85165794066317,
+ "grad_norm": 0.057175327092409134,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 3.720883846282959,
+ "step": 2365
+ },
+ {
+ "epoch": 32.86561954624782,
+ "grad_norm": 0.05690391734242439,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 3.706535577774048,
+ "step": 2366
+ },
+ {
+ "epoch": 32.87958115183246,
+ "grad_norm": 0.05758208781480789,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 3.743523120880127,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8935427574171,
+ "grad_norm": 0.05831100419163704,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 3.726163387298584,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90750436300174,
+ "grad_norm": 0.05664364993572235,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 3.691837787628174,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92146596858639,
+ "grad_norm": 0.05847351998090744,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 3.7386794090270996,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93542757417103,
+ "grad_norm": 0.05673616752028465,
+ "learning_rate": 0.00016935382741164,
+ "loss": 3.72623348236084,
+ "step": 2371
+ },
+ {
+ "epoch": 32.94938917975567,
+ "grad_norm": 0.058967288583517075,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 3.733576536178589,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96335078534032,
+ "grad_norm": 0.056366659700870514,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 3.7006077766418457,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97731239092496,
+ "grad_norm": 0.05691685527563095,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 3.7104835510253906,
+ "step": 2374
+ },
+ {
+ "epoch": 32.9912739965096,
+ "grad_norm": 0.05809671804308891,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 3.7175397872924805,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.0423070453107357,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 2.3232760429382324,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 0.6105995178222656,
+ "eval_runtime": 61.6632,
+ "eval_samples_per_second": 39.602,
+ "eval_steps_per_second": 0.632,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01396160558464,
+ "grad_norm": 0.05560882017016411,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 3.6668033599853516,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02792321116929,
+ "grad_norm": 0.053790636360645294,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 3.6379659175872803,
+ "step": 2378
+ },
+ {
+ "epoch": 33.04188481675393,
+ "grad_norm": 0.056180961430072784,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 3.6418814659118652,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05584642233857,
+ "grad_norm": 0.054904621094465256,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 3.651149034500122,
+ "step": 2380
+ },
+ {
+ "epoch": 33.06980802792321,
+ "grad_norm": 0.05298716947436333,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 3.686598777770996,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083769633507856,
+ "grad_norm": 0.053739335387945175,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 3.6473655700683594,
+ "step": 2382
+ },
+ {
+ "epoch": 33.0977312390925,
+ "grad_norm": 0.05417626351118088,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 3.6908068656921387,
+ "step": 2383
+ },
+ {
+ "epoch": 33.11169284467714,
+ "grad_norm": 0.0533599928021431,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 3.6613330841064453,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12565445026178,
+ "grad_norm": 0.05309661850333214,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 3.6534783840179443,
+ "step": 2385
+ },
+ {
+ "epoch": 33.139616055846425,
+ "grad_norm": 0.05432313680648804,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 3.664217948913574,
+ "step": 2386
+ },
+ {
+ "epoch": 33.153577661431065,
+ "grad_norm": 0.054764073342084885,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 3.660922050476074,
+ "step": 2387
+ },
+ {
+ "epoch": 33.167539267015705,
+ "grad_norm": 0.05396303907036781,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 3.662687063217163,
+ "step": 2388
+ },
+ {
+ "epoch": 33.181500872600346,
+ "grad_norm": 0.05540551245212555,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 3.6671347618103027,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19546247818499,
+ "grad_norm": 0.05510672926902771,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 3.6410186290740967,
+ "step": 2390
+ },
+ {
+ "epoch": 33.20942408376963,
+ "grad_norm": 0.055180903524160385,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 3.653822422027588,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223385689354274,
+ "grad_norm": 0.054320868104696274,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 3.6835601329803467,
+ "step": 2392
+ },
+ {
+ "epoch": 33.23734729493892,
+ "grad_norm": 0.05568838492035866,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 3.671053171157837,
+ "step": 2393
+ },
+ {
+ "epoch": 33.25130890052356,
+ "grad_norm": 0.056060485541820526,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 3.682260751724243,
+ "step": 2394
+ },
+ {
+ "epoch": 33.2652705061082,
+ "grad_norm": 0.054811324924230576,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 3.6742751598358154,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27923211169284,
+ "grad_norm": 0.05570835620164871,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 3.6490235328674316,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29319371727749,
+ "grad_norm": 0.05547381564974785,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 3.6795434951782227,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30715532286213,
+ "grad_norm": 0.05463161692023277,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 3.655003547668457,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32111692844677,
+ "grad_norm": 0.05583210662007332,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 3.680175304412842,
+ "step": 2399
+ },
+ {
+ "epoch": 33.33507853403141,
+ "grad_norm": 0.056169651448726654,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 3.667175054550171,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34904013961606,
+ "grad_norm": 0.055788591504096985,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 3.6673946380615234,
+ "step": 2401
+ },
+ {
+ "epoch": 33.3630017452007,
+ "grad_norm": 0.05556197836995125,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 3.650700569152832,
+ "step": 2402
+ },
+ {
+ "epoch": 33.37696335078534,
+ "grad_norm": 0.05749715119600296,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 3.6977670192718506,
+ "step": 2403
+ },
+ {
+ "epoch": 33.390924956369986,
+ "grad_norm": 0.05672239512205124,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 3.6604905128479004,
+ "step": 2404
+ },
+ {
+ "epoch": 33.404886561954626,
+ "grad_norm": 0.05661430582404137,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 3.64540958404541,
+ "step": 2405
+ },
+ {
+ "epoch": 33.41884816753927,
+ "grad_norm": 0.05520932003855705,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 3.6648027896881104,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43280977312391,
+ "grad_norm": 0.0557708777487278,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 3.6694397926330566,
+ "step": 2407
+ },
+ {
+ "epoch": 33.446771378708554,
+ "grad_norm": 0.05510152503848076,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 3.6723294258117676,
+ "step": 2408
+ },
+ {
+ "epoch": 33.460732984293195,
+ "grad_norm": 0.05714292824268341,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 3.6610422134399414,
+ "step": 2409
+ },
+ {
+ "epoch": 33.474694589877835,
+ "grad_norm": 0.05525747686624527,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 3.655020236968994,
+ "step": 2410
+ },
+ {
+ "epoch": 33.488656195462475,
+ "grad_norm": 0.05610140040516853,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 3.6707541942596436,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50261780104712,
+ "grad_norm": 0.05564311146736145,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 3.699244976043701,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51657940663176,
+ "grad_norm": 0.05607445910573006,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 3.6682703495025635,
+ "step": 2413
+ },
+ {
+ "epoch": 33.5305410122164,
+ "grad_norm": 0.05434789881110191,
+ "learning_rate": 0.000159218843594243,
+ "loss": 3.6799025535583496,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544502617801044,
+ "grad_norm": 0.055795323103666306,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 3.6433563232421875,
+ "step": 2415
+ },
+ {
+ "epoch": 33.55846422338569,
+ "grad_norm": 0.05676727741956711,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 3.665454864501953,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57242582897033,
+ "grad_norm": 0.05657052621245384,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 3.6618518829345703,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58638743455497,
+ "grad_norm": 0.056275662034749985,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 3.6946463584899902,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60034904013962,
+ "grad_norm": 0.056533295661211014,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 3.6695780754089355,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61431064572426,
+ "grad_norm": 0.05692996829748154,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 3.6850693225860596,
+ "step": 2420
+ },
+ {
+ "epoch": 33.6282722513089,
+ "grad_norm": 0.0577891543507576,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 3.668393611907959,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64223385689354,
+ "grad_norm": 0.05549122393131256,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 3.6772003173828125,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65619546247819,
+ "grad_norm": 0.05636308714747429,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 3.6818785667419434,
+ "step": 2423
+ },
+ {
+ "epoch": 33.67015706806283,
+ "grad_norm": 0.05681503564119339,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 3.6955409049987793,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68411867364747,
+ "grad_norm": 0.05746293440461159,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 3.6917951107025146,
+ "step": 2425
+ },
+ {
+ "epoch": 33.69808027923211,
+ "grad_norm": 0.056621894240379333,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 3.635894775390625,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712041884816756,
+ "grad_norm": 0.05611368268728256,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 3.679736375808716,
+ "step": 2427
+ },
+ {
+ "epoch": 33.726003490401396,
+ "grad_norm": 0.057769183069467545,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 3.6751227378845215,
+ "step": 2428
+ },
+ {
+ "epoch": 33.73996509598604,
+ "grad_norm": 0.05695759132504463,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 3.685178756713867,
+ "step": 2429
+ },
+ {
+ "epoch": 33.753926701570684,
+ "grad_norm": 0.05776882544159889,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 3.6355032920837402,
+ "step": 2430
+ },
+ {
+ "epoch": 33.767888307155324,
+ "grad_norm": 0.056043099611997604,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 3.668668270111084,
+ "step": 2431
+ },
+ {
+ "epoch": 33.781849912739965,
+ "grad_norm": 0.05634639039635658,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 3.6788253784179688,
+ "step": 2432
+ },
+ {
+ "epoch": 33.795811518324605,
+ "grad_norm": 0.05703914538025856,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 3.676163911819458,
+ "step": 2433
+ },
+ {
+ "epoch": 33.80977312390925,
+ "grad_norm": 0.05846793204545975,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 3.657705307006836,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82373472949389,
+ "grad_norm": 0.059191398322582245,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 3.70918607711792,
+ "step": 2435
+ },
+ {
+ "epoch": 33.83769633507853,
+ "grad_norm": 0.056207913905382156,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 3.690638542175293,
+ "step": 2436
+ },
+ {
+ "epoch": 33.85165794066317,
+ "grad_norm": 0.0590900182723999,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 3.6839122772216797,
+ "step": 2437
+ },
+ {
+ "epoch": 33.86561954624782,
+ "grad_norm": 0.05649874731898308,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 3.6689414978027344,
+ "step": 2438
+ },
+ {
+ "epoch": 33.87958115183246,
+ "grad_norm": 0.05947998911142349,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 3.653146743774414,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8935427574171,
+ "grad_norm": 0.05769696831703186,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 3.685544967651367,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90750436300174,
+ "grad_norm": 0.05837882682681084,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 3.6648831367492676,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92146596858639,
+ "grad_norm": 0.05676107108592987,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 3.7186074256896973,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93542757417103,
+ "grad_norm": 0.05771078169345856,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 3.678924083709717,
+ "step": 2443
+ },
+ {
+ "epoch": 33.94938917975567,
+ "grad_norm": 0.05686977133154869,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 3.7026078701019287,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96335078534032,
+ "grad_norm": 0.0571015290915966,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 3.6880578994750977,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97731239092496,
+ "grad_norm": 0.056974101811647415,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 3.6521928310394287,
+ "step": 2446
+ },
+ {
+ "epoch": 33.9912739965096,
+ "grad_norm": 0.05962125584483147,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 3.686126708984375,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.043099287897348404,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 2.308753490447998,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 0.6120397448539734,
+ "eval_runtime": 60.5299,
+ "eval_samples_per_second": 40.344,
+ "eval_steps_per_second": 0.644,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01396160558464,
+ "grad_norm": 0.05469869077205658,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 3.6253180503845215,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02792321116929,
+ "grad_norm": 0.05455945432186127,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 3.651854991912842,
+ "step": 2450
+ },
+ {
+ "epoch": 34.04188481675393,
+ "grad_norm": 0.05369517207145691,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 3.630152702331543,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05584642233857,
+ "grad_norm": 0.05498119816184044,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 3.6293938159942627,
+ "step": 2452
+ },
+ {
+ "epoch": 34.06980802792321,
+ "grad_norm": 0.05530959740281105,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 3.6548690795898438,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083769633507856,
+ "grad_norm": 0.055240705609321594,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 3.6408982276916504,
+ "step": 2454
+ },
+ {
+ "epoch": 34.0977312390925,
+ "grad_norm": 0.05358482152223587,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 3.6274750232696533,
+ "step": 2455
+ },
+ {
+ "epoch": 34.11169284467714,
+ "grad_norm": 0.05580627918243408,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 3.6357412338256836,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12565445026178,
+ "grad_norm": 0.05458527058362961,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 3.6637980937957764,
+ "step": 2457
+ },
+ {
+ "epoch": 34.139616055846425,
+ "grad_norm": 0.05548330023884773,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 3.627671718597412,
+ "step": 2458
+ },
+ {
+ "epoch": 34.153577661431065,
+ "grad_norm": 0.05482204258441925,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 3.642627239227295,
+ "step": 2459
+ },
+ {
+ "epoch": 34.167539267015705,
+ "grad_norm": 0.05435148626565933,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 3.6398377418518066,
+ "step": 2460
+ },
+ {
+ "epoch": 34.181500872600346,
+ "grad_norm": 0.05556599423289299,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 3.62994384765625,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19546247818499,
+ "grad_norm": 0.05394179746508598,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 3.593183994293213,
+ "step": 2462
+ },
+ {
+ "epoch": 34.20942408376963,
+ "grad_norm": 0.05472215637564659,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 3.645197868347168,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223385689354274,
+ "grad_norm": 0.05505116656422615,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 3.655001163482666,
+ "step": 2464
+ },
+ {
+ "epoch": 34.23734729493892,
+ "grad_norm": 0.05473910644650459,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 3.636812210083008,
+ "step": 2465
+ },
+ {
+ "epoch": 34.25130890052356,
+ "grad_norm": 0.05537406727671623,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 3.620631217956543,
+ "step": 2466
+ },
+ {
+ "epoch": 34.2652705061082,
+ "grad_norm": 0.054202884435653687,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 3.6500911712646484,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27923211169284,
+ "grad_norm": 0.0536428838968277,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 3.6268038749694824,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29319371727749,
+ "grad_norm": 0.053046368062496185,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 3.646955966949463,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30715532286213,
+ "grad_norm": 0.05279967561364174,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 3.6142497062683105,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32111692844677,
+ "grad_norm": 0.05348867550492287,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 3.643517017364502,
+ "step": 2471
+ },
+ {
+ "epoch": 34.33507853403141,
+ "grad_norm": 0.05406436324119568,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 3.6528172492980957,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34904013961606,
+ "grad_norm": 0.05249491706490517,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 3.6664018630981445,
+ "step": 2473
+ },
+ {
+ "epoch": 34.3630017452007,
+ "grad_norm": 0.053803686052560806,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 3.609346389770508,
+ "step": 2474
+ },
+ {
+ "epoch": 34.37696335078534,
+ "grad_norm": 0.054445188492536545,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 3.6647889614105225,
+ "step": 2475
+ },
+ {
+ "epoch": 34.390924956369986,
+ "grad_norm": 0.0550406239926815,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 3.656118154525757,
+ "step": 2476
+ },
+ {
+ "epoch": 34.404886561954626,
+ "grad_norm": 0.054776549339294434,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 3.6493523120880127,
+ "step": 2477
+ },
+ {
+ "epoch": 34.41884816753927,
+ "grad_norm": 0.054829373955726624,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 3.634632110595703,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43280977312391,
+ "grad_norm": 0.054575446993112564,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 3.636216878890991,
+ "step": 2479
+ },
+ {
+ "epoch": 34.446771378708554,
+ "grad_norm": 0.056146737188100815,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 3.6584529876708984,
+ "step": 2480
+ },
+ {
+ "epoch": 34.460732984293195,
+ "grad_norm": 0.054963648319244385,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 3.6363425254821777,
+ "step": 2481
+ },
+ {
+ "epoch": 34.474694589877835,
+ "grad_norm": 0.05659656599164009,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 3.6401073932647705,
+ "step": 2482
+ },
+ {
+ "epoch": 34.488656195462475,
+ "grad_norm": 0.05448099970817566,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 3.6161837577819824,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50261780104712,
+ "grad_norm": 0.057917334139347076,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 3.6802477836608887,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51657940663176,
+ "grad_norm": 0.056171346455812454,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 3.652339220046997,
+ "step": 2485
+ },
+ {
+ "epoch": 34.5305410122164,
+ "grad_norm": 0.05800231546163559,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 3.6437389850616455,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544502617801044,
+ "grad_norm": 0.05693773180246353,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 3.6268343925476074,
+ "step": 2487
+ },
+ {
+ "epoch": 34.55846422338569,
+ "grad_norm": 0.05424632877111435,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 3.6246471405029297,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57242582897033,
+ "grad_norm": 0.05645301938056946,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 3.649033308029175,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58638743455497,
+ "grad_norm": 0.056913044303655624,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 3.658982515335083,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60034904013962,
+ "grad_norm": 0.05770258232951164,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 3.621516704559326,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61431064572426,
+ "grad_norm": 0.05483509227633476,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 3.659385919570923,
+ "step": 2492
+ },
+ {
+ "epoch": 34.6282722513089,
+ "grad_norm": 0.057591512799263,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 3.6512298583984375,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64223385689354,
+ "grad_norm": 0.055053744465112686,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 3.630310297012329,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65619546247819,
+ "grad_norm": 0.056070517748594284,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 3.6766881942749023,
+ "step": 2495
+ },
+ {
+ "epoch": 34.67015706806283,
+ "grad_norm": 0.056556008756160736,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 3.6454544067382812,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68411867364747,
+ "grad_norm": 0.05624367296695709,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 3.6482372283935547,
+ "step": 2497
+ },
+ {
+ "epoch": 34.69808027923211,
+ "grad_norm": 0.056266799569129944,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 3.6528916358947754,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712041884816756,
+ "grad_norm": 0.05540672317147255,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 3.636371374130249,
+ "step": 2499
+ },
+ {
+ "epoch": 34.726003490401396,
+ "grad_norm": 0.05527781695127487,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 3.6440672874450684,
+ "step": 2500
+ },
+ {
+ "epoch": 34.73996509598604,
+ "grad_norm": 0.05645914003252983,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 3.6747167110443115,
+ "step": 2501
+ },
+ {
+ "epoch": 34.753926701570684,
+ "grad_norm": 0.055584684014320374,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 3.637603282928467,
+ "step": 2502
+ },
+ {
+ "epoch": 34.767888307155324,
+ "grad_norm": 0.0554841123521328,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 3.6546664237976074,
+ "step": 2503
+ },
+ {
+ "epoch": 34.781849912739965,
+ "grad_norm": 0.0557134784758091,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 3.6366615295410156,
+ "step": 2504
+ },
+ {
+ "epoch": 34.795811518324605,
+ "grad_norm": 0.056569263339042664,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 3.657125949859619,
+ "step": 2505
+ },
+ {
+ "epoch": 34.80977312390925,
+ "grad_norm": 0.05737468600273132,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 3.6409783363342285,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82373472949389,
+ "grad_norm": 0.05602499470114708,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 3.6670620441436768,
+ "step": 2507
+ },
+ {
+ "epoch": 34.83769633507853,
+ "grad_norm": 0.059535399079322815,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 3.6585612297058105,
+ "step": 2508
+ },
+ {
+ "epoch": 34.85165794066317,
+ "grad_norm": 0.05610577017068863,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 3.6544179916381836,
+ "step": 2509
+ },
+ {
+ "epoch": 34.86561954624782,
+ "grad_norm": 0.057512786239385605,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 3.658529281616211,
+ "step": 2510
+ },
+ {
+ "epoch": 34.87958115183246,
+ "grad_norm": 0.05605233088135719,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 3.622344493865967,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8935427574171,
+ "grad_norm": 0.055069468915462494,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 3.6812562942504883,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90750436300174,
+ "grad_norm": 0.05756533145904541,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 3.651491165161133,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92146596858639,
+ "grad_norm": 0.05596763640642166,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 3.633408784866333,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93542757417103,
+ "grad_norm": 0.057164326310157776,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 3.632655620574951,
+ "step": 2515
+ },
+ {
+ "epoch": 34.94938917975567,
+ "grad_norm": 0.059363316744565964,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 3.637542486190796,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96335078534032,
+ "grad_norm": 0.05572650581598282,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 3.666663885116577,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97731239092496,
+ "grad_norm": 0.05825378745794296,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 3.654174327850342,
+ "step": 2518
+ },
+ {
+ "epoch": 34.9912739965096,
+ "grad_norm": 0.058226607739925385,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 3.630509376525879,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.041674062609672546,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 2.2687463760375977,
+ "step": 2520
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.0753029862273843e+18,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2520/training_args.bin b/runs/l2r90-wd0033-ssl/checkpoint-2520/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..313c41733ff55284c06034a59bb885b30e40ffc8
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2520/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7ddd0bdddba68d3e19ab33477000f3fc684fe4ee8e3580eeb8c02646e5fc27ac
+size 4856
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2880/chat_template.jinja b/runs/l2r90-wd0033-ssl/checkpoint-2880/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2880/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2880/config.json b/runs/l2r90-wd0033-ssl/checkpoint-2880/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2880/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2880/generation_config.json b/runs/l2r90-wd0033-ssl/checkpoint-2880/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2880/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2880/model.safetensors b/runs/l2r90-wd0033-ssl/checkpoint-2880/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..1b7dffc137fbe487a5ec32ec141b36b74f7dda61
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2880/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:fde69e47285d2f9aad6c4a6ffd38113453a42e2a20cbec16fb1a51938bad1fdb
+size 583356232
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2880/optimizer.pt b/runs/l2r90-wd0033-ssl/checkpoint-2880/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..2638f5aacc79eb6873ca97de52d7f6a53895d567
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2880/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:556efaacdd8562c34a828aa3399d9c90e348f2ada0f6682db00be36a762667b3
+size 1166827898
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2880/rng_state_0.pth b/runs/l2r90-wd0033-ssl/checkpoint-2880/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..d0a0edf9dd5d265c92220d4e979a77083deb97af
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2880/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:85c8b26cb86c92e574450b60beab12a61e8d26df90092a5c2f1efeb69e097455
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2880/rng_state_1.pth b/runs/l2r90-wd0033-ssl/checkpoint-2880/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..4fa7c978646ed846fd10feda28e54269b8f8067a
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2880/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1ac830d44175ec31da3f7539f6c1a5e88e6ae26472bd7fcea1121a59a58457da
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2880/scheduler.pt b/runs/l2r90-wd0033-ssl/checkpoint-2880/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..5c7f6b411db3da6cc58af5b4530e84dddd5abe00
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2880/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2e88932b62643ef7a711b903ac04665eab5638dccef14b4bec61d20b2791e67b
+size 1064
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2880/tokenizer.json b/runs/l2r90-wd0033-ssl/checkpoint-2880/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2880/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2880/tokenizer_config.json b/runs/l2r90-wd0033-ssl/checkpoint-2880/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2880/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2880/trainer_state.json b/runs/l2r90-wd0033-ssl/checkpoint-2880/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..150ccb4acb422b3e74ee5a3ad5206aa9c470f000
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2880/trainer_state.json
@@ -0,0 +1,20498 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 40.0,
+ "eval_steps": 500,
+ "global_step": 2880,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013961605584642234,
+ "grad_norm": 0.26733073592185974,
+ "learning_rate": 0.0,
+ "loss": 7.990724086761475,
+ "step": 1
+ },
+ {
+ "epoch": 0.027923211169284468,
+ "grad_norm": 0.26712340116500854,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 7.990110874176025,
+ "step": 2
+ },
+ {
+ "epoch": 0.041884816753926704,
+ "grad_norm": 0.2646932601928711,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 7.956935405731201,
+ "step": 3
+ },
+ {
+ "epoch": 0.055846422338568937,
+ "grad_norm": 0.24684545397758484,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 7.890539646148682,
+ "step": 4
+ },
+ {
+ "epoch": 0.06980802792321117,
+ "grad_norm": 0.21016643941402435,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 7.8146138191223145,
+ "step": 5
+ },
+ {
+ "epoch": 0.08376963350785341,
+ "grad_norm": 0.17992742359638214,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 7.724307060241699,
+ "step": 6
+ },
+ {
+ "epoch": 0.09773123909249563,
+ "grad_norm": 0.14356893301010132,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 7.645662307739258,
+ "step": 7
+ },
+ {
+ "epoch": 0.11169284467713787,
+ "grad_norm": 0.10830964893102646,
+ "learning_rate": 4.2e-05,
+ "loss": 7.582019329071045,
+ "step": 8
+ },
+ {
+ "epoch": 0.1256544502617801,
+ "grad_norm": 0.08265161514282227,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 7.5230793952941895,
+ "step": 9
+ },
+ {
+ "epoch": 0.13961605584642234,
+ "grad_norm": 0.06423710286617279,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 7.473812103271484,
+ "step": 10
+ },
+ {
+ "epoch": 0.15357766143106458,
+ "grad_norm": 0.05132246017456055,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 7.440786838531494,
+ "step": 11
+ },
+ {
+ "epoch": 0.16753926701570682,
+ "grad_norm": 0.040572404861450195,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 7.41290283203125,
+ "step": 12
+ },
+ {
+ "epoch": 0.18150087260034903,
+ "grad_norm": 0.03138705715537071,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 7.39749002456665,
+ "step": 13
+ },
+ {
+ "epoch": 0.19546247818499127,
+ "grad_norm": 0.02553008496761322,
+ "learning_rate": 7.8e-05,
+ "loss": 7.387092590332031,
+ "step": 14
+ },
+ {
+ "epoch": 0.2094240837696335,
+ "grad_norm": 0.022178977727890015,
+ "learning_rate": 8.4e-05,
+ "loss": 7.378993511199951,
+ "step": 15
+ },
+ {
+ "epoch": 0.22338568935427575,
+ "grad_norm": 0.019915293902158737,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 7.37220573425293,
+ "step": 16
+ },
+ {
+ "epoch": 0.23734729493891799,
+ "grad_norm": 0.017390629276633263,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 7.366759300231934,
+ "step": 17
+ },
+ {
+ "epoch": 0.2513089005235602,
+ "grad_norm": 0.015472771599888802,
+ "learning_rate": 0.000102,
+ "loss": 7.3652663230896,
+ "step": 18
+ },
+ {
+ "epoch": 0.26527050610820246,
+ "grad_norm": 0.013599214144051075,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 7.359714984893799,
+ "step": 19
+ },
+ {
+ "epoch": 0.2792321116928447,
+ "grad_norm": 0.012442861683666706,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 7.353302001953125,
+ "step": 20
+ },
+ {
+ "epoch": 0.2931937172774869,
+ "grad_norm": 0.010471965186297894,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 7.357417583465576,
+ "step": 21
+ },
+ {
+ "epoch": 0.30715532286212915,
+ "grad_norm": 0.009647224098443985,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 7.351763725280762,
+ "step": 22
+ },
+ {
+ "epoch": 0.32111692844677137,
+ "grad_norm": 0.008228210732340813,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 7.345434188842773,
+ "step": 23
+ },
+ {
+ "epoch": 0.33507853403141363,
+ "grad_norm": 0.007454514969140291,
+ "learning_rate": 0.000138,
+ "loss": 7.342520236968994,
+ "step": 24
+ },
+ {
+ "epoch": 0.34904013961605584,
+ "grad_norm": 0.0069856527261435986,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 7.338870048522949,
+ "step": 25
+ },
+ {
+ "epoch": 0.36300174520069806,
+ "grad_norm": 0.006091665010899305,
+ "learning_rate": 0.00015,
+ "loss": 7.328121662139893,
+ "step": 26
+ },
+ {
+ "epoch": 0.3769633507853403,
+ "grad_norm": 0.005525320768356323,
+ "learning_rate": 0.000156,
+ "loss": 7.335029602050781,
+ "step": 27
+ },
+ {
+ "epoch": 0.39092495636998253,
+ "grad_norm": 0.0061133443377912045,
+ "learning_rate": 0.000162,
+ "loss": 7.319417953491211,
+ "step": 28
+ },
+ {
+ "epoch": 0.4048865619546248,
+ "grad_norm": 0.006040909793227911,
+ "learning_rate": 0.000168,
+ "loss": 7.326376438140869,
+ "step": 29
+ },
+ {
+ "epoch": 0.418848167539267,
+ "grad_norm": 0.006003216840326786,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 7.319552421569824,
+ "step": 30
+ },
+ {
+ "epoch": 0.4328097731239092,
+ "grad_norm": 0.007096003741025925,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 7.323486328125,
+ "step": 31
+ },
+ {
+ "epoch": 0.4467713787085515,
+ "grad_norm": 0.007725459989160299,
+ "learning_rate": 0.000186,
+ "loss": 7.3190717697143555,
+ "step": 32
+ },
+ {
+ "epoch": 0.4607329842931937,
+ "grad_norm": 0.007338670082390308,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 7.317154884338379,
+ "step": 33
+ },
+ {
+ "epoch": 0.47469458987783597,
+ "grad_norm": 0.007709544617682695,
+ "learning_rate": 0.000198,
+ "loss": 7.311826705932617,
+ "step": 34
+ },
+ {
+ "epoch": 0.4886561954624782,
+ "grad_norm": 0.009249784983694553,
+ "learning_rate": 0.000204,
+ "loss": 7.322498798370361,
+ "step": 35
+ },
+ {
+ "epoch": 0.5026178010471204,
+ "grad_norm": 0.006950767710804939,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 7.314571380615234,
+ "step": 36
+ },
+ {
+ "epoch": 0.5165794066317626,
+ "grad_norm": 0.007583985570818186,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 7.309540748596191,
+ "step": 37
+ },
+ {
+ "epoch": 0.5305410122164049,
+ "grad_norm": 0.014731820672750473,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 7.307004928588867,
+ "step": 38
+ },
+ {
+ "epoch": 0.5445026178010471,
+ "grad_norm": 0.016276473179459572,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 7.299642562866211,
+ "step": 39
+ },
+ {
+ "epoch": 0.5584642233856894,
+ "grad_norm": 0.020018689334392548,
+ "learning_rate": 0.000234,
+ "loss": 7.298466682434082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5724258289703316,
+ "grad_norm": 0.014941536821424961,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 7.289968967437744,
+ "step": 41
+ },
+ {
+ "epoch": 0.5863874345549738,
+ "grad_norm": 0.018299812451004982,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 7.276564598083496,
+ "step": 42
+ },
+ {
+ "epoch": 0.6003490401396161,
+ "grad_norm": 0.01615075021982193,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.2575554847717285,
+ "step": 43
+ },
+ {
+ "epoch": 0.6143106457242583,
+ "grad_norm": 0.013063831254839897,
+ "learning_rate": 0.000258,
+ "loss": 7.253885269165039,
+ "step": 44
+ },
+ {
+ "epoch": 0.6282722513089005,
+ "grad_norm": 0.015211676247417927,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.245779037475586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6422338568935427,
+ "grad_norm": 0.02093500830233097,
+ "learning_rate": 0.00027,
+ "loss": 7.225130081176758,
+ "step": 46
+ },
+ {
+ "epoch": 0.6561954624781849,
+ "grad_norm": 0.01808053068816662,
+ "learning_rate": 0.000276,
+ "loss": 7.209654331207275,
+ "step": 47
+ },
+ {
+ "epoch": 0.6701570680628273,
+ "grad_norm": 0.02209644205868244,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.197124481201172,
+ "step": 48
+ },
+ {
+ "epoch": 0.6841186736474695,
+ "grad_norm": 0.023524878546595573,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.1841936111450195,
+ "step": 49
+ },
+ {
+ "epoch": 0.6980802792321117,
+ "grad_norm": 0.031148087233304977,
+ "learning_rate": 0.000294,
+ "loss": 7.188040733337402,
+ "step": 50
+ },
+ {
+ "epoch": 0.7120418848167539,
+ "grad_norm": 0.02931929938495159,
+ "learning_rate": 0.0003,
+ "loss": 7.17371940612793,
+ "step": 51
+ },
+ {
+ "epoch": 0.7260034904013961,
+ "grad_norm": 0.01611483097076416,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.1714324951171875,
+ "step": 52
+ },
+ {
+ "epoch": 0.7399650959860384,
+ "grad_norm": 0.02619907818734646,
+ "learning_rate": 0.000312,
+ "loss": 7.158261299133301,
+ "step": 53
+ },
+ {
+ "epoch": 0.7539267015706806,
+ "grad_norm": 0.049462560564279556,
+ "learning_rate": 0.000318,
+ "loss": 7.167076110839844,
+ "step": 54
+ },
+ {
+ "epoch": 0.7678883071553229,
+ "grad_norm": 0.09835070371627808,
+ "learning_rate": 0.000324,
+ "loss": 7.154292106628418,
+ "step": 55
+ },
+ {
+ "epoch": 0.7818499127399651,
+ "grad_norm": 0.11406510323286057,
+ "learning_rate": 0.00033,
+ "loss": 7.162016868591309,
+ "step": 56
+ },
+ {
+ "epoch": 0.7958115183246073,
+ "grad_norm": 0.027026303112506866,
+ "learning_rate": 0.000336,
+ "loss": 7.141416549682617,
+ "step": 57
+ },
+ {
+ "epoch": 0.8097731239092496,
+ "grad_norm": 0.06690579652786255,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.13959264755249,
+ "step": 58
+ },
+ {
+ "epoch": 0.8237347294938918,
+ "grad_norm": 0.020919431000947952,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.126513481140137,
+ "step": 59
+ },
+ {
+ "epoch": 0.837696335078534,
+ "grad_norm": 0.05618130415678024,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.124345779418945,
+ "step": 60
+ },
+ {
+ "epoch": 0.8516579406631762,
+ "grad_norm": 0.03402302414178848,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.100615501403809,
+ "step": 61
+ },
+ {
+ "epoch": 0.8656195462478184,
+ "grad_norm": 0.03646906837821007,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.116507530212402,
+ "step": 62
+ },
+ {
+ "epoch": 0.8795811518324608,
+ "grad_norm": 0.03076397068798542,
+ "learning_rate": 0.000372,
+ "loss": 7.097659111022949,
+ "step": 63
+ },
+ {
+ "epoch": 0.893542757417103,
+ "grad_norm": 0.03329646959900856,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.097290992736816,
+ "step": 64
+ },
+ {
+ "epoch": 0.9075043630017452,
+ "grad_norm": 0.01998630352318287,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.092268943786621,
+ "step": 65
+ },
+ {
+ "epoch": 0.9214659685863874,
+ "grad_norm": 0.03278940171003342,
+ "learning_rate": 0.00039,
+ "loss": 7.088292121887207,
+ "step": 66
+ },
+ {
+ "epoch": 0.9354275741710296,
+ "grad_norm": 0.024620421230793,
+ "learning_rate": 0.000396,
+ "loss": 7.078650951385498,
+ "step": 67
+ },
+ {
+ "epoch": 0.9493891797556719,
+ "grad_norm": 0.03750383481383324,
+ "learning_rate": 0.000402,
+ "loss": 7.052778244018555,
+ "step": 68
+ },
+ {
+ "epoch": 0.9633507853403142,
+ "grad_norm": 0.07558075338602066,
+ "learning_rate": 0.000408,
+ "loss": 7.045991897583008,
+ "step": 69
+ },
+ {
+ "epoch": 0.9773123909249564,
+ "grad_norm": 0.16021482646465302,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.023152828216553,
+ "step": 70
+ },
+ {
+ "epoch": 0.9912739965095986,
+ "grad_norm": 0.24981893599033356,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.067540645599365,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.10255520790815353,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 4.425640106201172,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8861758708953857,
+ "eval_runtime": 58.2699,
+ "eval_samples_per_second": 41.908,
+ "eval_steps_per_second": 0.669,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139616055846423,
+ "grad_norm": 0.13374797999858856,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.073373317718506,
+ "step": 73
+ },
+ {
+ "epoch": 1.0279232111692844,
+ "grad_norm": 0.05079387128353119,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.0243306159973145,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418848167539267,
+ "grad_norm": 0.10056068748235703,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.014476299285889,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558464223385688,
+ "grad_norm": 0.06995758414268494,
+ "learning_rate": 0.00045,
+ "loss": 7.011931419372559,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698080279232112,
+ "grad_norm": 0.0483892522752285,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 6.9822797775268555,
+ "step": 77
+ },
+ {
+ "epoch": 1.0837696335078535,
+ "grad_norm": 0.08408170193433762,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 6.964348793029785,
+ "step": 78
+ },
+ {
+ "epoch": 1.0977312390924956,
+ "grad_norm": 0.08027710020542145,
+ "learning_rate": 0.000468,
+ "loss": 6.96852445602417,
+ "step": 79
+ },
+ {
+ "epoch": 1.111692844677138,
+ "grad_norm": 0.05647159367799759,
+ "learning_rate": 0.000474,
+ "loss": 6.944484710693359,
+ "step": 80
+ },
+ {
+ "epoch": 1.12565445026178,
+ "grad_norm": 0.07464330643415451,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 6.9246745109558105,
+ "step": 81
+ },
+ {
+ "epoch": 1.1396160558464223,
+ "grad_norm": 0.04451674595475197,
+ "learning_rate": 0.000486,
+ "loss": 6.895403861999512,
+ "step": 82
+ },
+ {
+ "epoch": 1.1535776614310647,
+ "grad_norm": 0.05820256471633911,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 6.887338638305664,
+ "step": 83
+ },
+ {
+ "epoch": 1.1675392670157068,
+ "grad_norm": 0.02925296127796173,
+ "learning_rate": 0.000498,
+ "loss": 6.857797622680664,
+ "step": 84
+ },
+ {
+ "epoch": 1.181500872600349,
+ "grad_norm": 0.05096987634897232,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 6.854822158813477,
+ "step": 85
+ },
+ {
+ "epoch": 1.1954624781849912,
+ "grad_norm": 0.03483197093009949,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 6.840914726257324,
+ "step": 86
+ },
+ {
+ "epoch": 1.2094240837696335,
+ "grad_norm": 0.03323264792561531,
+ "learning_rate": 0.000516,
+ "loss": 6.834497451782227,
+ "step": 87
+ },
+ {
+ "epoch": 1.2233856893542758,
+ "grad_norm": 0.03290868178009987,
+ "learning_rate": 0.000522,
+ "loss": 6.820746898651123,
+ "step": 88
+ },
+ {
+ "epoch": 1.237347294938918,
+ "grad_norm": 0.025398844853043556,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 6.795746803283691,
+ "step": 89
+ },
+ {
+ "epoch": 1.2513089005235603,
+ "grad_norm": 0.030700596049427986,
+ "learning_rate": 0.000534,
+ "loss": 6.7867889404296875,
+ "step": 90
+ },
+ {
+ "epoch": 1.2652705061082026,
+ "grad_norm": 0.019589420408010483,
+ "learning_rate": 0.00054,
+ "loss": 6.77443790435791,
+ "step": 91
+ },
+ {
+ "epoch": 1.2792321116928447,
+ "grad_norm": 0.03313920646905899,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 6.762271881103516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2931937172774868,
+ "grad_norm": 0.025389468297362328,
+ "learning_rate": 0.000552,
+ "loss": 6.750258445739746,
+ "step": 93
+ },
+ {
+ "epoch": 1.307155322862129,
+ "grad_norm": 0.040296655148267746,
+ "learning_rate": 0.000558,
+ "loss": 6.729808807373047,
+ "step": 94
+ },
+ {
+ "epoch": 1.3211169284467714,
+ "grad_norm": 0.04735933244228363,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 6.737982749938965,
+ "step": 95
+ },
+ {
+ "epoch": 1.3350785340314135,
+ "grad_norm": 0.05529680848121643,
+ "learning_rate": 0.00057,
+ "loss": 6.735714912414551,
+ "step": 96
+ },
+ {
+ "epoch": 1.3490401396160558,
+ "grad_norm": 0.07148554176092148,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 6.708859920501709,
+ "step": 97
+ },
+ {
+ "epoch": 1.3630017452006982,
+ "grad_norm": 0.059708479791879654,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 6.708860397338867,
+ "step": 98
+ },
+ {
+ "epoch": 1.3769633507853403,
+ "grad_norm": 0.027111195027828217,
+ "learning_rate": 0.000588,
+ "loss": 6.686685085296631,
+ "step": 99
+ },
+ {
+ "epoch": 1.3909249563699826,
+ "grad_norm": 0.0500967763364315,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 6.676325798034668,
+ "step": 100
+ },
+ {
+ "epoch": 1.404886561954625,
+ "grad_norm": 0.044565796852111816,
+ "learning_rate": 0.0006,
+ "loss": 6.674530029296875,
+ "step": 101
+ },
+ {
+ "epoch": 1.418848167539267,
+ "grad_norm": 0.03608180209994316,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 6.653741359710693,
+ "step": 102
+ },
+ {
+ "epoch": 1.4328097731239091,
+ "grad_norm": 0.030394982546567917,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 6.643152236938477,
+ "step": 103
+ },
+ {
+ "epoch": 1.4467713787085514,
+ "grad_norm": 0.040185656398534775,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 6.628686904907227,
+ "step": 104
+ },
+ {
+ "epoch": 1.4607329842931938,
+ "grad_norm": 0.021049322560429573,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 6.610864639282227,
+ "step": 105
+ },
+ {
+ "epoch": 1.4746945898778359,
+ "grad_norm": 0.04164277762174606,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 6.602880954742432,
+ "step": 106
+ },
+ {
+ "epoch": 1.4886561954624782,
+ "grad_norm": 0.0389864556491375,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 6.599456787109375,
+ "step": 107
+ },
+ {
+ "epoch": 1.5026178010471205,
+ "grad_norm": 0.05805973336100578,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 6.588700294494629,
+ "step": 108
+ },
+ {
+ "epoch": 1.5165794066317626,
+ "grad_norm": 0.08825036883354187,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 6.587496757507324,
+ "step": 109
+ },
+ {
+ "epoch": 1.530541012216405,
+ "grad_norm": 0.12521523237228394,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 6.626501083374023,
+ "step": 110
+ },
+ {
+ "epoch": 1.5445026178010473,
+ "grad_norm": 0.07429645210504532,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 6.590241432189941,
+ "step": 111
+ },
+ {
+ "epoch": 1.5584642233856894,
+ "grad_norm": 0.06463009119033813,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 6.578714370727539,
+ "step": 112
+ },
+ {
+ "epoch": 1.5724258289703315,
+ "grad_norm": 0.07508203387260437,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 6.561270713806152,
+ "step": 113
+ },
+ {
+ "epoch": 1.5863874345549738,
+ "grad_norm": 0.04125197231769562,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 6.558821678161621,
+ "step": 114
+ },
+ {
+ "epoch": 1.600349040139616,
+ "grad_norm": 0.05457386001944542,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.541611671447754,
+ "step": 115
+ },
+ {
+ "epoch": 1.6143106457242582,
+ "grad_norm": 0.03684385493397713,
+ "learning_rate": 0.000599973080557108,
+ "loss": 6.547738075256348,
+ "step": 116
+ },
+ {
+ "epoch": 1.6282722513089005,
+ "grad_norm": 0.05888746306300163,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.529491424560547,
+ "step": 117
+ },
+ {
+ "epoch": 1.6422338568935428,
+ "grad_norm": 0.056668370962142944,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.519444942474365,
+ "step": 118
+ },
+ {
+ "epoch": 1.656195462478185,
+ "grad_norm": 0.09123896062374115,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.522984981536865,
+ "step": 119
+ },
+ {
+ "epoch": 1.6701570680628273,
+ "grad_norm": 0.12380023300647736,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.515375137329102,
+ "step": 120
+ },
+ {
+ "epoch": 1.6841186736474696,
+ "grad_norm": 0.06953322887420654,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.500491142272949,
+ "step": 121
+ },
+ {
+ "epoch": 1.6980802792321117,
+ "grad_norm": 0.08613869547843933,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.5097198486328125,
+ "step": 122
+ },
+ {
+ "epoch": 1.7120418848167538,
+ "grad_norm": 0.08296383172273636,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.487521171569824,
+ "step": 123
+ },
+ {
+ "epoch": 1.7260034904013961,
+ "grad_norm": 0.05744783207774162,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.485134124755859,
+ "step": 124
+ },
+ {
+ "epoch": 1.7399650959860384,
+ "grad_norm": 0.04108566418290138,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.466182708740234,
+ "step": 125
+ },
+ {
+ "epoch": 1.7539267015706805,
+ "grad_norm": 0.05374931916594505,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.479011535644531,
+ "step": 126
+ },
+ {
+ "epoch": 1.7678883071553229,
+ "grad_norm": 0.03867914527654648,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.465902328491211,
+ "step": 127
+ },
+ {
+ "epoch": 1.7818499127399652,
+ "grad_norm": 0.05477374792098999,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.453082084655762,
+ "step": 128
+ },
+ {
+ "epoch": 1.7958115183246073,
+ "grad_norm": 0.06508298963308334,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.446089744567871,
+ "step": 129
+ },
+ {
+ "epoch": 1.8097731239092496,
+ "grad_norm": 0.06388624757528305,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.451748847961426,
+ "step": 130
+ },
+ {
+ "epoch": 1.823734729493892,
+ "grad_norm": 0.04223188757896423,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.4492387771606445,
+ "step": 131
+ },
+ {
+ "epoch": 1.837696335078534,
+ "grad_norm": 0.04528658092021942,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.409390926361084,
+ "step": 132
+ },
+ {
+ "epoch": 1.8516579406631761,
+ "grad_norm": 0.04962967336177826,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.415774345397949,
+ "step": 133
+ },
+ {
+ "epoch": 1.8656195462478184,
+ "grad_norm": 0.031944390386343,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.411528587341309,
+ "step": 134
+ },
+ {
+ "epoch": 1.8795811518324608,
+ "grad_norm": 0.028411149978637695,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.392329216003418,
+ "step": 135
+ },
+ {
+ "epoch": 1.8935427574171029,
+ "grad_norm": 0.03516148403286934,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.392238616943359,
+ "step": 136
+ },
+ {
+ "epoch": 1.9075043630017452,
+ "grad_norm": 0.027800120413303375,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.39179801940918,
+ "step": 137
+ },
+ {
+ "epoch": 1.9214659685863875,
+ "grad_norm": 0.02362987957894802,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.378960609436035,
+ "step": 138
+ },
+ {
+ "epoch": 1.9354275741710296,
+ "grad_norm": 0.033999308943748474,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.36802339553833,
+ "step": 139
+ },
+ {
+ "epoch": 1.949389179755672,
+ "grad_norm": 0.03290879726409912,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.366308689117432,
+ "step": 140
+ },
+ {
+ "epoch": 1.9633507853403143,
+ "grad_norm": 0.0286843404173851,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.350399971008301,
+ "step": 141
+ },
+ {
+ "epoch": 1.9773123909249564,
+ "grad_norm": 0.027145877480506897,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.33477783203125,
+ "step": 142
+ },
+ {
+ "epoch": 1.9912739965095985,
+ "grad_norm": 0.03530878946185112,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.322859764099121,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.028133150190114975,
+ "learning_rate": 0.000599778806120077,
+ "loss": 3.955288887023926,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7927101254463196,
+ "eval_runtime": 60.8372,
+ "eval_samples_per_second": 40.14,
+ "eval_steps_per_second": 0.641,
+ "step": 144
+ },
+ {
+ "epoch": 2.013961605584642,
+ "grad_norm": 0.03054557554423809,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.311975479125977,
+ "step": 145
+ },
+ {
+ "epoch": 2.0279232111692846,
+ "grad_norm": 0.05406459793448448,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.3030853271484375,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418848167539267,
+ "grad_norm": 0.10911019891500473,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.329056739807129,
+ "step": 147
+ },
+ {
+ "epoch": 2.055846422338569,
+ "grad_norm": 0.13948315382003784,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.362421035766602,
+ "step": 148
+ },
+ {
+ "epoch": 2.0698080279232114,
+ "grad_norm": 0.13176991045475006,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.3888139724731445,
+ "step": 149
+ },
+ {
+ "epoch": 2.0837696335078535,
+ "grad_norm": 0.09615885466337204,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.35154914855957,
+ "step": 150
+ },
+ {
+ "epoch": 2.0977312390924956,
+ "grad_norm": 0.06462561339139938,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.335790634155273,
+ "step": 151
+ },
+ {
+ "epoch": 2.1116928446771377,
+ "grad_norm": 0.05186621472239494,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.324396133422852,
+ "step": 152
+ },
+ {
+ "epoch": 2.1256544502617802,
+ "grad_norm": 0.06095981225371361,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.318549156188965,
+ "step": 153
+ },
+ {
+ "epoch": 2.1396160558464223,
+ "grad_norm": 0.036121442914009094,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.300607681274414,
+ "step": 154
+ },
+ {
+ "epoch": 2.1535776614310644,
+ "grad_norm": 0.055716633796691895,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.307400703430176,
+ "step": 155
+ },
+ {
+ "epoch": 2.167539267015707,
+ "grad_norm": 0.037626396864652634,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.287364959716797,
+ "step": 156
+ },
+ {
+ "epoch": 2.181500872600349,
+ "grad_norm": 0.04329098388552666,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.28244686126709,
+ "step": 157
+ },
+ {
+ "epoch": 2.195462478184991,
+ "grad_norm": 0.031868450343608856,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.2602410316467285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2094240837696333,
+ "grad_norm": 0.03647344186902046,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.263988971710205,
+ "step": 159
+ },
+ {
+ "epoch": 2.223385689354276,
+ "grad_norm": 0.03141765296459198,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.263205051422119,
+ "step": 160
+ },
+ {
+ "epoch": 2.237347294938918,
+ "grad_norm": 0.028253229334950447,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.253107070922852,
+ "step": 161
+ },
+ {
+ "epoch": 2.25130890052356,
+ "grad_norm": 0.03608971834182739,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.2571702003479,
+ "step": 162
+ },
+ {
+ "epoch": 2.2652705061082026,
+ "grad_norm": 0.02644292451441288,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.241732597351074,
+ "step": 163
+ },
+ {
+ "epoch": 2.2792321116928447,
+ "grad_norm": 0.028604455292224884,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.230289459228516,
+ "step": 164
+ },
+ {
+ "epoch": 2.2931937172774868,
+ "grad_norm": 0.023704614490270615,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.226101398468018,
+ "step": 165
+ },
+ {
+ "epoch": 2.3071553228621293,
+ "grad_norm": 0.020691027864813805,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.2096405029296875,
+ "step": 166
+ },
+ {
+ "epoch": 2.3211169284467714,
+ "grad_norm": 0.025598427280783653,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.22075080871582,
+ "step": 167
+ },
+ {
+ "epoch": 2.3350785340314135,
+ "grad_norm": 0.030546002089977264,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.216886520385742,
+ "step": 168
+ },
+ {
+ "epoch": 2.349040139616056,
+ "grad_norm": 0.026066245511174202,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.2022504806518555,
+ "step": 169
+ },
+ {
+ "epoch": 2.363001745200698,
+ "grad_norm": 0.023802239447832108,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.189509391784668,
+ "step": 170
+ },
+ {
+ "epoch": 2.3769633507853403,
+ "grad_norm": 0.029805002734065056,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.186277389526367,
+ "step": 171
+ },
+ {
+ "epoch": 2.3909249563699824,
+ "grad_norm": 0.043327830731868744,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.164876937866211,
+ "step": 172
+ },
+ {
+ "epoch": 2.404886561954625,
+ "grad_norm": 0.058787211775779724,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.179936408996582,
+ "step": 173
+ },
+ {
+ "epoch": 2.418848167539267,
+ "grad_norm": 0.0787397176027298,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.1830244064331055,
+ "step": 174
+ },
+ {
+ "epoch": 2.432809773123909,
+ "grad_norm": 0.0899854376912117,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.190343856811523,
+ "step": 175
+ },
+ {
+ "epoch": 2.4467713787085517,
+ "grad_norm": 0.06520707905292511,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.185057163238525,
+ "step": 176
+ },
+ {
+ "epoch": 2.4607329842931938,
+ "grad_norm": 0.1002504974603653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.171279430389404,
+ "step": 177
+ },
+ {
+ "epoch": 2.474694589877836,
+ "grad_norm": 0.09591004997491837,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.174382209777832,
+ "step": 178
+ },
+ {
+ "epoch": 2.488656195462478,
+ "grad_norm": 0.12127121537923813,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.185717582702637,
+ "step": 179
+ },
+ {
+ "epoch": 2.5026178010471205,
+ "grad_norm": 0.09640590846538544,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.196681976318359,
+ "step": 180
+ },
+ {
+ "epoch": 2.5165794066317626,
+ "grad_norm": 0.048341453075408936,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.158774375915527,
+ "step": 181
+ },
+ {
+ "epoch": 2.530541012216405,
+ "grad_norm": 0.08206334710121155,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.17445707321167,
+ "step": 182
+ },
+ {
+ "epoch": 2.5445026178010473,
+ "grad_norm": 0.038934968411922455,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.170527458190918,
+ "step": 183
+ },
+ {
+ "epoch": 2.5584642233856894,
+ "grad_norm": 0.060987457633018494,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.1501994132995605,
+ "step": 184
+ },
+ {
+ "epoch": 2.5724258289703315,
+ "grad_norm": 0.033269546926021576,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.144186019897461,
+ "step": 185
+ },
+ {
+ "epoch": 2.5863874345549736,
+ "grad_norm": 0.046137236058712006,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.1236677169799805,
+ "step": 186
+ },
+ {
+ "epoch": 2.600349040139616,
+ "grad_norm": 0.038265109062194824,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.121979236602783,
+ "step": 187
+ },
+ {
+ "epoch": 2.614310645724258,
+ "grad_norm": 0.039128001779317856,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.124156951904297,
+ "step": 188
+ },
+ {
+ "epoch": 2.6282722513089007,
+ "grad_norm": 0.025303540751338005,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.108036994934082,
+ "step": 189
+ },
+ {
+ "epoch": 2.642233856893543,
+ "grad_norm": 0.04441308230161667,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.1211161613464355,
+ "step": 190
+ },
+ {
+ "epoch": 2.656195462478185,
+ "grad_norm": 0.03083532303571701,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.094561576843262,
+ "step": 191
+ },
+ {
+ "epoch": 2.670157068062827,
+ "grad_norm": 0.02951296605169773,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.091608047485352,
+ "step": 192
+ },
+ {
+ "epoch": 2.6841186736474696,
+ "grad_norm": 0.02385987527668476,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.087080001831055,
+ "step": 193
+ },
+ {
+ "epoch": 2.6980802792321117,
+ "grad_norm": 0.03315780311822891,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.088132381439209,
+ "step": 194
+ },
+ {
+ "epoch": 2.712041884816754,
+ "grad_norm": 0.02410835586488247,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.081460952758789,
+ "step": 195
+ },
+ {
+ "epoch": 2.7260034904013963,
+ "grad_norm": 0.033069487661123276,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.0696539878845215,
+ "step": 196
+ },
+ {
+ "epoch": 2.7399650959860384,
+ "grad_norm": 0.031288471072912216,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.047612190246582,
+ "step": 197
+ },
+ {
+ "epoch": 2.7539267015706805,
+ "grad_norm": 0.03399677947163582,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.019026756286621,
+ "step": 198
+ },
+ {
+ "epoch": 2.7678883071553226,
+ "grad_norm": 0.031948331743478775,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.050133228302002,
+ "step": 199
+ },
+ {
+ "epoch": 2.781849912739965,
+ "grad_norm": 0.049206435680389404,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.045332431793213,
+ "step": 200
+ },
+ {
+ "epoch": 2.7958115183246073,
+ "grad_norm": 0.07528849691152573,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.059202194213867,
+ "step": 201
+ },
+ {
+ "epoch": 2.80977312390925,
+ "grad_norm": 0.11815470457077026,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.077486038208008,
+ "step": 202
+ },
+ {
+ "epoch": 2.823734729493892,
+ "grad_norm": 0.0867348462343216,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.045670032501221,
+ "step": 203
+ },
+ {
+ "epoch": 2.837696335078534,
+ "grad_norm": 0.05716705322265625,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.034373760223389,
+ "step": 204
+ },
+ {
+ "epoch": 2.851657940663176,
+ "grad_norm": 0.05944284051656723,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.049911975860596,
+ "step": 205
+ },
+ {
+ "epoch": 2.8656195462478182,
+ "grad_norm": 0.03220470994710922,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.039861679077148,
+ "step": 206
+ },
+ {
+ "epoch": 2.8795811518324608,
+ "grad_norm": 0.057081177830696106,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.010714530944824,
+ "step": 207
+ },
+ {
+ "epoch": 2.893542757417103,
+ "grad_norm": 0.04037516191601753,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.018877983093262,
+ "step": 208
+ },
+ {
+ "epoch": 2.9075043630017454,
+ "grad_norm": 0.036492880433797836,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.009493827819824,
+ "step": 209
+ },
+ {
+ "epoch": 2.9214659685863875,
+ "grad_norm": 0.04785023257136345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.0140862464904785,
+ "step": 210
+ },
+ {
+ "epoch": 2.9354275741710296,
+ "grad_norm": 0.07721036672592163,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.030359268188477,
+ "step": 211
+ },
+ {
+ "epoch": 2.9493891797556717,
+ "grad_norm": 0.14220190048217773,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.045989990234375,
+ "step": 212
+ },
+ {
+ "epoch": 2.9633507853403143,
+ "grad_norm": 0.10265158116817474,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.043968200683594,
+ "step": 213
+ },
+ {
+ "epoch": 2.9773123909249564,
+ "grad_norm": 0.051398951560258865,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.022278785705566,
+ "step": 214
+ },
+ {
+ "epoch": 2.9912739965095985,
+ "grad_norm": 0.06044618785381317,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.017296314239502,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.03284750506281853,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 3.7544798851013184,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7531244158744812,
+ "eval_runtime": 60.3986,
+ "eval_samples_per_second": 40.431,
+ "eval_steps_per_second": 0.646,
+ "step": 216
+ },
+ {
+ "epoch": 3.013961605584642,
+ "grad_norm": 0.03895840048789978,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 5.9985857009887695,
+ "step": 217
+ },
+ {
+ "epoch": 3.0279232111692846,
+ "grad_norm": 0.04554992541670799,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 5.984710693359375,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418848167539267,
+ "grad_norm": 0.03359116613864899,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 5.984078407287598,
+ "step": 219
+ },
+ {
+ "epoch": 3.055846422338569,
+ "grad_norm": 0.035171881318092346,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 5.979424476623535,
+ "step": 220
+ },
+ {
+ "epoch": 3.0698080279232114,
+ "grad_norm": 0.032774072140455246,
+ "learning_rate": 0.000598278794687186,
+ "loss": 5.94467830657959,
+ "step": 221
+ },
+ {
+ "epoch": 3.0837696335078535,
+ "grad_norm": 0.032842908054590225,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 5.95986270904541,
+ "step": 222
+ },
+ {
+ "epoch": 3.0977312390924956,
+ "grad_norm": 0.02870488539338112,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 5.951186656951904,
+ "step": 223
+ },
+ {
+ "epoch": 3.1116928446771377,
+ "grad_norm": 0.030777374282479286,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 5.942493438720703,
+ "step": 224
+ },
+ {
+ "epoch": 3.1256544502617802,
+ "grad_norm": 0.026439018547534943,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 5.938275337219238,
+ "step": 225
+ },
+ {
+ "epoch": 3.1396160558464223,
+ "grad_norm": 0.028689472004771233,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 5.931982040405273,
+ "step": 226
+ },
+ {
+ "epoch": 3.1535776614310644,
+ "grad_norm": 0.028100362047553062,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 5.919692039489746,
+ "step": 227
+ },
+ {
+ "epoch": 3.167539267015707,
+ "grad_norm": 0.02407177910208702,
+ "learning_rate": 0.00059807235435179,
+ "loss": 5.909029006958008,
+ "step": 228
+ },
+ {
+ "epoch": 3.181500872600349,
+ "grad_norm": 0.0237748846411705,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 5.9131598472595215,
+ "step": 229
+ },
+ {
+ "epoch": 3.195462478184991,
+ "grad_norm": 0.022418556734919548,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 5.909396171569824,
+ "step": 230
+ },
+ {
+ "epoch": 3.2094240837696333,
+ "grad_norm": 0.024656571447849274,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 5.885243892669678,
+ "step": 231
+ },
+ {
+ "epoch": 3.223385689354276,
+ "grad_norm": 0.02136118710041046,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 5.898958206176758,
+ "step": 232
+ },
+ {
+ "epoch": 3.237347294938918,
+ "grad_norm": 0.025290116667747498,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 5.886986255645752,
+ "step": 233
+ },
+ {
+ "epoch": 3.25130890052356,
+ "grad_norm": 0.02656404674053192,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 5.897647857666016,
+ "step": 234
+ },
+ {
+ "epoch": 3.2652705061082026,
+ "grad_norm": 0.019188793376088142,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 5.875659942626953,
+ "step": 235
+ },
+ {
+ "epoch": 3.2792321116928447,
+ "grad_norm": 0.02308926172554493,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 5.859380722045898,
+ "step": 236
+ },
+ {
+ "epoch": 3.2931937172774868,
+ "grad_norm": 0.029775725677609444,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 5.873961448669434,
+ "step": 237
+ },
+ {
+ "epoch": 3.3071553228621293,
+ "grad_norm": 0.03875259310007095,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 5.863171100616455,
+ "step": 238
+ },
+ {
+ "epoch": 3.3211169284467714,
+ "grad_norm": 0.052446842193603516,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 5.886547088623047,
+ "step": 239
+ },
+ {
+ "epoch": 3.3350785340314135,
+ "grad_norm": 0.05583837628364563,
+ "learning_rate": 0.000597691359743243,
+ "loss": 5.8849287033081055,
+ "step": 240
+ },
+ {
+ "epoch": 3.349040139616056,
+ "grad_norm": 0.034729067236185074,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 5.863419532775879,
+ "step": 241
+ },
+ {
+ "epoch": 3.363001745200698,
+ "grad_norm": 0.03590013459324837,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 5.852169990539551,
+ "step": 242
+ },
+ {
+ "epoch": 3.3769633507853403,
+ "grad_norm": 0.035041507333517075,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 5.836690902709961,
+ "step": 243
+ },
+ {
+ "epoch": 3.3909249563699824,
+ "grad_norm": 0.02857939340174198,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 5.836485862731934,
+ "step": 244
+ },
+ {
+ "epoch": 3.404886561954625,
+ "grad_norm": 0.03176642581820488,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 5.860719203948975,
+ "step": 245
+ },
+ {
+ "epoch": 3.418848167539267,
+ "grad_norm": 0.026105748489499092,
+ "learning_rate": 0.000597488038559129,
+ "loss": 5.8354949951171875,
+ "step": 246
+ },
+ {
+ "epoch": 3.432809773123909,
+ "grad_norm": 0.029094981029629707,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 5.846247673034668,
+ "step": 247
+ },
+ {
+ "epoch": 3.4467713787085517,
+ "grad_norm": 0.02388785220682621,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 5.837607383728027,
+ "step": 248
+ },
+ {
+ "epoch": 3.4607329842931938,
+ "grad_norm": 0.03094472736120224,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 5.832090377807617,
+ "step": 249
+ },
+ {
+ "epoch": 3.474694589877836,
+ "grad_norm": 0.03227658197283745,
+ "learning_rate": 0.000597347745466109,
+ "loss": 5.83722448348999,
+ "step": 250
+ },
+ {
+ "epoch": 3.488656195462478,
+ "grad_norm": 0.042813923209905624,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 5.851722717285156,
+ "step": 251
+ },
+ {
+ "epoch": 3.5026178010471205,
+ "grad_norm": 0.05825525149703026,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.795719146728516,
+ "step": 252
+ },
+ {
+ "epoch": 3.5165794066317626,
+ "grad_norm": 0.0867447778582573,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.813750267028809,
+ "step": 253
+ },
+ {
+ "epoch": 3.530541012216405,
+ "grad_norm": 0.09192710369825363,
+ "learning_rate": 0.000597203657966319,
+ "loss": 5.823488235473633,
+ "step": 254
+ },
+ {
+ "epoch": 3.5445026178010473,
+ "grad_norm": 0.06069248169660568,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 5.814342021942139,
+ "step": 255
+ },
+ {
+ "epoch": 3.5584642233856894,
+ "grad_norm": 0.06259952485561371,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 5.834272384643555,
+ "step": 256
+ },
+ {
+ "epoch": 3.5724258289703315,
+ "grad_norm": 0.0513639897108078,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.807951927185059,
+ "step": 257
+ },
+ {
+ "epoch": 3.5863874345549736,
+ "grad_norm": 0.03327798470854759,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.790238857269287,
+ "step": 258
+ },
+ {
+ "epoch": 3.600349040139616,
+ "grad_norm": 0.03963128477334976,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.7956695556640625,
+ "step": 259
+ },
+ {
+ "epoch": 3.614310645724258,
+ "grad_norm": 0.031153453513979912,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 5.7874531745910645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6282722513089007,
+ "grad_norm": 0.031419433653354645,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 5.779650688171387,
+ "step": 261
+ },
+ {
+ "epoch": 3.642233856893543,
+ "grad_norm": 0.0288684144616127,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.776577949523926,
+ "step": 262
+ },
+ {
+ "epoch": 3.656195462478185,
+ "grad_norm": 0.028315918520092964,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.778571128845215,
+ "step": 263
+ },
+ {
+ "epoch": 3.670157068062827,
+ "grad_norm": 0.024977795779705048,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 5.776470184326172,
+ "step": 264
+ },
+ {
+ "epoch": 3.6841186736474696,
+ "grad_norm": 0.026822632178664207,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.755108833312988,
+ "step": 265
+ },
+ {
+ "epoch": 3.6980802792321117,
+ "grad_norm": 0.02738630585372448,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.74041748046875,
+ "step": 266
+ },
+ {
+ "epoch": 3.712041884816754,
+ "grad_norm": 0.02361280657351017,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.764027118682861,
+ "step": 267
+ },
+ {
+ "epoch": 3.7260034904013963,
+ "grad_norm": 0.029926009476184845,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.780390739440918,
+ "step": 268
+ },
+ {
+ "epoch": 3.7399650959860384,
+ "grad_norm": 0.03519507870078087,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.758105278015137,
+ "step": 269
+ },
+ {
+ "epoch": 3.7539267015706805,
+ "grad_norm": 0.030114633962512016,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.757230758666992,
+ "step": 270
+ },
+ {
+ "epoch": 3.7678883071553226,
+ "grad_norm": 0.02467748522758484,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.7615461349487305,
+ "step": 271
+ },
+ {
+ "epoch": 3.781849912739965,
+ "grad_norm": 0.03425183147192001,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.749448776245117,
+ "step": 272
+ },
+ {
+ "epoch": 3.7958115183246073,
+ "grad_norm": 0.046307582408189774,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.7356977462768555,
+ "step": 273
+ },
+ {
+ "epoch": 3.80977312390925,
+ "grad_norm": 0.054102856665849686,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.741869926452637,
+ "step": 274
+ },
+ {
+ "epoch": 3.823734729493892,
+ "grad_norm": 0.07026689499616623,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.721103668212891,
+ "step": 275
+ },
+ {
+ "epoch": 3.837696335078534,
+ "grad_norm": 0.0733039602637291,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.731414794921875,
+ "step": 276
+ },
+ {
+ "epoch": 3.851657940663176,
+ "grad_norm": 0.05342920497059822,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.734922885894775,
+ "step": 277
+ },
+ {
+ "epoch": 3.8656195462478182,
+ "grad_norm": 0.0504680834710598,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.732746124267578,
+ "step": 278
+ },
+ {
+ "epoch": 3.8795811518324608,
+ "grad_norm": 0.038991592824459076,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.719473838806152,
+ "step": 279
+ },
+ {
+ "epoch": 3.893542757417103,
+ "grad_norm": 0.04434404894709587,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.709481239318848,
+ "step": 280
+ },
+ {
+ "epoch": 3.9075043630017454,
+ "grad_norm": 0.047285184264183044,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.693531036376953,
+ "step": 281
+ },
+ {
+ "epoch": 3.9214659685863875,
+ "grad_norm": 0.045081138610839844,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.728959083557129,
+ "step": 282
+ },
+ {
+ "epoch": 3.9354275741710296,
+ "grad_norm": 0.025133663788437843,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.719696998596191,
+ "step": 283
+ },
+ {
+ "epoch": 3.9493891797556717,
+ "grad_norm": 0.03353596478700638,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.700235366821289,
+ "step": 284
+ },
+ {
+ "epoch": 3.9633507853403143,
+ "grad_norm": 0.035162411630153656,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.684836387634277,
+ "step": 285
+ },
+ {
+ "epoch": 3.9773123909249564,
+ "grad_norm": 0.04025450348854065,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.692097187042236,
+ "step": 286
+ },
+ {
+ "epoch": 3.9912739965095985,
+ "grad_norm": 0.04995397478342056,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.6766862869262695,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.031619708985090256,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 3.5519747734069824,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.714687705039978,
+ "eval_runtime": 59.637,
+ "eval_samples_per_second": 40.948,
+ "eval_steps_per_second": 0.654,
+ "step": 288
+ },
+ {
+ "epoch": 4.013961605584642,
+ "grad_norm": 0.036328598856925964,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.666294097900391,
+ "step": 289
+ },
+ {
+ "epoch": 4.027923211169284,
+ "grad_norm": 0.07326336950063705,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.6879167556762695,
+ "step": 290
+ },
+ {
+ "epoch": 4.041884816753926,
+ "grad_norm": 0.07947526127099991,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.674883842468262,
+ "step": 291
+ },
+ {
+ "epoch": 4.055846422338569,
+ "grad_norm": 0.04599885269999504,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.688086986541748,
+ "step": 292
+ },
+ {
+ "epoch": 4.069808027923211,
+ "grad_norm": 0.0465150885283947,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.649483680725098,
+ "step": 293
+ },
+ {
+ "epoch": 4.0837696335078535,
+ "grad_norm": 0.038542453199625015,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.653055667877197,
+ "step": 294
+ },
+ {
+ "epoch": 4.097731239092496,
+ "grad_norm": 0.03659164905548096,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.65955924987793,
+ "step": 295
+ },
+ {
+ "epoch": 4.111692844677138,
+ "grad_norm": 0.03824038803577423,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.650853157043457,
+ "step": 296
+ },
+ {
+ "epoch": 4.12565445026178,
+ "grad_norm": 0.03591204807162285,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.643499374389648,
+ "step": 297
+ },
+ {
+ "epoch": 4.139616055846423,
+ "grad_norm": 0.03269682824611664,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.657600402832031,
+ "step": 298
+ },
+ {
+ "epoch": 4.153577661431065,
+ "grad_norm": 0.03343537449836731,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.6352081298828125,
+ "step": 299
+ },
+ {
+ "epoch": 4.167539267015707,
+ "grad_norm": 0.034283801913261414,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.626757621765137,
+ "step": 300
+ },
+ {
+ "epoch": 4.181500872600349,
+ "grad_norm": 0.034614723175764084,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.6316680908203125,
+ "step": 301
+ },
+ {
+ "epoch": 4.195462478184991,
+ "grad_norm": 0.0340128131210804,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.648160934448242,
+ "step": 302
+ },
+ {
+ "epoch": 4.209424083769633,
+ "grad_norm": 0.03068280965089798,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.619738578796387,
+ "step": 303
+ },
+ {
+ "epoch": 4.223385689354275,
+ "grad_norm": 0.030269058421254158,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.609238624572754,
+ "step": 304
+ },
+ {
+ "epoch": 4.237347294938918,
+ "grad_norm": 0.03612440824508667,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.602194786071777,
+ "step": 305
+ },
+ {
+ "epoch": 4.2513089005235605,
+ "grad_norm": 0.044292811304330826,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.61159610748291,
+ "step": 306
+ },
+ {
+ "epoch": 4.265270506108203,
+ "grad_norm": 0.04449291527271271,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.6274237632751465,
+ "step": 307
+ },
+ {
+ "epoch": 4.279232111692845,
+ "grad_norm": 0.05014493316411972,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.589034080505371,
+ "step": 308
+ },
+ {
+ "epoch": 4.293193717277487,
+ "grad_norm": 0.05286206305027008,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.600770950317383,
+ "step": 309
+ },
+ {
+ "epoch": 4.307155322862129,
+ "grad_norm": 0.048459846526384354,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.5796098709106445,
+ "step": 310
+ },
+ {
+ "epoch": 4.321116928446771,
+ "grad_norm": 0.05329663306474686,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.590407371520996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335078534031414,
+ "grad_norm": 0.052595119923353195,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.583693504333496,
+ "step": 312
+ },
+ {
+ "epoch": 4.349040139616056,
+ "grad_norm": 0.05164601653814316,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.603281021118164,
+ "step": 313
+ },
+ {
+ "epoch": 4.363001745200698,
+ "grad_norm": 0.04657137766480446,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.591344833374023,
+ "step": 314
+ },
+ {
+ "epoch": 4.37696335078534,
+ "grad_norm": 0.03804599866271019,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.580578804016113,
+ "step": 315
+ },
+ {
+ "epoch": 4.390924956369982,
+ "grad_norm": 0.03835320845246315,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.597836971282959,
+ "step": 316
+ },
+ {
+ "epoch": 4.4048865619546245,
+ "grad_norm": 0.034059133380651474,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.567502021789551,
+ "step": 317
+ },
+ {
+ "epoch": 4.418848167539267,
+ "grad_norm": 0.03350386768579483,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.552022933959961,
+ "step": 318
+ },
+ {
+ "epoch": 4.43280977312391,
+ "grad_norm": 0.032574303448200226,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.559177875518799,
+ "step": 319
+ },
+ {
+ "epoch": 4.446771378708552,
+ "grad_norm": 0.030893854796886444,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.549288272857666,
+ "step": 320
+ },
+ {
+ "epoch": 4.460732984293194,
+ "grad_norm": 0.03255036100745201,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.5400390625,
+ "step": 321
+ },
+ {
+ "epoch": 4.474694589877836,
+ "grad_norm": 0.03009258769452572,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.5328474044799805,
+ "step": 322
+ },
+ {
+ "epoch": 4.488656195462478,
+ "grad_norm": 0.03178320452570915,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.5594482421875,
+ "step": 323
+ },
+ {
+ "epoch": 4.50261780104712,
+ "grad_norm": 0.03360465541481972,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.5402727127075195,
+ "step": 324
+ },
+ {
+ "epoch": 4.516579406631763,
+ "grad_norm": 0.04275049269199371,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.534474849700928,
+ "step": 325
+ },
+ {
+ "epoch": 4.530541012216405,
+ "grad_norm": 0.04140641540288925,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.5324859619140625,
+ "step": 326
+ },
+ {
+ "epoch": 4.544502617801047,
+ "grad_norm": 0.04769204929471016,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.4919352531433105,
+ "step": 327
+ },
+ {
+ "epoch": 4.558464223385689,
+ "grad_norm": 0.04402285814285278,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.520858287811279,
+ "step": 328
+ },
+ {
+ "epoch": 4.5724258289703315,
+ "grad_norm": 0.038526229560375214,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.523778438568115,
+ "step": 329
+ },
+ {
+ "epoch": 4.5863874345549736,
+ "grad_norm": 0.05422377213835716,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.517973899841309,
+ "step": 330
+ },
+ {
+ "epoch": 4.600349040139616,
+ "grad_norm": 0.07236721366643906,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.548083782196045,
+ "step": 331
+ },
+ {
+ "epoch": 4.614310645724259,
+ "grad_norm": 0.06861061602830887,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.535577774047852,
+ "step": 332
+ },
+ {
+ "epoch": 4.628272251308901,
+ "grad_norm": 0.07148340344429016,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.54155158996582,
+ "step": 333
+ },
+ {
+ "epoch": 4.642233856893543,
+ "grad_norm": 0.07838302850723267,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.552852630615234,
+ "step": 334
+ },
+ {
+ "epoch": 4.656195462478185,
+ "grad_norm": 0.10946758091449738,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.555066108703613,
+ "step": 335
+ },
+ {
+ "epoch": 4.670157068062827,
+ "grad_norm": 0.10143530368804932,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.584932804107666,
+ "step": 336
+ },
+ {
+ "epoch": 4.684118673647469,
+ "grad_norm": 0.08926606178283691,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.570462226867676,
+ "step": 337
+ },
+ {
+ "epoch": 4.698080279232112,
+ "grad_norm": 0.07060826569795609,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.5716753005981445,
+ "step": 338
+ },
+ {
+ "epoch": 4.712041884816754,
+ "grad_norm": 0.0418456569314003,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.556496620178223,
+ "step": 339
+ },
+ {
+ "epoch": 4.726003490401396,
+ "grad_norm": 0.05952411890029907,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.567801475524902,
+ "step": 340
+ },
+ {
+ "epoch": 4.739965095986038,
+ "grad_norm": 0.042999885976314545,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.537403106689453,
+ "step": 341
+ },
+ {
+ "epoch": 4.7539267015706805,
+ "grad_norm": 0.04426449537277222,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.552387237548828,
+ "step": 342
+ },
+ {
+ "epoch": 4.767888307155323,
+ "grad_norm": 0.040835145860910416,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.53051233291626,
+ "step": 343
+ },
+ {
+ "epoch": 4.781849912739965,
+ "grad_norm": 0.03819291293621063,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.531174659729004,
+ "step": 344
+ },
+ {
+ "epoch": 4.795811518324607,
+ "grad_norm": 0.030088599771261215,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.499328136444092,
+ "step": 345
+ },
+ {
+ "epoch": 4.80977312390925,
+ "grad_norm": 0.029984362423419952,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.504540920257568,
+ "step": 346
+ },
+ {
+ "epoch": 4.823734729493892,
+ "grad_norm": 0.027402915060520172,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.50899076461792,
+ "step": 347
+ },
+ {
+ "epoch": 4.837696335078534,
+ "grad_norm": 0.025096213445067406,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.480479717254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.851657940663176,
+ "grad_norm": 0.026828132569789886,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.501729965209961,
+ "step": 349
+ },
+ {
+ "epoch": 4.865619546247818,
+ "grad_norm": 0.023715168237686157,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.466860771179199,
+ "step": 350
+ },
+ {
+ "epoch": 4.879581151832461,
+ "grad_norm": 0.024503640830516815,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.491827964782715,
+ "step": 351
+ },
+ {
+ "epoch": 4.893542757417103,
+ "grad_norm": 0.020891105756163597,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.478384971618652,
+ "step": 352
+ },
+ {
+ "epoch": 4.907504363001745,
+ "grad_norm": 0.021373765543103218,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.4612135887146,
+ "step": 353
+ },
+ {
+ "epoch": 4.9214659685863875,
+ "grad_norm": 0.021773340180516243,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.4547319412231445,
+ "step": 354
+ },
+ {
+ "epoch": 4.93542757417103,
+ "grad_norm": 0.019637586548924446,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.431741714477539,
+ "step": 355
+ },
+ {
+ "epoch": 4.949389179755672,
+ "grad_norm": 0.023108696565032005,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.448612213134766,
+ "step": 356
+ },
+ {
+ "epoch": 4.963350785340314,
+ "grad_norm": 0.021935103461146355,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.447112083435059,
+ "step": 357
+ },
+ {
+ "epoch": 4.977312390924956,
+ "grad_norm": 0.023260753601789474,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.436091423034668,
+ "step": 358
+ },
+ {
+ "epoch": 4.991273996509599,
+ "grad_norm": 0.021355057135224342,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.408421993255615,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.014992325566709042,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 3.4004223346710205,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6834986805915833,
+ "eval_runtime": 60.4148,
+ "eval_samples_per_second": 40.421,
+ "eval_steps_per_second": 0.646,
+ "step": 360
+ },
+ {
+ "epoch": 5.013961605584642,
+ "grad_norm": 0.02471218630671501,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.396238803863525,
+ "step": 361
+ },
+ {
+ "epoch": 5.027923211169284,
+ "grad_norm": 0.027672411873936653,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.4048919677734375,
+ "step": 362
+ },
+ {
+ "epoch": 5.041884816753926,
+ "grad_norm": 0.031560689210891724,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.399665832519531,
+ "step": 363
+ },
+ {
+ "epoch": 5.055846422338569,
+ "grad_norm": 0.03459061309695244,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.396360397338867,
+ "step": 364
+ },
+ {
+ "epoch": 5.069808027923211,
+ "grad_norm": 0.04040960222482681,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.405309677124023,
+ "step": 365
+ },
+ {
+ "epoch": 5.0837696335078535,
+ "grad_norm": 0.05444670096039772,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.408894062042236,
+ "step": 366
+ },
+ {
+ "epoch": 5.097731239092496,
+ "grad_norm": 0.05816745012998581,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.403566360473633,
+ "step": 367
+ },
+ {
+ "epoch": 5.111692844677138,
+ "grad_norm": 0.06330069154500961,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.395961761474609,
+ "step": 368
+ },
+ {
+ "epoch": 5.12565445026178,
+ "grad_norm": 0.07458917796611786,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.400968551635742,
+ "step": 369
+ },
+ {
+ "epoch": 5.139616055846423,
+ "grad_norm": 0.06231356039643288,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.412624359130859,
+ "step": 370
+ },
+ {
+ "epoch": 5.153577661431065,
+ "grad_norm": 0.05212346464395523,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.394794464111328,
+ "step": 371
+ },
+ {
+ "epoch": 5.167539267015707,
+ "grad_norm": 0.050003375858068466,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.380125045776367,
+ "step": 372
+ },
+ {
+ "epoch": 5.181500872600349,
+ "grad_norm": 0.048393815755844116,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.390806198120117,
+ "step": 373
+ },
+ {
+ "epoch": 5.195462478184991,
+ "grad_norm": 0.04200794920325279,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.387467384338379,
+ "step": 374
+ },
+ {
+ "epoch": 5.209424083769633,
+ "grad_norm": 0.04931448772549629,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.37552547454834,
+ "step": 375
+ },
+ {
+ "epoch": 5.223385689354275,
+ "grad_norm": 0.03851857781410217,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.373126983642578,
+ "step": 376
+ },
+ {
+ "epoch": 5.237347294938918,
+ "grad_norm": 0.03458717465400696,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.346826553344727,
+ "step": 377
+ },
+ {
+ "epoch": 5.2513089005235605,
+ "grad_norm": 0.031107677146792412,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.353341102600098,
+ "step": 378
+ },
+ {
+ "epoch": 5.265270506108203,
+ "grad_norm": 0.03077058307826519,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.358076095581055,
+ "step": 379
+ },
+ {
+ "epoch": 5.279232111692845,
+ "grad_norm": 0.028170401230454445,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.354019641876221,
+ "step": 380
+ },
+ {
+ "epoch": 5.293193717277487,
+ "grad_norm": 0.02871990203857422,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.368258476257324,
+ "step": 381
+ },
+ {
+ "epoch": 5.307155322862129,
+ "grad_norm": 0.03046584501862526,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.342505931854248,
+ "step": 382
+ },
+ {
+ "epoch": 5.321116928446771,
+ "grad_norm": 0.02781563811004162,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.339411735534668,
+ "step": 383
+ },
+ {
+ "epoch": 5.335078534031414,
+ "grad_norm": 0.028825288638472557,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.341948509216309,
+ "step": 384
+ },
+ {
+ "epoch": 5.349040139616056,
+ "grad_norm": 0.026684079319238663,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.340881824493408,
+ "step": 385
+ },
+ {
+ "epoch": 5.363001745200698,
+ "grad_norm": 0.02429167740046978,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.342733383178711,
+ "step": 386
+ },
+ {
+ "epoch": 5.37696335078534,
+ "grad_norm": 0.02855239436030388,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.313878536224365,
+ "step": 387
+ },
+ {
+ "epoch": 5.390924956369982,
+ "grad_norm": 0.029651548713445663,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.312665939331055,
+ "step": 388
+ },
+ {
+ "epoch": 5.4048865619546245,
+ "grad_norm": 0.027191132307052612,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3086838722229,
+ "step": 389
+ },
+ {
+ "epoch": 5.418848167539267,
+ "grad_norm": 0.026719525456428528,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.305359840393066,
+ "step": 390
+ },
+ {
+ "epoch": 5.43280977312391,
+ "grad_norm": 0.026763100177049637,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.3164381980896,
+ "step": 391
+ },
+ {
+ "epoch": 5.446771378708552,
+ "grad_norm": 0.03125109151005745,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.3083953857421875,
+ "step": 392
+ },
+ {
+ "epoch": 5.460732984293194,
+ "grad_norm": 0.0394827201962471,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.275996208190918,
+ "step": 393
+ },
+ {
+ "epoch": 5.474694589877836,
+ "grad_norm": 0.040579475462436676,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.292995929718018,
+ "step": 394
+ },
+ {
+ "epoch": 5.488656195462478,
+ "grad_norm": 0.037223003804683685,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.298526763916016,
+ "step": 395
+ },
+ {
+ "epoch": 5.50261780104712,
+ "grad_norm": 0.04207765311002731,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.287985801696777,
+ "step": 396
+ },
+ {
+ "epoch": 5.516579406631763,
+ "grad_norm": 0.048299964517354965,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.244608402252197,
+ "step": 397
+ },
+ {
+ "epoch": 5.530541012216405,
+ "grad_norm": 0.04650307446718216,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.285453796386719,
+ "step": 398
+ },
+ {
+ "epoch": 5.544502617801047,
+ "grad_norm": 0.045339711010456085,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.312187671661377,
+ "step": 399
+ },
+ {
+ "epoch": 5.558464223385689,
+ "grad_norm": 0.041067708283662796,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.292692184448242,
+ "step": 400
+ },
+ {
+ "epoch": 5.5724258289703315,
+ "grad_norm": 0.05455026403069496,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.253364562988281,
+ "step": 401
+ },
+ {
+ "epoch": 5.5863874345549736,
+ "grad_norm": 0.06660915911197662,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.262646675109863,
+ "step": 402
+ },
+ {
+ "epoch": 5.600349040139616,
+ "grad_norm": 0.06634793430566788,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.274322986602783,
+ "step": 403
+ },
+ {
+ "epoch": 5.614310645724259,
+ "grad_norm": 0.08996330946683884,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.287951469421387,
+ "step": 404
+ },
+ {
+ "epoch": 5.628272251308901,
+ "grad_norm": 0.08526470512151718,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.3007097244262695,
+ "step": 405
+ },
+ {
+ "epoch": 5.642233856893543,
+ "grad_norm": 0.08766836673021317,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.332538604736328,
+ "step": 406
+ },
+ {
+ "epoch": 5.656195462478185,
+ "grad_norm": 0.07055048644542694,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.296169281005859,
+ "step": 407
+ },
+ {
+ "epoch": 5.670157068062827,
+ "grad_norm": 0.07520575076341629,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.309683322906494,
+ "step": 408
+ },
+ {
+ "epoch": 5.684118673647469,
+ "grad_norm": 0.06115756183862686,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.311301231384277,
+ "step": 409
+ },
+ {
+ "epoch": 5.698080279232112,
+ "grad_norm": 0.05862593650817871,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.28066349029541,
+ "step": 410
+ },
+ {
+ "epoch": 5.712041884816754,
+ "grad_norm": 0.06169601157307625,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.286840438842773,
+ "step": 411
+ },
+ {
+ "epoch": 5.726003490401396,
+ "grad_norm": 0.05270719155669212,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.302594184875488,
+ "step": 412
+ },
+ {
+ "epoch": 5.739965095986038,
+ "grad_norm": 0.03912634402513504,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.262563705444336,
+ "step": 413
+ },
+ {
+ "epoch": 5.7539267015706805,
+ "grad_norm": 0.03215811401605606,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.269310474395752,
+ "step": 414
+ },
+ {
+ "epoch": 5.767888307155323,
+ "grad_norm": 0.03236516937613487,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.268766403198242,
+ "step": 415
+ },
+ {
+ "epoch": 5.781849912739965,
+ "grad_norm": 0.027429452165961266,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.283801555633545,
+ "step": 416
+ },
+ {
+ "epoch": 5.795811518324607,
+ "grad_norm": 0.02990538813173771,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.22291374206543,
+ "step": 417
+ },
+ {
+ "epoch": 5.80977312390925,
+ "grad_norm": 0.030410146340727806,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.2281599044799805,
+ "step": 418
+ },
+ {
+ "epoch": 5.823734729493892,
+ "grad_norm": 0.030068187043070793,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.250294208526611,
+ "step": 419
+ },
+ {
+ "epoch": 5.837696335078534,
+ "grad_norm": 0.02800513058900833,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.226508617401123,
+ "step": 420
+ },
+ {
+ "epoch": 5.851657940663176,
+ "grad_norm": 0.023823332041502,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.247838020324707,
+ "step": 421
+ },
+ {
+ "epoch": 5.865619546247818,
+ "grad_norm": 0.024974249303340912,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.2392473220825195,
+ "step": 422
+ },
+ {
+ "epoch": 5.879581151832461,
+ "grad_norm": 0.023320546373724937,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.209543704986572,
+ "step": 423
+ },
+ {
+ "epoch": 5.893542757417103,
+ "grad_norm": 0.024201491847634315,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.219725608825684,
+ "step": 424
+ },
+ {
+ "epoch": 5.907504363001745,
+ "grad_norm": 0.021649666130542755,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.203141212463379,
+ "step": 425
+ },
+ {
+ "epoch": 5.9214659685863875,
+ "grad_norm": 0.0248002577573061,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.179058074951172,
+ "step": 426
+ },
+ {
+ "epoch": 5.93542757417103,
+ "grad_norm": 0.028549587354063988,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.202126502990723,
+ "step": 427
+ },
+ {
+ "epoch": 5.949389179755672,
+ "grad_norm": 0.0331057533621788,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.208107948303223,
+ "step": 428
+ },
+ {
+ "epoch": 5.963350785340314,
+ "grad_norm": 0.031843364238739014,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.184683799743652,
+ "step": 429
+ },
+ {
+ "epoch": 5.977312390924956,
+ "grad_norm": 0.024927377700805664,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.180486679077148,
+ "step": 430
+ },
+ {
+ "epoch": 5.991273996509599,
+ "grad_norm": 0.026685267686843872,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.186746597290039,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.018545351922512054,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 3.217043876647949,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.653471827507019,
+ "eval_runtime": 61.1586,
+ "eval_samples_per_second": 39.929,
+ "eval_steps_per_second": 0.638,
+ "step": 432
+ },
+ {
+ "epoch": 6.013961605584642,
+ "grad_norm": 0.025291118770837784,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.156534194946289,
+ "step": 433
+ },
+ {
+ "epoch": 6.027923211169284,
+ "grad_norm": 0.03774752840399742,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.144041538238525,
+ "step": 434
+ },
+ {
+ "epoch": 6.041884816753926,
+ "grad_norm": 0.04244692996144295,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.158862113952637,
+ "step": 435
+ },
+ {
+ "epoch": 6.055846422338569,
+ "grad_norm": 0.039782389998435974,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.166165351867676,
+ "step": 436
+ },
+ {
+ "epoch": 6.069808027923211,
+ "grad_norm": 0.05489896610379219,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.155766487121582,
+ "step": 437
+ },
+ {
+ "epoch": 6.0837696335078535,
+ "grad_norm": 0.059844326227903366,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.153733253479004,
+ "step": 438
+ },
+ {
+ "epoch": 6.097731239092496,
+ "grad_norm": 0.05539938807487488,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.130112648010254,
+ "step": 439
+ },
+ {
+ "epoch": 6.111692844677138,
+ "grad_norm": 0.05283183231949806,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.136910915374756,
+ "step": 440
+ },
+ {
+ "epoch": 6.12565445026178,
+ "grad_norm": 0.054407261312007904,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.165216445922852,
+ "step": 441
+ },
+ {
+ "epoch": 6.139616055846423,
+ "grad_norm": 0.0509788915514946,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.140888214111328,
+ "step": 442
+ },
+ {
+ "epoch": 6.153577661431065,
+ "grad_norm": 0.05018748342990875,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.13115930557251,
+ "step": 443
+ },
+ {
+ "epoch": 6.167539267015707,
+ "grad_norm": 0.047012437134981155,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.146507263183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181500872600349,
+ "grad_norm": 0.04943335801362991,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.135608673095703,
+ "step": 445
+ },
+ {
+ "epoch": 6.195462478184991,
+ "grad_norm": 0.05304260924458504,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.136394023895264,
+ "step": 446
+ },
+ {
+ "epoch": 6.209424083769633,
+ "grad_norm": 0.04554047808051109,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.124661922454834,
+ "step": 447
+ },
+ {
+ "epoch": 6.223385689354275,
+ "grad_norm": 0.040672432631254196,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.129773139953613,
+ "step": 448
+ },
+ {
+ "epoch": 6.237347294938918,
+ "grad_norm": 0.03442879393696785,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.128028392791748,
+ "step": 449
+ },
+ {
+ "epoch": 6.2513089005235605,
+ "grad_norm": 0.03927934914827347,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.092485427856445,
+ "step": 450
+ },
+ {
+ "epoch": 6.265270506108203,
+ "grad_norm": 0.029380977153778076,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.077669143676758,
+ "step": 451
+ },
+ {
+ "epoch": 6.279232111692845,
+ "grad_norm": 0.03331821784377098,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.138857841491699,
+ "step": 452
+ },
+ {
+ "epoch": 6.293193717277487,
+ "grad_norm": 0.02939118817448616,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.082937240600586,
+ "step": 453
+ },
+ {
+ "epoch": 6.307155322862129,
+ "grad_norm": 0.02730412967503071,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.093183517456055,
+ "step": 454
+ },
+ {
+ "epoch": 6.321116928446771,
+ "grad_norm": 0.02964780665934086,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.1028151512146,
+ "step": 455
+ },
+ {
+ "epoch": 6.335078534031414,
+ "grad_norm": 0.026350129395723343,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.066362380981445,
+ "step": 456
+ },
+ {
+ "epoch": 6.349040139616056,
+ "grad_norm": 0.027010295540094376,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.070037841796875,
+ "step": 457
+ },
+ {
+ "epoch": 6.363001745200698,
+ "grad_norm": 0.02906128577888012,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.107708930969238,
+ "step": 458
+ },
+ {
+ "epoch": 6.37696335078534,
+ "grad_norm": 0.028619443997740746,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.073182582855225,
+ "step": 459
+ },
+ {
+ "epoch": 6.390924956369982,
+ "grad_norm": 0.029981834813952446,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.076602935791016,
+ "step": 460
+ },
+ {
+ "epoch": 6.4048865619546245,
+ "grad_norm": 0.03255423903465271,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.063301086425781,
+ "step": 461
+ },
+ {
+ "epoch": 6.418848167539267,
+ "grad_norm": 0.03863866627216339,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.041812896728516,
+ "step": 462
+ },
+ {
+ "epoch": 6.43280977312391,
+ "grad_norm": 0.03860306739807129,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.06298828125,
+ "step": 463
+ },
+ {
+ "epoch": 6.446771378708552,
+ "grad_norm": 0.03378435596823692,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.0434370040893555,
+ "step": 464
+ },
+ {
+ "epoch": 6.460732984293194,
+ "grad_norm": 0.03276759013533592,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.036187171936035,
+ "step": 465
+ },
+ {
+ "epoch": 6.474694589877836,
+ "grad_norm": 0.04130513221025467,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.039098739624023,
+ "step": 466
+ },
+ {
+ "epoch": 6.488656195462478,
+ "grad_norm": 0.05389940366148949,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.062798976898193,
+ "step": 467
+ },
+ {
+ "epoch": 6.50261780104712,
+ "grad_norm": 0.05881704390048981,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.067028522491455,
+ "step": 468
+ },
+ {
+ "epoch": 6.516579406631763,
+ "grad_norm": 0.061897799372673035,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.056947231292725,
+ "step": 469
+ },
+ {
+ "epoch": 6.530541012216405,
+ "grad_norm": 0.06549323350191116,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.089262008666992,
+ "step": 470
+ },
+ {
+ "epoch": 6.544502617801047,
+ "grad_norm": 0.048082590103149414,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.055979251861572,
+ "step": 471
+ },
+ {
+ "epoch": 6.558464223385689,
+ "grad_norm": 0.03768027946352959,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.0660600662231445,
+ "step": 472
+ },
+ {
+ "epoch": 6.5724258289703315,
+ "grad_norm": 0.043922603130340576,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 5.036745071411133,
+ "step": 473
+ },
+ {
+ "epoch": 6.5863874345549736,
+ "grad_norm": 0.044848836958408356,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.073650360107422,
+ "step": 474
+ },
+ {
+ "epoch": 6.600349040139616,
+ "grad_norm": 0.05841026455163956,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.053045272827148,
+ "step": 475
+ },
+ {
+ "epoch": 6.614310645724259,
+ "grad_norm": 0.0650017261505127,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 5.052206039428711,
+ "step": 476
+ },
+ {
+ "epoch": 6.628272251308901,
+ "grad_norm": 0.048979002982378006,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 5.054277420043945,
+ "step": 477
+ },
+ {
+ "epoch": 6.642233856893543,
+ "grad_norm": 0.05085799843072891,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 5.063844680786133,
+ "step": 478
+ },
+ {
+ "epoch": 6.656195462478185,
+ "grad_norm": 0.04336017370223999,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 5.045474052429199,
+ "step": 479
+ },
+ {
+ "epoch": 6.670157068062827,
+ "grad_norm": 0.04113380238413811,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 5.031338214874268,
+ "step": 480
+ },
+ {
+ "epoch": 6.684118673647469,
+ "grad_norm": 0.03199538215994835,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 5.030025959014893,
+ "step": 481
+ },
+ {
+ "epoch": 6.698080279232112,
+ "grad_norm": 0.03043113648891449,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 5.033163070678711,
+ "step": 482
+ },
+ {
+ "epoch": 6.712041884816754,
+ "grad_norm": 0.031694378703832626,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 5.004739761352539,
+ "step": 483
+ },
+ {
+ "epoch": 6.726003490401396,
+ "grad_norm": 0.030243348330259323,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 5.012566089630127,
+ "step": 484
+ },
+ {
+ "epoch": 6.739965095986038,
+ "grad_norm": 0.02937416359782219,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 5.039803504943848,
+ "step": 485
+ },
+ {
+ "epoch": 6.7539267015706805,
+ "grad_norm": 0.029616443440318108,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 5.026117324829102,
+ "step": 486
+ },
+ {
+ "epoch": 6.767888307155323,
+ "grad_norm": 0.029567131772637367,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 5.006569862365723,
+ "step": 487
+ },
+ {
+ "epoch": 6.781849912739965,
+ "grad_norm": 0.029443370178341866,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.9941487312316895,
+ "step": 488
+ },
+ {
+ "epoch": 6.795811518324607,
+ "grad_norm": 0.03434352949261665,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.976431846618652,
+ "step": 489
+ },
+ {
+ "epoch": 6.80977312390925,
+ "grad_norm": 0.03845234215259552,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.993168354034424,
+ "step": 490
+ },
+ {
+ "epoch": 6.823734729493892,
+ "grad_norm": 0.05058208107948303,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.030162334442139,
+ "step": 491
+ },
+ {
+ "epoch": 6.837696335078534,
+ "grad_norm": 0.05961710214614868,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 5.007053375244141,
+ "step": 492
+ },
+ {
+ "epoch": 6.851657940663176,
+ "grad_norm": 0.056259628385305405,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 5.020356178283691,
+ "step": 493
+ },
+ {
+ "epoch": 6.865619546247818,
+ "grad_norm": 0.048628658056259155,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 5.037652969360352,
+ "step": 494
+ },
+ {
+ "epoch": 6.879581151832461,
+ "grad_norm": 0.040836453437805176,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.990390777587891,
+ "step": 495
+ },
+ {
+ "epoch": 6.893542757417103,
+ "grad_norm": 0.04112066701054573,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 5.017822742462158,
+ "step": 496
+ },
+ {
+ "epoch": 6.907504363001745,
+ "grad_norm": 0.03956829756498337,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.993971824645996,
+ "step": 497
+ },
+ {
+ "epoch": 6.9214659685863875,
+ "grad_norm": 0.03887917473912239,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.982889175415039,
+ "step": 498
+ },
+ {
+ "epoch": 6.93542757417103,
+ "grad_norm": 0.031012587249279022,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.961717128753662,
+ "step": 499
+ },
+ {
+ "epoch": 6.949389179755672,
+ "grad_norm": 0.03443489596247673,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.9904046058654785,
+ "step": 500
+ },
+ {
+ "epoch": 6.963350785340314,
+ "grad_norm": 0.035175010561943054,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.983619213104248,
+ "step": 501
+ },
+ {
+ "epoch": 6.977312390924956,
+ "grad_norm": 0.031724050641059875,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 5.000522613525391,
+ "step": 502
+ },
+ {
+ "epoch": 6.991273996509599,
+ "grad_norm": 0.032037995755672455,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.947064399719238,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.01883271522819996,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 3.112166166305542,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6296830177307129,
+ "eval_runtime": 59.5495,
+ "eval_samples_per_second": 41.008,
+ "eval_steps_per_second": 0.655,
+ "step": 504
+ },
+ {
+ "epoch": 7.013961605584642,
+ "grad_norm": 0.031091563403606415,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.919820785522461,
+ "step": 505
+ },
+ {
+ "epoch": 7.027923211169284,
+ "grad_norm": 0.04074421525001526,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.945765972137451,
+ "step": 506
+ },
+ {
+ "epoch": 7.041884816753926,
+ "grad_norm": 0.04225052148103714,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.935208797454834,
+ "step": 507
+ },
+ {
+ "epoch": 7.055846422338569,
+ "grad_norm": 0.0399051196873188,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.951314926147461,
+ "step": 508
+ },
+ {
+ "epoch": 7.069808027923211,
+ "grad_norm": 0.04785890504717827,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.92843770980835,
+ "step": 509
+ },
+ {
+ "epoch": 7.0837696335078535,
+ "grad_norm": 0.04922301694750786,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.901020050048828,
+ "step": 510
+ },
+ {
+ "epoch": 7.097731239092496,
+ "grad_norm": 0.05307295173406601,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.961151599884033,
+ "step": 511
+ },
+ {
+ "epoch": 7.111692844677138,
+ "grad_norm": 0.07003362476825714,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.9427876472473145,
+ "step": 512
+ },
+ {
+ "epoch": 7.12565445026178,
+ "grad_norm": 0.06498776376247406,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.947838306427002,
+ "step": 513
+ },
+ {
+ "epoch": 7.139616055846423,
+ "grad_norm": 0.058723628520965576,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.938859939575195,
+ "step": 514
+ },
+ {
+ "epoch": 7.153577661431065,
+ "grad_norm": 0.06590746343135834,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.962221145629883,
+ "step": 515
+ },
+ {
+ "epoch": 7.167539267015707,
+ "grad_norm": 0.052714601159095764,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.931011199951172,
+ "step": 516
+ },
+ {
+ "epoch": 7.181500872600349,
+ "grad_norm": 0.04296804964542389,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.93586540222168,
+ "step": 517
+ },
+ {
+ "epoch": 7.195462478184991,
+ "grad_norm": 0.03699173778295517,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.910763263702393,
+ "step": 518
+ },
+ {
+ "epoch": 7.209424083769633,
+ "grad_norm": 0.0317913293838501,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.9256439208984375,
+ "step": 519
+ },
+ {
+ "epoch": 7.223385689354275,
+ "grad_norm": 0.030313925817608833,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.938329696655273,
+ "step": 520
+ },
+ {
+ "epoch": 7.237347294938918,
+ "grad_norm": 0.02879229187965393,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.924966812133789,
+ "step": 521
+ },
+ {
+ "epoch": 7.2513089005235605,
+ "grad_norm": 0.02689451351761818,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.9079742431640625,
+ "step": 522
+ },
+ {
+ "epoch": 7.265270506108203,
+ "grad_norm": 0.028403930366039276,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.913238525390625,
+ "step": 523
+ },
+ {
+ "epoch": 7.279232111692845,
+ "grad_norm": 0.025912854820489883,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.896944046020508,
+ "step": 524
+ },
+ {
+ "epoch": 7.293193717277487,
+ "grad_norm": 0.024899760261178017,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.884920120239258,
+ "step": 525
+ },
+ {
+ "epoch": 7.307155322862129,
+ "grad_norm": 0.026912573724985123,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.9181060791015625,
+ "step": 526
+ },
+ {
+ "epoch": 7.321116928446771,
+ "grad_norm": 0.029539721086621284,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.914867401123047,
+ "step": 527
+ },
+ {
+ "epoch": 7.335078534031414,
+ "grad_norm": 0.028527872636914253,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.907217979431152,
+ "step": 528
+ },
+ {
+ "epoch": 7.349040139616056,
+ "grad_norm": 0.025879016146063805,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.905783176422119,
+ "step": 529
+ },
+ {
+ "epoch": 7.363001745200698,
+ "grad_norm": 0.026395199820399284,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.880505561828613,
+ "step": 530
+ },
+ {
+ "epoch": 7.37696335078534,
+ "grad_norm": 0.025072351098060608,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.870449066162109,
+ "step": 531
+ },
+ {
+ "epoch": 7.390924956369982,
+ "grad_norm": 0.02751019410789013,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.888952255249023,
+ "step": 532
+ },
+ {
+ "epoch": 7.4048865619546245,
+ "grad_norm": 0.025810658931732178,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.894552230834961,
+ "step": 533
+ },
+ {
+ "epoch": 7.418848167539267,
+ "grad_norm": 0.02399870567023754,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.8768229484558105,
+ "step": 534
+ },
+ {
+ "epoch": 7.43280977312391,
+ "grad_norm": 0.024639368057250977,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.9030914306640625,
+ "step": 535
+ },
+ {
+ "epoch": 7.446771378708552,
+ "grad_norm": 0.02543291077017784,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.86467170715332,
+ "step": 536
+ },
+ {
+ "epoch": 7.460732984293194,
+ "grad_norm": 0.02702544629573822,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.888283729553223,
+ "step": 537
+ },
+ {
+ "epoch": 7.474694589877836,
+ "grad_norm": 0.028904743492603302,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.866081237792969,
+ "step": 538
+ },
+ {
+ "epoch": 7.488656195462478,
+ "grad_norm": 0.03055848367512226,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.890739440917969,
+ "step": 539
+ },
+ {
+ "epoch": 7.50261780104712,
+ "grad_norm": 0.03410084918141365,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.862918853759766,
+ "step": 540
+ },
+ {
+ "epoch": 7.516579406631763,
+ "grad_norm": 0.03738495334982872,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.864728927612305,
+ "step": 541
+ },
+ {
+ "epoch": 7.530541012216405,
+ "grad_norm": 0.037646882236003876,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.865185737609863,
+ "step": 542
+ },
+ {
+ "epoch": 7.544502617801047,
+ "grad_norm": 0.03418618440628052,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.866114139556885,
+ "step": 543
+ },
+ {
+ "epoch": 7.558464223385689,
+ "grad_norm": 0.036924004554748535,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.883186340332031,
+ "step": 544
+ },
+ {
+ "epoch": 7.5724258289703315,
+ "grad_norm": 0.03590598329901695,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.863725662231445,
+ "step": 545
+ },
+ {
+ "epoch": 7.5863874345549736,
+ "grad_norm": 0.037363454699516296,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.841030120849609,
+ "step": 546
+ },
+ {
+ "epoch": 7.600349040139616,
+ "grad_norm": 0.046537864953279495,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.873837471008301,
+ "step": 547
+ },
+ {
+ "epoch": 7.614310645724259,
+ "grad_norm": 0.057226769626140594,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.865679740905762,
+ "step": 548
+ },
+ {
+ "epoch": 7.628272251308901,
+ "grad_norm": 0.06175961345434189,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.888645648956299,
+ "step": 549
+ },
+ {
+ "epoch": 7.642233856893543,
+ "grad_norm": 0.05063694715499878,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.877828598022461,
+ "step": 550
+ },
+ {
+ "epoch": 7.656195462478185,
+ "grad_norm": 0.03950628638267517,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.88637638092041,
+ "step": 551
+ },
+ {
+ "epoch": 7.670157068062827,
+ "grad_norm": 0.04183337092399597,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.8597917556762695,
+ "step": 552
+ },
+ {
+ "epoch": 7.684118673647469,
+ "grad_norm": 0.04642999917268753,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.880380153656006,
+ "step": 553
+ },
+ {
+ "epoch": 7.698080279232112,
+ "grad_norm": 0.05333106964826584,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.853932857513428,
+ "step": 554
+ },
+ {
+ "epoch": 7.712041884816754,
+ "grad_norm": 0.04764975979924202,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.870235443115234,
+ "step": 555
+ },
+ {
+ "epoch": 7.726003490401396,
+ "grad_norm": 0.043072231113910675,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.822697639465332,
+ "step": 556
+ },
+ {
+ "epoch": 7.739965095986038,
+ "grad_norm": 0.03913857042789459,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.849332809448242,
+ "step": 557
+ },
+ {
+ "epoch": 7.7539267015706805,
+ "grad_norm": 0.034836407750844955,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.842280387878418,
+ "step": 558
+ },
+ {
+ "epoch": 7.767888307155323,
+ "grad_norm": 0.033344294875860214,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.837811470031738,
+ "step": 559
+ },
+ {
+ "epoch": 7.781849912739965,
+ "grad_norm": 0.03342701122164726,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.876040458679199,
+ "step": 560
+ },
+ {
+ "epoch": 7.795811518324607,
+ "grad_norm": 0.034638747572898865,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.830340385437012,
+ "step": 561
+ },
+ {
+ "epoch": 7.80977312390925,
+ "grad_norm": 0.03358135372400284,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.828607559204102,
+ "step": 562
+ },
+ {
+ "epoch": 7.823734729493892,
+ "grad_norm": 0.036001041531562805,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.827905654907227,
+ "step": 563
+ },
+ {
+ "epoch": 7.837696335078534,
+ "grad_norm": 0.038079630583524704,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.866955280303955,
+ "step": 564
+ },
+ {
+ "epoch": 7.851657940663176,
+ "grad_norm": 0.03728470206260681,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.828039169311523,
+ "step": 565
+ },
+ {
+ "epoch": 7.865619546247818,
+ "grad_norm": 0.03920375928282738,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.841631889343262,
+ "step": 566
+ },
+ {
+ "epoch": 7.879581151832461,
+ "grad_norm": 0.032353926450014114,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.861110210418701,
+ "step": 567
+ },
+ {
+ "epoch": 7.893542757417103,
+ "grad_norm": 0.03241422027349472,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.818441390991211,
+ "step": 568
+ },
+ {
+ "epoch": 7.907504363001745,
+ "grad_norm": 0.03879080340266228,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.823749542236328,
+ "step": 569
+ },
+ {
+ "epoch": 7.9214659685863875,
+ "grad_norm": 0.03486498072743416,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.830697059631348,
+ "step": 570
+ },
+ {
+ "epoch": 7.93542757417103,
+ "grad_norm": 0.03279713913798332,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.815736770629883,
+ "step": 571
+ },
+ {
+ "epoch": 7.949389179755672,
+ "grad_norm": 0.03313925117254257,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.8364033699035645,
+ "step": 572
+ },
+ {
+ "epoch": 7.963350785340314,
+ "grad_norm": 0.03405062481760979,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.816010475158691,
+ "step": 573
+ },
+ {
+ "epoch": 7.977312390924956,
+ "grad_norm": 0.03259236365556717,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.813413619995117,
+ "step": 574
+ },
+ {
+ "epoch": 7.991273996509599,
+ "grad_norm": 0.030408985912799835,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.833840370178223,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.022680889815092087,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 2.99565052986145,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6126086115837097,
+ "eval_runtime": 60.2662,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 576
+ },
+ {
+ "epoch": 8.013961605584642,
+ "grad_norm": 0.029732126742601395,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.7816314697265625,
+ "step": 577
+ },
+ {
+ "epoch": 8.027923211169284,
+ "grad_norm": 0.04357453063130379,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.7575225830078125,
+ "step": 578
+ },
+ {
+ "epoch": 8.041884816753926,
+ "grad_norm": 0.051903221756219864,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.802519798278809,
+ "step": 579
+ },
+ {
+ "epoch": 8.055846422338568,
+ "grad_norm": 0.0539076030254364,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.799004554748535,
+ "step": 580
+ },
+ {
+ "epoch": 8.06980802792321,
+ "grad_norm": 0.043642807751894,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.7896528244018555,
+ "step": 581
+ },
+ {
+ "epoch": 8.083769633507853,
+ "grad_norm": 0.0469554178416729,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.765430927276611,
+ "step": 582
+ },
+ {
+ "epoch": 8.097731239092496,
+ "grad_norm": 0.042060062289237976,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.784132957458496,
+ "step": 583
+ },
+ {
+ "epoch": 8.111692844677139,
+ "grad_norm": 0.035101912915706635,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.767234802246094,
+ "step": 584
+ },
+ {
+ "epoch": 8.12565445026178,
+ "grad_norm": 0.042576879262924194,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.762052536010742,
+ "step": 585
+ },
+ {
+ "epoch": 8.139616055846423,
+ "grad_norm": 0.04231419041752815,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.752150535583496,
+ "step": 586
+ },
+ {
+ "epoch": 8.153577661431065,
+ "grad_norm": 0.046912599354982376,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.771881103515625,
+ "step": 587
+ },
+ {
+ "epoch": 8.167539267015707,
+ "grad_norm": 0.051271144300699234,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.773284912109375,
+ "step": 588
+ },
+ {
+ "epoch": 8.181500872600349,
+ "grad_norm": 0.04655958339571953,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.769917964935303,
+ "step": 589
+ },
+ {
+ "epoch": 8.195462478184991,
+ "grad_norm": 0.04463687539100647,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.77938175201416,
+ "step": 590
+ },
+ {
+ "epoch": 8.209424083769633,
+ "grad_norm": 0.047959886491298676,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.785962104797363,
+ "step": 591
+ },
+ {
+ "epoch": 8.223385689354275,
+ "grad_norm": 0.04414986073970795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.786281585693359,
+ "step": 592
+ },
+ {
+ "epoch": 8.237347294938917,
+ "grad_norm": 0.044150348752737045,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.777070045471191,
+ "step": 593
+ },
+ {
+ "epoch": 8.25130890052356,
+ "grad_norm": 0.037002332508563995,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.77547550201416,
+ "step": 594
+ },
+ {
+ "epoch": 8.265270506108202,
+ "grad_norm": 0.03417367860674858,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.782101154327393,
+ "step": 595
+ },
+ {
+ "epoch": 8.279232111692846,
+ "grad_norm": 0.03069210797548294,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.745595455169678,
+ "step": 596
+ },
+ {
+ "epoch": 8.293193717277488,
+ "grad_norm": 0.03228718787431717,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.7663655281066895,
+ "step": 597
+ },
+ {
+ "epoch": 8.30715532286213,
+ "grad_norm": 0.028603695333003998,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.767274856567383,
+ "step": 598
+ },
+ {
+ "epoch": 8.321116928446772,
+ "grad_norm": 0.03098377026617527,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.7853684425354,
+ "step": 599
+ },
+ {
+ "epoch": 8.335078534031414,
+ "grad_norm": 0.029242439195513725,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.738029956817627,
+ "step": 600
+ },
+ {
+ "epoch": 8.349040139616056,
+ "grad_norm": 0.02751081995666027,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.76801872253418,
+ "step": 601
+ },
+ {
+ "epoch": 8.363001745200698,
+ "grad_norm": 0.02838125079870224,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.743884086608887,
+ "step": 602
+ },
+ {
+ "epoch": 8.37696335078534,
+ "grad_norm": 0.029203062877058983,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.746668815612793,
+ "step": 603
+ },
+ {
+ "epoch": 8.390924956369982,
+ "grad_norm": 0.02679944597184658,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.760764122009277,
+ "step": 604
+ },
+ {
+ "epoch": 8.404886561954624,
+ "grad_norm": 0.028955848887562752,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.7353668212890625,
+ "step": 605
+ },
+ {
+ "epoch": 8.418848167539267,
+ "grad_norm": 0.02880851738154888,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.766837120056152,
+ "step": 606
+ },
+ {
+ "epoch": 8.432809773123909,
+ "grad_norm": 0.02495497092604637,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.78959321975708,
+ "step": 607
+ },
+ {
+ "epoch": 8.44677137870855,
+ "grad_norm": 0.026064753532409668,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.736433029174805,
+ "step": 608
+ },
+ {
+ "epoch": 8.460732984293193,
+ "grad_norm": 0.027566388249397278,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.782382965087891,
+ "step": 609
+ },
+ {
+ "epoch": 8.474694589877837,
+ "grad_norm": 0.02967359498143196,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.735286712646484,
+ "step": 610
+ },
+ {
+ "epoch": 8.488656195462479,
+ "grad_norm": 0.03079676441848278,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.725196838378906,
+ "step": 611
+ },
+ {
+ "epoch": 8.502617801047121,
+ "grad_norm": 0.032531846314668655,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.719333648681641,
+ "step": 612
+ },
+ {
+ "epoch": 8.516579406631763,
+ "grad_norm": 0.03120352514088154,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.723845958709717,
+ "step": 613
+ },
+ {
+ "epoch": 8.530541012216405,
+ "grad_norm": 0.03328424692153931,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.7669782638549805,
+ "step": 614
+ },
+ {
+ "epoch": 8.544502617801047,
+ "grad_norm": 0.035601649433374405,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.714925765991211,
+ "step": 615
+ },
+ {
+ "epoch": 8.55846422338569,
+ "grad_norm": 0.03675796464085579,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.715317726135254,
+ "step": 616
+ },
+ {
+ "epoch": 8.572425828970331,
+ "grad_norm": 0.03701988235116005,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.724606990814209,
+ "step": 617
+ },
+ {
+ "epoch": 8.586387434554974,
+ "grad_norm": 0.036719705909490585,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.746044635772705,
+ "step": 618
+ },
+ {
+ "epoch": 8.600349040139616,
+ "grad_norm": 0.034756604582071304,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.74885368347168,
+ "step": 619
+ },
+ {
+ "epoch": 8.614310645724258,
+ "grad_norm": 0.03237922117114067,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.721858501434326,
+ "step": 620
+ },
+ {
+ "epoch": 8.6282722513089,
+ "grad_norm": 0.028420142829418182,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.733814239501953,
+ "step": 621
+ },
+ {
+ "epoch": 8.642233856893542,
+ "grad_norm": 0.026203038170933723,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.726088523864746,
+ "step": 622
+ },
+ {
+ "epoch": 8.656195462478184,
+ "grad_norm": 0.02627231739461422,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.727700233459473,
+ "step": 623
+ },
+ {
+ "epoch": 8.670157068062828,
+ "grad_norm": 0.02920261025428772,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.731454849243164,
+ "step": 624
+ },
+ {
+ "epoch": 8.68411867364747,
+ "grad_norm": 0.02792907878756523,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.734154224395752,
+ "step": 625
+ },
+ {
+ "epoch": 8.698080279232112,
+ "grad_norm": 0.026770325377583504,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.715337753295898,
+ "step": 626
+ },
+ {
+ "epoch": 8.712041884816754,
+ "grad_norm": 0.027346540242433548,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.716512680053711,
+ "step": 627
+ },
+ {
+ "epoch": 8.726003490401396,
+ "grad_norm": 0.031965773552656174,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.7072038650512695,
+ "step": 628
+ },
+ {
+ "epoch": 8.739965095986038,
+ "grad_norm": 0.03708025813102722,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.749785900115967,
+ "step": 629
+ },
+ {
+ "epoch": 8.75392670157068,
+ "grad_norm": 0.04846490919589996,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.735586166381836,
+ "step": 630
+ },
+ {
+ "epoch": 8.767888307155323,
+ "grad_norm": 0.053352758288383484,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.720524787902832,
+ "step": 631
+ },
+ {
+ "epoch": 8.781849912739965,
+ "grad_norm": 0.040559202432632446,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.719901084899902,
+ "step": 632
+ },
+ {
+ "epoch": 8.795811518324607,
+ "grad_norm": 0.03333032503724098,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.708139419555664,
+ "step": 633
+ },
+ {
+ "epoch": 8.809773123909249,
+ "grad_norm": 0.030532941222190857,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.728184223175049,
+ "step": 634
+ },
+ {
+ "epoch": 8.823734729493891,
+ "grad_norm": 0.03337649628520012,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.726020336151123,
+ "step": 635
+ },
+ {
+ "epoch": 8.837696335078533,
+ "grad_norm": 0.03352045267820358,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.7196855545043945,
+ "step": 636
+ },
+ {
+ "epoch": 8.851657940663177,
+ "grad_norm": 0.03662872314453125,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.732823371887207,
+ "step": 637
+ },
+ {
+ "epoch": 8.86561954624782,
+ "grad_norm": 0.037085745483636856,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.728443145751953,
+ "step": 638
+ },
+ {
+ "epoch": 8.879581151832461,
+ "grad_norm": 0.036635689437389374,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.713988780975342,
+ "step": 639
+ },
+ {
+ "epoch": 8.893542757417103,
+ "grad_norm": 0.03719659522175789,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.715950965881348,
+ "step": 640
+ },
+ {
+ "epoch": 8.907504363001745,
+ "grad_norm": 0.03719159588217735,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.7124924659729,
+ "step": 641
+ },
+ {
+ "epoch": 8.921465968586388,
+ "grad_norm": 0.04275721684098244,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.708404064178467,
+ "step": 642
+ },
+ {
+ "epoch": 8.93542757417103,
+ "grad_norm": 0.04204598814249039,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.715372085571289,
+ "step": 643
+ },
+ {
+ "epoch": 8.949389179755672,
+ "grad_norm": 0.03938259929418564,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.729745864868164,
+ "step": 644
+ },
+ {
+ "epoch": 8.963350785340314,
+ "grad_norm": 0.036226850003004074,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.733749866485596,
+ "step": 645
+ },
+ {
+ "epoch": 8.977312390924956,
+ "grad_norm": 0.03542874753475189,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.706825256347656,
+ "step": 646
+ },
+ {
+ "epoch": 8.991273996509598,
+ "grad_norm": 0.029055381193757057,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.718567848205566,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.01990380510687828,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 2.943183183670044,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.601699709892273,
+ "eval_runtime": 60.0221,
+ "eval_samples_per_second": 40.685,
+ "eval_steps_per_second": 0.65,
+ "step": 648
+ },
+ {
+ "epoch": 9.013961605584642,
+ "grad_norm": 0.030904127284884453,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.660327911376953,
+ "step": 649
+ },
+ {
+ "epoch": 9.027923211169284,
+ "grad_norm": 0.0441313236951828,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.701565742492676,
+ "step": 650
+ },
+ {
+ "epoch": 9.041884816753926,
+ "grad_norm": 0.047547727823257446,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.666440486907959,
+ "step": 651
+ },
+ {
+ "epoch": 9.055846422338568,
+ "grad_norm": 0.04549463838338852,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.660548210144043,
+ "step": 652
+ },
+ {
+ "epoch": 9.06980802792321,
+ "grad_norm": 0.050346389412879944,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.640229225158691,
+ "step": 653
+ },
+ {
+ "epoch": 9.083769633507853,
+ "grad_norm": 0.058505453169345856,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.673881530761719,
+ "step": 654
+ },
+ {
+ "epoch": 9.097731239092496,
+ "grad_norm": 0.051989227533340454,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.693943977355957,
+ "step": 655
+ },
+ {
+ "epoch": 9.111692844677139,
+ "grad_norm": 0.052564945071935654,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.644315242767334,
+ "step": 656
+ },
+ {
+ "epoch": 9.12565445026178,
+ "grad_norm": 0.05647841840982437,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.681768417358398,
+ "step": 657
+ },
+ {
+ "epoch": 9.139616055846423,
+ "grad_norm": 0.058640867471694946,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.676457405090332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153577661431065,
+ "grad_norm": 0.05386025458574295,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.659588813781738,
+ "step": 659
+ },
+ {
+ "epoch": 9.167539267015707,
+ "grad_norm": 0.048699021339416504,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.680031776428223,
+ "step": 660
+ },
+ {
+ "epoch": 9.181500872600349,
+ "grad_norm": 0.04374176263809204,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.672671318054199,
+ "step": 661
+ },
+ {
+ "epoch": 9.195462478184991,
+ "grad_norm": 0.042337436228990555,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.673431873321533,
+ "step": 662
+ },
+ {
+ "epoch": 9.209424083769633,
+ "grad_norm": 0.04564661160111427,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.632595062255859,
+ "step": 663
+ },
+ {
+ "epoch": 9.223385689354275,
+ "grad_norm": 0.04404835030436516,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.678891181945801,
+ "step": 664
+ },
+ {
+ "epoch": 9.237347294938917,
+ "grad_norm": 0.040642790496349335,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.633418560028076,
+ "step": 665
+ },
+ {
+ "epoch": 9.25130890052356,
+ "grad_norm": 0.036492060869932175,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.670109748840332,
+ "step": 666
+ },
+ {
+ "epoch": 9.265270506108202,
+ "grad_norm": 0.03735879808664322,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.608570098876953,
+ "step": 667
+ },
+ {
+ "epoch": 9.279232111692846,
+ "grad_norm": 0.034563641995191574,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.632505416870117,
+ "step": 668
+ },
+ {
+ "epoch": 9.293193717277488,
+ "grad_norm": 0.03522484749555588,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.6475629806518555,
+ "step": 669
+ },
+ {
+ "epoch": 9.30715532286213,
+ "grad_norm": 0.03216472268104553,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.627873420715332,
+ "step": 670
+ },
+ {
+ "epoch": 9.321116928446772,
+ "grad_norm": 0.029155759140849113,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.652706146240234,
+ "step": 671
+ },
+ {
+ "epoch": 9.335078534031414,
+ "grad_norm": 0.031060541048645973,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.644545078277588,
+ "step": 672
+ },
+ {
+ "epoch": 9.349040139616056,
+ "grad_norm": 0.034979186952114105,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.642645359039307,
+ "step": 673
+ },
+ {
+ "epoch": 9.363001745200698,
+ "grad_norm": 0.035258810967206955,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.647383689880371,
+ "step": 674
+ },
+ {
+ "epoch": 9.37696335078534,
+ "grad_norm": 0.031552575528621674,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.667779445648193,
+ "step": 675
+ },
+ {
+ "epoch": 9.390924956369982,
+ "grad_norm": 0.03193951025605202,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.63625431060791,
+ "step": 676
+ },
+ {
+ "epoch": 9.404886561954624,
+ "grad_norm": 0.029811711981892586,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.658000946044922,
+ "step": 677
+ },
+ {
+ "epoch": 9.418848167539267,
+ "grad_norm": 0.028464138507843018,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.6545515060424805,
+ "step": 678
+ },
+ {
+ "epoch": 9.432809773123909,
+ "grad_norm": 0.0315735749900341,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.641153335571289,
+ "step": 679
+ },
+ {
+ "epoch": 9.44677137870855,
+ "grad_norm": 0.033412106335163116,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.626931667327881,
+ "step": 680
+ },
+ {
+ "epoch": 9.460732984293193,
+ "grad_norm": 0.03757626190781593,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.671254634857178,
+ "step": 681
+ },
+ {
+ "epoch": 9.474694589877837,
+ "grad_norm": 0.03910793736577034,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.651954174041748,
+ "step": 682
+ },
+ {
+ "epoch": 9.488656195462479,
+ "grad_norm": 0.04015549272298813,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.613324165344238,
+ "step": 683
+ },
+ {
+ "epoch": 9.502617801047121,
+ "grad_norm": 0.040689144283533096,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.639243125915527,
+ "step": 684
+ },
+ {
+ "epoch": 9.516579406631763,
+ "grad_norm": 0.04075361415743828,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.649250030517578,
+ "step": 685
+ },
+ {
+ "epoch": 9.530541012216405,
+ "grad_norm": 0.03914607688784599,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.66444730758667,
+ "step": 686
+ },
+ {
+ "epoch": 9.544502617801047,
+ "grad_norm": 0.033253733068704605,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.647335052490234,
+ "step": 687
+ },
+ {
+ "epoch": 9.55846422338569,
+ "grad_norm": 0.03179008886218071,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.6727800369262695,
+ "step": 688
+ },
+ {
+ "epoch": 9.572425828970331,
+ "grad_norm": 0.03477025404572487,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.655778884887695,
+ "step": 689
+ },
+ {
+ "epoch": 9.586387434554974,
+ "grad_norm": 0.030094800516963005,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.680851936340332,
+ "step": 690
+ },
+ {
+ "epoch": 9.600349040139616,
+ "grad_norm": 0.030153103172779083,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.654638290405273,
+ "step": 691
+ },
+ {
+ "epoch": 9.614310645724258,
+ "grad_norm": 0.029009278863668442,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.625916481018066,
+ "step": 692
+ },
+ {
+ "epoch": 9.6282722513089,
+ "grad_norm": 0.027248440310359,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.656370162963867,
+ "step": 693
+ },
+ {
+ "epoch": 9.642233856893542,
+ "grad_norm": 0.027727462351322174,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.631473541259766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656195462478184,
+ "grad_norm": 0.026710307225584984,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.636524200439453,
+ "step": 695
+ },
+ {
+ "epoch": 9.670157068062828,
+ "grad_norm": 0.02726181596517563,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.629701614379883,
+ "step": 696
+ },
+ {
+ "epoch": 9.68411867364747,
+ "grad_norm": 0.027146652340888977,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.640240669250488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698080279232112,
+ "grad_norm": 0.027165887877345085,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.64462947845459,
+ "step": 698
+ },
+ {
+ "epoch": 9.712041884816754,
+ "grad_norm": 0.03025982715189457,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.6271820068359375,
+ "step": 699
+ },
+ {
+ "epoch": 9.726003490401396,
+ "grad_norm": 0.02837025746703148,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.629697799682617,
+ "step": 700
+ },
+ {
+ "epoch": 9.739965095986038,
+ "grad_norm": 0.026777995750308037,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.646146774291992,
+ "step": 701
+ },
+ {
+ "epoch": 9.75392670157068,
+ "grad_norm": 0.024626802653074265,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.641671657562256,
+ "step": 702
+ },
+ {
+ "epoch": 9.767888307155323,
+ "grad_norm": 0.025329895317554474,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.626345157623291,
+ "step": 703
+ },
+ {
+ "epoch": 9.781849912739965,
+ "grad_norm": 0.027804844081401825,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.621457099914551,
+ "step": 704
+ },
+ {
+ "epoch": 9.795811518324607,
+ "grad_norm": 0.028763437643647194,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.6124114990234375,
+ "step": 705
+ },
+ {
+ "epoch": 9.809773123909249,
+ "grad_norm": 0.030769916251301765,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.604399681091309,
+ "step": 706
+ },
+ {
+ "epoch": 9.823734729493891,
+ "grad_norm": 0.03060949593782425,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.640804290771484,
+ "step": 707
+ },
+ {
+ "epoch": 9.837696335078533,
+ "grad_norm": 0.029832778498530388,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.593376159667969,
+ "step": 708
+ },
+ {
+ "epoch": 9.851657940663177,
+ "grad_norm": 0.02952776849269867,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.619533538818359,
+ "step": 709
+ },
+ {
+ "epoch": 9.86561954624782,
+ "grad_norm": 0.02894212119281292,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.640516757965088,
+ "step": 710
+ },
+ {
+ "epoch": 9.879581151832461,
+ "grad_norm": 0.026435624808073044,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.6342997550964355,
+ "step": 711
+ },
+ {
+ "epoch": 9.893542757417103,
+ "grad_norm": 0.02591666206717491,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.630884170532227,
+ "step": 712
+ },
+ {
+ "epoch": 9.907504363001745,
+ "grad_norm": 0.02702309936285019,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.632085800170898,
+ "step": 713
+ },
+ {
+ "epoch": 9.921465968586388,
+ "grad_norm": 0.026766657829284668,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.6231231689453125,
+ "step": 714
+ },
+ {
+ "epoch": 9.93542757417103,
+ "grad_norm": 0.026666559278964996,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.601619243621826,
+ "step": 715
+ },
+ {
+ "epoch": 9.949389179755672,
+ "grad_norm": 0.026801064610481262,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.641724109649658,
+ "step": 716
+ },
+ {
+ "epoch": 9.963350785340314,
+ "grad_norm": 0.03152798116207123,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.642511367797852,
+ "step": 717
+ },
+ {
+ "epoch": 9.977312390924956,
+ "grad_norm": 0.035406943410634995,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.635779857635498,
+ "step": 718
+ },
+ {
+ "epoch": 9.991273996509598,
+ "grad_norm": 0.037504684180021286,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.629463195800781,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02263174019753933,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 2.876767635345459,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.5942904949188232,
+ "eval_runtime": 60.2086,
+ "eval_samples_per_second": 40.559,
+ "eval_steps_per_second": 0.648,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03707852587103844,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.5765533447265625,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.05454680323600769,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.576363563537598,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.04832114651799202,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.586915969848633,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.050469402223825455,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.5265092849731445,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0622565895318985,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.592833518981934,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0640949010848999,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.5855560302734375,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.06502196192741394,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.591236114501953,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.06167149916291237,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.56747579574585,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.06747374683618546,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.60452938079834,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.07210271060466766,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.561588764190674,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.0733448714017868,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.610766410827637,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.0731528028845787,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.589570045471191,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.0822535902261734,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.631077289581299,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.06921929866075516,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.628293991088867,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.07443586736917496,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.627321243286133,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.07153813540935516,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.641068935394287,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.069196417927742,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.615875720977783,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06483697891235352,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.6165313720703125,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.05482150986790657,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.599890232086182,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.05364019051194191,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.61989164352417,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.053216829895973206,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.564305305480957,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05740506947040558,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.582683563232422,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.05587657541036606,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.61264705657959,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.050252825021743774,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.648050308227539,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.050955042243003845,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.615367889404297,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04903746768832207,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.617378234863281,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04252410680055618,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.609130859375,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.03801053389906883,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.625149250030518,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.03422372043132782,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.606106281280518,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.030507879331707954,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.6125993728637695,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.0340375155210495,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.5690202713012695,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.030507374554872513,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.6042985916137695,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027383938431739807,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.585570812225342,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.028792303055524826,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.566786766052246,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02866457775235176,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.566376686096191,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.028657952323555946,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.581750869750977,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.028227943927049637,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.576788902282715,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.024757765233516693,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.59373664855957,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.027442635968327522,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.570978164672852,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.02590034157037735,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.571286201477051,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.025696570053696632,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.542209625244141,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.025478003546595573,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.5900421142578125,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.02563696913421154,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.544609546661377,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.025326019152998924,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.5808515548706055,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.024502627551555634,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.560579776763916,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.02687518671154976,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.567154884338379,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.025280235335230827,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.560011863708496,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.026266315951943398,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.5758442878723145,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.026058891788125038,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.583309173583984,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.02620471641421318,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.547737121582031,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02626362442970276,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.556955337524414,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.024503158405423164,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.580735206604004,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02580532431602478,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.564000129699707,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.024437161162495613,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.545594215393066,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.026664750650525093,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.56415319442749,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.02499818615615368,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.550689697265625,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.027514580637216568,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.5548248291015625,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.026799403131008148,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.548295021057129,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.02685476280748844,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.5660600662231445,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.026104001328349113,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.556082248687744,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.027479661628603935,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.5580244064331055,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.02638247236609459,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.550073623657227,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.025853842496871948,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.578088760375977,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.02636682614684105,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.58807373046875,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.02684398740530014,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.561017036437988,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.028538964688777924,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.548762321472168,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.028327809646725655,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.548969745635986,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.025131452828645706,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.564047813415527,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.026498543098568916,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.553997039794922,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.030081963166594505,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.5787787437438965,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.03201213851571083,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.566424369812012,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.020518779754638672,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8421616554260254,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.5899484157562256,
+ "eval_runtime": 60.2671,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03692612797021866,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.492715835571289,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.050258126109838486,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.480323791503906,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.04179152473807335,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.484371185302734,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.04160204157233238,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.4905524253845215,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.03974860534071922,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.503803253173828,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.037031810730695724,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.494629859924316,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.038762833923101425,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.495111465454102,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.0412563718855381,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.490113735198975,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.041371624916791916,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.490643501281738,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04151557758450508,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.49881649017334,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04037577286362648,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.492178916931152,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.03676290065050125,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.477460861206055,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.038055796176195145,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.509150981903076,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.041152238845825195,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.47133731842041,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.03728092834353447,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.508072376251221,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.035338856279850006,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.505619049072266,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.037813492119312286,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.509489059448242,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.03789111226797104,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.508646488189697,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.03880828619003296,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.52567720413208,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.03611127287149429,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.498889446258545,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.036903638392686844,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.498601913452148,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03620576485991478,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.490840911865234,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03555903211236,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.506166458129883,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.03441924974322319,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.498001575469971,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.0334339402616024,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.491082191467285,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.031418606638908386,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.527408599853516,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03073921985924244,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.514107704162598,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.029589885845780373,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.48323917388916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.02974501997232437,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.51338529586792,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.029247552156448364,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.504205226898193,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.03132032975554466,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.505674362182617,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03088037669658661,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.477198600769043,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.031054604798555374,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.498721122741699,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03293661028146744,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.527826309204102,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.03181406110525131,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.495999336242676,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.029380083084106445,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.530351638793945,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.029296347871422768,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.492367744445801,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.02958036959171295,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.493037700653076,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03078395500779152,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.47504186630249,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03280661255121231,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.518409729003906,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.03323954716324806,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.460753917694092,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.032254140824079514,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.474285125732422,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.03018898330628872,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.5162353515625,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.02802245132625103,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.538684844970703,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.02821732498705387,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.4850287437438965,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.029189350083470345,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.50770378112793,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.02851196378469467,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.499988555908203,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.02741030789911747,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.499852180480957,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.028708625584840775,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.465211868286133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.028519680723547935,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.502464294433594,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.027846841141581535,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.503248691558838,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030456233769655228,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.490391731262207,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029952052980661392,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.50056266784668,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.02925778552889824,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.49215030670166,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.029679695144295692,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.523841857910156,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.031178541481494904,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.499871730804443,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.031264498829841614,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.50546932220459,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03482109308242798,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.494480133056641,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.03341105580329895,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.532899856567383,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028623664751648903,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.5016279220581055,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028600146993994713,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.511589050292969,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.029127901419997215,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.502058982849121,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.027898237109184265,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.493182182312012,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.028683282434940338,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.519793510437012,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.02887023612856865,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.457289695739746,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.02811296097934246,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.49398946762085,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.028485111892223358,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.478344440460205,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02914245054125786,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.487666130065918,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.028030497953295708,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.518488883972168,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.027973609045147896,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.515271186828613,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.027458522468805313,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.490856647491455,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.01909293793141842,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.822120189666748,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5866912603378296,
+ "eval_runtime": 61.3185,
+ "eval_samples_per_second": 39.825,
+ "eval_steps_per_second": 0.636,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03683992475271225,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.438363075256348,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.05042923241853714,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.443560600280762,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.055914171040058136,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.422282695770264,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.057007886469364166,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.442644119262695,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.05999710038304329,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.44780969619751,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06472347676753998,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.485074996948242,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.06923902779817581,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.457515716552734,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.0710320770740509,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.465476989746094,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.06942574679851532,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.426226615905762,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.06867868453264236,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.438092231750488,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.0688796266913414,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.446664810180664,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.0690719485282898,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.456482887268066,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.06068912148475647,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.4730730056762695,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.05411486327648163,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.4596452713012695,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.056623928248882294,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.418197154998779,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.05634097009897232,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.482892036437988,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.05056057870388031,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.457589149475098,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.04460986703634262,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.451439380645752,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.045022401958703995,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.452855110168457,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.04054102674126625,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.417323112487793,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.03682882338762283,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.465713977813721,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.03528819605708122,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.4502339363098145,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.03568802773952484,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.422399520874023,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.0363088883459568,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.447678089141846,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.03460463136434555,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.437664985656738,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.034751344472169876,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.489151954650879,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.033658601343631744,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.4471635818481445,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.03433239087462425,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.455332279205322,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.032074104994535446,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.435157775878906,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.032168325036764145,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.435118675231934,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.03007678873836994,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.42781925201416,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03094375506043434,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.433105945587158,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.032186444848775864,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.383000373840332,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.030610304325819016,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.473257064819336,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.030257435515522957,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.453041076660156,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.0301323514431715,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.396578788757324,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.030934512615203857,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.445223808288574,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.029907869175076485,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.442287445068359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03163841366767883,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.434650421142578,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.029627414420247078,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.454374313354492,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.031026875600218773,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.413784027099609,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03099491260945797,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.458438873291016,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.02981731668114662,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.438560962677002,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.02927570603787899,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.44493293762207,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.029779717326164246,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.445121765136719,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.02909068390727043,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.4736127853393555,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.029217200353741646,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.434604644775391,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.029872911050915718,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.454591751098633,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.028742222115397453,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.468222141265869,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.030313383787870407,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.45759391784668,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.03014662116765976,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.446331977844238,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029481127858161926,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.478209495544434,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.028485266491770744,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.444445610046387,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.02857700176537037,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.430604934692383,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.029038602486252785,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.43759298324585,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.029576119035482407,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.474579811096191,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.02897879108786583,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.440483093261719,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.029799485579133034,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.4839277267456055,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.030710609629750252,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.466696739196777,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.030270954594016075,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.455197334289551,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.03170182928442955,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.454664707183838,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030684730038046837,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.437619209289551,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.030972573906183243,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.431796073913574,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028914963826537132,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.460618019104004,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.029810264706611633,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.407312393188477,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.029806939885020256,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.466520309448242,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.03020886518061161,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.427349090576172,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030598826706409454,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.427891731262207,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03192012757062912,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.448113918304443,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.030492989346385002,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.450060844421387,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03188254311680794,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.43007230758667,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02081947773694992,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.782536506652832,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5860040187835693,
+ "eval_runtime": 60.6725,
+ "eval_samples_per_second": 40.249,
+ "eval_steps_per_second": 0.643,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.036721356213092804,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.358441352844238,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.056633755564689636,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.361527919769287,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.06175022944808006,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.3827924728393555,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05159056559205055,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.354856491088867,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.04885871335864067,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.359966278076172,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.04407693073153496,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.364612579345703,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.04891050234436989,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.344021320343018,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.04643657058477402,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.366066932678223,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.05241481214761734,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.3390607833862305,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.05837438255548477,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.390261650085449,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.058482322841882706,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.3856201171875,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.06507553160190582,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.414223670959473,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.0575365275144577,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.402576446533203,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05055592209100723,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.389563083648682,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.047167956829071045,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.402750015258789,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.04473405331373215,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.358315467834473,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.042440690100193024,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.392551898956299,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04465719312429428,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.373558044433594,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.04337528720498085,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.37944221496582,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.04322915896773338,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.38493013381958,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.039873406291007996,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.389793395996094,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.0413178876042366,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.427964210510254,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.04567280039191246,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.404172897338867,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04154926910996437,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.391359329223633,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.03774876892566681,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.381457328796387,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03588850423693657,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.392526626586914,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.03757261857390404,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.39273738861084,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03849692642688751,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.409083366394043,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03787294402718544,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.3776044845581055,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.034496866166591644,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.413462162017822,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.03973205015063286,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.359318733215332,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.036192040890455246,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.392844200134277,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.03436632081866264,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.421384811401367,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.035658642649650574,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.386757850646973,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.03434896469116211,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.364978313446045,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03642290085554123,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.385326385498047,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.03427736461162567,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.406352996826172,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.034845270216464996,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.370953559875488,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03390975296497345,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.3799943923950195,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.03134961053729057,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.419156551361084,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.03512260690331459,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.404660701751709,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03319067507982254,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.388535022735596,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03113507106900215,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.403522968292236,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03559132665395737,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.382323265075684,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.03579750657081604,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.37620210647583,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.033675212413072586,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.411443710327148,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.03220067173242569,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.443470001220703,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031610481441020966,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.406038284301758,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.02958764135837555,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.41604471206665,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.0318237729370594,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4253387451171875,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03260074555873871,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.370792388916016,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.03359333053231239,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.397639274597168,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032516930252313614,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.4140777587890625,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.03203396871685982,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.408872604370117,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.03201935440301895,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.402975082397461,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.03330211713910103,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.394018173217773,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.0348181314766407,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.380587100982666,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.03174913674592972,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.372378349304199,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.029864462092518806,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.42018985748291,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.033312153071165085,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.385973930358887,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03203144669532776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.395167827606201,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.031578030437231064,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.3982062339782715,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.030885562300682068,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.400731086730957,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03235667943954468,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.389830589294434,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.032664328813552856,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.401277542114258,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.030354196205735207,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.422520637512207,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.031162403523921967,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.393032073974609,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.032187070697546005,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.464164733886719,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.03445067256689072,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.368991851806641,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03398612141609192,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.418752670288086,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.029730670154094696,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.41429328918457,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022440090775489807,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7315595149993896,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5853753089904785,
+ "eval_runtime": 60.7174,
+ "eval_samples_per_second": 40.219,
+ "eval_steps_per_second": 0.642,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.03754177689552307,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.304028511047363,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.051616180688142776,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.34659481048584,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055258627980947495,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.310445785522461,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.049489911645650864,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3416619300842285,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05377533286809921,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.3143310546875,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.05443088337779045,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.320600986480713,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.050222691148519516,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.334789752960205,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.051904670894145966,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.316756248474121,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.05376426503062248,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.274116516113281,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.053313180804252625,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.321177005767822,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.058234598487615585,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.3169264793396,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05730101093649864,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.307013511657715,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.058140840381383896,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.332961082458496,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05654815956950188,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.315462112426758,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.05129384249448776,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.338476181030273,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.047652702778577805,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.321232795715332,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.050912003964185715,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.318971633911133,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.0514022521674633,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.3334150314331055,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.05457981303334236,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.325802803039551,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04678885266184807,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.307281494140625,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.05048161372542381,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.299831390380859,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.05334261804819107,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.313231468200684,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.05055771768093109,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.36859130859375,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04955561086535454,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.301530361175537,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.04548008739948273,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.339517593383789,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.046585116535425186,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.333036422729492,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.0503716766834259,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.364710807800293,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.040569406002759933,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.369089126586914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.03754054754972458,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.3349432945251465,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03775954991579056,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.333134174346924,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.04066266492009163,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.354358673095703,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.040490977466106415,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.360981464385986,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.0426657609641552,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.357726097106934,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.0415605828166008,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.363457679748535,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.04020509123802185,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.372447967529297,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.04101516678929329,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.292452812194824,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.03546427935361862,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.33892297744751,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.03661105036735535,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.329778671264648,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03565476834774017,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.351012229919434,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.03591311350464821,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.376980781555176,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.03665630891919136,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.354767799377441,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.0371236577630043,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.374916076660156,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.03464849665760994,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.3713059425354,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03345152735710144,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.350670337677002,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.03357568010687828,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.3573713302612305,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03503486514091492,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.362038612365723,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.03455005958676338,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.3263773918151855,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.03173859417438507,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.351345062255859,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03301836550235748,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.398364067077637,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.033733855932950974,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.335268020629883,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034080155193805695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.371492385864258,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.034396376460790634,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.382076740264893,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.03303401917219162,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.36623477935791,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03297369182109833,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3148393630981445,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.034819211810827255,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.358110427856445,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.033389780670404434,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.367427825927734,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.032363444566726685,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.3813676834106445,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03509577363729477,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.386520862579346,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.034685730934143066,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.33738374710083,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.034755002707242966,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.360980987548828,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03461688384413719,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.396650314331055,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03392560034990311,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.356130123138428,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.03342147544026375,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.359145164489746,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.032852694392204285,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.3796162605285645,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03472859039902687,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.361110687255859,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03277098760008812,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.361811637878418,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.03341882303357124,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.369524955749512,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.032347049564123154,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.369096755981445,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03161193057894707,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.339844703674316,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.03322301432490349,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.376675605773926,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03419092297554016,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.346864223480225,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02461826801300049,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7236275672912598,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.5859372019767761,
+ "eval_runtime": 59.0139,
+ "eval_samples_per_second": 41.38,
+ "eval_steps_per_second": 0.661,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013961605584642,
+ "grad_norm": 0.039559684693813324,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 4.289841651916504,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027923211169284,
+ "grad_norm": 0.05554218962788582,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 4.255925178527832,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041884816753926,
+ "grad_norm": 0.05330996215343475,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 4.272125244140625,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055846422338568,
+ "grad_norm": 0.053247954696416855,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 4.290765285491943,
+ "step": 1084
+ },
+ {
+ "epoch": 15.06980802792321,
+ "grad_norm": 0.05471150949597359,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 4.271744728088379,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083769633507853,
+ "grad_norm": 0.05976719409227371,
+ "learning_rate": 0.000491285979372924,
+ "loss": 4.28203821182251,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097731239092496,
+ "grad_norm": 0.05369880795478821,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 4.275313377380371,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111692844677139,
+ "grad_norm": 0.052529267966747284,
+ "learning_rate": 0.000490873338088198,
+ "loss": 4.25678825378418,
+ "step": 1088
+ },
+ {
+ "epoch": 15.12565445026178,
+ "grad_norm": 0.05184522643685341,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 4.260510444641113,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139616055846423,
+ "grad_norm": 0.051606856286525726,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 4.283233642578125,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153577661431065,
+ "grad_norm": 0.05318903923034668,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 4.286595344543457,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167539267015707,
+ "grad_norm": 0.058525051921606064,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 4.282374382019043,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181500872600349,
+ "grad_norm": 0.054776228964328766,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 4.2913126945495605,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195462478184991,
+ "grad_norm": 0.04907633364200592,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 4.297187805175781,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209424083769633,
+ "grad_norm": 0.05163136497139931,
+ "learning_rate": 0.000489424334303338,
+ "loss": 4.2968645095825195,
+ "step": 1095
+ },
+ {
+ "epoch": 15.223385689354275,
+ "grad_norm": 0.053157489746809006,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 4.285839080810547,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237347294938917,
+ "grad_norm": 0.05251535773277283,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 4.311246871948242,
+ "step": 1097
+ },
+ {
+ "epoch": 15.25130890052356,
+ "grad_norm": 0.048785097897052765,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 4.300364971160889,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265270506108202,
+ "grad_norm": 0.050393108278512955,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 4.299890995025635,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279232111692846,
+ "grad_norm": 0.04894037917256355,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 4.312449932098389,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293193717277488,
+ "grad_norm": 0.05274952948093414,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 4.2889814376831055,
+ "step": 1101
+ },
+ {
+ "epoch": 15.30715532286213,
+ "grad_norm": 0.05946570262312889,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 4.2731170654296875,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321116928446772,
+ "grad_norm": 0.054245397448539734,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 4.272463798522949,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335078534031414,
+ "grad_norm": 0.042092181742191315,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 4.299953937530518,
+ "step": 1104
+ },
+ {
+ "epoch": 15.349040139616056,
+ "grad_norm": 0.040222618728876114,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 4.292524814605713,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363001745200698,
+ "grad_norm": 0.04298696666955948,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 4.31342887878418,
+ "step": 1106
+ },
+ {
+ "epoch": 15.37696335078534,
+ "grad_norm": 0.04271050915122032,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 4.277663707733154,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390924956369982,
+ "grad_norm": 0.042364370077848434,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 4.311102390289307,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404886561954624,
+ "grad_norm": 0.04343319311738014,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 4.311610221862793,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418848167539267,
+ "grad_norm": 0.041321512311697006,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 4.29677677154541,
+ "step": 1110
+ },
+ {
+ "epoch": 15.432809773123909,
+ "grad_norm": 0.04107942432165146,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 4.3163018226623535,
+ "step": 1111
+ },
+ {
+ "epoch": 15.44677137870855,
+ "grad_norm": 0.03844515234231949,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 4.311301231384277,
+ "step": 1112
+ },
+ {
+ "epoch": 15.460732984293193,
+ "grad_norm": 0.03731007128953934,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 4.31325626373291,
+ "step": 1113
+ },
+ {
+ "epoch": 15.474694589877837,
+ "grad_norm": 0.03807826340198517,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 4.296201705932617,
+ "step": 1114
+ },
+ {
+ "epoch": 15.488656195462479,
+ "grad_norm": 0.040137600153684616,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 4.320395469665527,
+ "step": 1115
+ },
+ {
+ "epoch": 15.502617801047121,
+ "grad_norm": 0.038484394550323486,
+ "learning_rate": 0.000485033394934934,
+ "loss": 4.313858985900879,
+ "step": 1116
+ },
+ {
+ "epoch": 15.516579406631763,
+ "grad_norm": 0.03603338822722435,
+ "learning_rate": 0.000484822676912974,
+ "loss": 4.297666549682617,
+ "step": 1117
+ },
+ {
+ "epoch": 15.530541012216405,
+ "grad_norm": 0.038647182285785675,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 4.305002689361572,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544502617801047,
+ "grad_norm": 0.03907139599323273,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 4.329583644866943,
+ "step": 1119
+ },
+ {
+ "epoch": 15.55846422338569,
+ "grad_norm": 0.03621622920036316,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 4.290896415710449,
+ "step": 1120
+ },
+ {
+ "epoch": 15.572425828970331,
+ "grad_norm": 0.037496887147426605,
+ "learning_rate": 0.000483978341612154,
+ "loss": 4.33336877822876,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586387434554974,
+ "grad_norm": 0.03778311610221863,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 4.285354137420654,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600349040139616,
+ "grad_norm": 0.03782997280359268,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 4.29697847366333,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614310645724258,
+ "grad_norm": 0.03665657714009285,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 4.320103168487549,
+ "step": 1124
+ },
+ {
+ "epoch": 15.6282722513089,
+ "grad_norm": 0.03796127438545227,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 4.317153453826904,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642233856893542,
+ "grad_norm": 0.036425214260816574,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 4.293642520904541,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656195462478184,
+ "grad_norm": 0.03629784658551216,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 4.291100025177002,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670157068062828,
+ "grad_norm": 0.03674762696027756,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 4.295144557952881,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68411867364747,
+ "grad_norm": 0.036291178315877914,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 4.313133239746094,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698080279232112,
+ "grad_norm": 0.03817860782146454,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 4.294533729553223,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712041884816754,
+ "grad_norm": 0.036819685250520706,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 4.299263954162598,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726003490401396,
+ "grad_norm": 0.035103823989629745,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 4.287391185760498,
+ "step": 1132
+ },
+ {
+ "epoch": 15.739965095986038,
+ "grad_norm": 0.035701096057891846,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 4.318216323852539,
+ "step": 1133
+ },
+ {
+ "epoch": 15.75392670157068,
+ "grad_norm": 0.03639856353402138,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 4.301465034484863,
+ "step": 1134
+ },
+ {
+ "epoch": 15.767888307155323,
+ "grad_norm": 0.03593013063073158,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 4.311532974243164,
+ "step": 1135
+ },
+ {
+ "epoch": 15.781849912739965,
+ "grad_norm": 0.03564770147204399,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 4.3400726318359375,
+ "step": 1136
+ },
+ {
+ "epoch": 15.795811518324607,
+ "grad_norm": 0.03598494827747345,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 4.30976676940918,
+ "step": 1137
+ },
+ {
+ "epoch": 15.809773123909249,
+ "grad_norm": 0.034391652792692184,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 4.328464508056641,
+ "step": 1138
+ },
+ {
+ "epoch": 15.823734729493891,
+ "grad_norm": 0.03445672616362572,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 4.310579299926758,
+ "step": 1139
+ },
+ {
+ "epoch": 15.837696335078533,
+ "grad_norm": 0.03643282502889633,
+ "learning_rate": 0.000479936075927473,
+ "loss": 4.320905685424805,
+ "step": 1140
+ },
+ {
+ "epoch": 15.851657940663177,
+ "grad_norm": 0.03489850088953972,
+ "learning_rate": 0.000479721889242305,
+ "loss": 4.317229270935059,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86561954624782,
+ "grad_norm": 0.03609171137213707,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 4.309203147888184,
+ "step": 1142
+ },
+ {
+ "epoch": 15.879581151832461,
+ "grad_norm": 0.03705338388681412,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 4.287387371063232,
+ "step": 1143
+ },
+ {
+ "epoch": 15.893542757417103,
+ "grad_norm": 0.03697717934846878,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 4.302860260009766,
+ "step": 1144
+ },
+ {
+ "epoch": 15.907504363001745,
+ "grad_norm": 0.035338208079338074,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 4.3152265548706055,
+ "step": 1145
+ },
+ {
+ "epoch": 15.921465968586388,
+ "grad_norm": 0.03548634052276611,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 4.327335357666016,
+ "step": 1146
+ },
+ {
+ "epoch": 15.93542757417103,
+ "grad_norm": 0.035512328147888184,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 4.3067803382873535,
+ "step": 1147
+ },
+ {
+ "epoch": 15.949389179755672,
+ "grad_norm": 0.03799289092421532,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 4.320701599121094,
+ "step": 1148
+ },
+ {
+ "epoch": 15.963350785340314,
+ "grad_norm": 0.03866339474916458,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 4.321087837219238,
+ "step": 1149
+ },
+ {
+ "epoch": 15.977312390924956,
+ "grad_norm": 0.036002084612846375,
+ "learning_rate": 0.000477787822648007,
+ "loss": 4.287045478820801,
+ "step": 1150
+ },
+ {
+ "epoch": 15.991273996509598,
+ "grad_norm": 0.03433441370725632,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 4.2562665939331055,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.02580358274281025,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 2.67783260345459,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.5870338082313538,
+ "eval_runtime": 60.0469,
+ "eval_samples_per_second": 40.668,
+ "eval_steps_per_second": 0.649,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013961605584644,
+ "grad_norm": 0.04105810821056366,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 4.224295616149902,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027923211169284,
+ "grad_norm": 0.049538832157850266,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 4.222184181213379,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041884816753928,
+ "grad_norm": 0.05089569091796875,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 4.2067060470581055,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05584642233857,
+ "grad_norm": 0.04600589722394943,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 4.238527774810791,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069808027923212,
+ "grad_norm": 0.043311409652233124,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 4.242960453033447,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083769633507853,
+ "grad_norm": 0.047897014766931534,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 4.199069023132324,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097731239092496,
+ "grad_norm": 0.05633604899048805,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 4.211665153503418,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111692844677137,
+ "grad_norm": 0.056110311299562454,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 4.20163631439209,
+ "step": 1160
+ },
+ {
+ "epoch": 16.12565445026178,
+ "grad_norm": 0.05426819995045662,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 4.21889591217041,
+ "step": 1161
+ },
+ {
+ "epoch": 16.13961605584642,
+ "grad_norm": 0.05489835515618324,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 4.220406532287598,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153577661431065,
+ "grad_norm": 0.053941354155540466,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 4.231909275054932,
+ "step": 1163
+ },
+ {
+ "epoch": 16.167539267015705,
+ "grad_norm": 0.058548808097839355,
+ "learning_rate": 0.000474756648870413,
+ "loss": 4.237003326416016,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18150087260035,
+ "grad_norm": 0.05370597913861275,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 4.240085124969482,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195462478184993,
+ "grad_norm": 0.05684991553425789,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 4.203842639923096,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209424083769633,
+ "grad_norm": 0.060745254158973694,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 4.247998237609863,
+ "step": 1167
+ },
+ {
+ "epoch": 16.223385689354277,
+ "grad_norm": 0.058296725153923035,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 4.226161003112793,
+ "step": 1168
+ },
+ {
+ "epoch": 16.237347294938917,
+ "grad_norm": 0.05310584232211113,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 4.229870796203613,
+ "step": 1169
+ },
+ {
+ "epoch": 16.25130890052356,
+ "grad_norm": 0.04710390046238899,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 4.240084648132324,
+ "step": 1170
+ },
+ {
+ "epoch": 16.2652705061082,
+ "grad_norm": 0.05003240332007408,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 4.240972518920898,
+ "step": 1171
+ },
+ {
+ "epoch": 16.279232111692846,
+ "grad_norm": 0.05552851781249046,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 4.248024940490723,
+ "step": 1172
+ },
+ {
+ "epoch": 16.293193717277486,
+ "grad_norm": 0.049967192113399506,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 4.279229164123535,
+ "step": 1173
+ },
+ {
+ "epoch": 16.30715532286213,
+ "grad_norm": 0.04883681237697601,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 4.269807815551758,
+ "step": 1174
+ },
+ {
+ "epoch": 16.32111692844677,
+ "grad_norm": 0.04931480810046196,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 4.232419490814209,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335078534031414,
+ "grad_norm": 0.048746511340141296,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 4.267911911010742,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349040139616054,
+ "grad_norm": 0.04678189381957054,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 4.262979507446289,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363001745200698,
+ "grad_norm": 0.048893529921770096,
+ "learning_rate": 0.000471698352726896,
+ "loss": 4.254408359527588,
+ "step": 1178
+ },
+ {
+ "epoch": 16.376963350785342,
+ "grad_norm": 0.045027732849121094,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 4.2445902824401855,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390924956369982,
+ "grad_norm": 0.04324870929121971,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 4.2685136795043945,
+ "step": 1180
+ },
+ {
+ "epoch": 16.404886561954626,
+ "grad_norm": 0.0470522940158844,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 4.237505912780762,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418848167539267,
+ "grad_norm": 0.04896962642669678,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 4.2523345947265625,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43280977312391,
+ "grad_norm": 0.04562787711620331,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 4.2192792892456055,
+ "step": 1183
+ },
+ {
+ "epoch": 16.44677137870855,
+ "grad_norm": 0.04414575919508934,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 4.250238418579102,
+ "step": 1184
+ },
+ {
+ "epoch": 16.460732984293195,
+ "grad_norm": 0.04363945871591568,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 4.256250381469727,
+ "step": 1185
+ },
+ {
+ "epoch": 16.474694589877835,
+ "grad_norm": 0.04159797728061676,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 4.274903297424316,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48865619546248,
+ "grad_norm": 0.04248544201254845,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 4.278222560882568,
+ "step": 1187
+ },
+ {
+ "epoch": 16.50261780104712,
+ "grad_norm": 0.04379890486598015,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 4.267505645751953,
+ "step": 1188
+ },
+ {
+ "epoch": 16.516579406631763,
+ "grad_norm": 0.04319411888718605,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 4.2607269287109375,
+ "step": 1189
+ },
+ {
+ "epoch": 16.530541012216403,
+ "grad_norm": 0.04143810644745827,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 4.267977714538574,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544502617801047,
+ "grad_norm": 0.03914840891957283,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 4.260498523712158,
+ "step": 1191
+ },
+ {
+ "epoch": 16.55846422338569,
+ "grad_norm": 0.04107543081045151,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 4.2550482749938965,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57242582897033,
+ "grad_norm": 0.041316065937280655,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 4.3075456619262695,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586387434554975,
+ "grad_norm": 0.04069342464208603,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 4.234392166137695,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600349040139616,
+ "grad_norm": 0.03816857561469078,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 4.251038551330566,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61431064572426,
+ "grad_norm": 0.04095374792814255,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 4.284790992736816,
+ "step": 1196
+ },
+ {
+ "epoch": 16.6282722513089,
+ "grad_norm": 0.03788160905241966,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 4.255716323852539,
+ "step": 1197
+ },
+ {
+ "epoch": 16.642233856893544,
+ "grad_norm": 0.03784061595797539,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 4.269606590270996,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656195462478184,
+ "grad_norm": 0.04087168723344803,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 4.231165885925293,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670157068062828,
+ "grad_norm": 0.040546853095293045,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 4.288422584533691,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68411867364747,
+ "grad_norm": 0.03777262568473816,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 4.277492523193359,
+ "step": 1201
+ },
+ {
+ "epoch": 16.698080279232112,
+ "grad_norm": 0.039857733994722366,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 4.239383697509766,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712041884816752,
+ "grad_norm": 0.04387792944908142,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 4.300265312194824,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726003490401396,
+ "grad_norm": 0.04214833676815033,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 4.255545616149902,
+ "step": 1204
+ },
+ {
+ "epoch": 16.739965095986037,
+ "grad_norm": 0.04290056228637695,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 4.275530815124512,
+ "step": 1205
+ },
+ {
+ "epoch": 16.75392670157068,
+ "grad_norm": 0.03975554555654526,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 4.282685279846191,
+ "step": 1206
+ },
+ {
+ "epoch": 16.767888307155324,
+ "grad_norm": 0.039464011788368225,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 4.285504341125488,
+ "step": 1207
+ },
+ {
+ "epoch": 16.781849912739965,
+ "grad_norm": 0.039714012295007706,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 4.273477554321289,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79581151832461,
+ "grad_norm": 0.03936111181974411,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 4.304205894470215,
+ "step": 1209
+ },
+ {
+ "epoch": 16.80977312390925,
+ "grad_norm": 0.03782064467668533,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 4.307382583618164,
+ "step": 1210
+ },
+ {
+ "epoch": 16.823734729493893,
+ "grad_norm": 0.040340717881917953,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 4.308620452880859,
+ "step": 1211
+ },
+ {
+ "epoch": 16.837696335078533,
+ "grad_norm": 0.0403442308306694,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 4.248626708984375,
+ "step": 1212
+ },
+ {
+ "epoch": 16.851657940663177,
+ "grad_norm": 0.038351960480213165,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 4.261248588562012,
+ "step": 1213
+ },
+ {
+ "epoch": 16.865619546247817,
+ "grad_norm": 0.03965333476662636,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 4.278291702270508,
+ "step": 1214
+ },
+ {
+ "epoch": 16.87958115183246,
+ "grad_norm": 0.040909092873334885,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 4.241847991943359,
+ "step": 1215
+ },
+ {
+ "epoch": 16.8935427574171,
+ "grad_norm": 0.038684047758579254,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 4.281346321105957,
+ "step": 1216
+ },
+ {
+ "epoch": 16.907504363001745,
+ "grad_norm": 0.03698103874921799,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 4.3085432052612305,
+ "step": 1217
+ },
+ {
+ "epoch": 16.921465968586386,
+ "grad_norm": 0.03908613696694374,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 4.24220085144043,
+ "step": 1218
+ },
+ {
+ "epoch": 16.93542757417103,
+ "grad_norm": 0.03873388469219208,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 4.282486915588379,
+ "step": 1219
+ },
+ {
+ "epoch": 16.949389179755673,
+ "grad_norm": 0.03709883987903595,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 4.279009819030762,
+ "step": 1220
+ },
+ {
+ "epoch": 16.963350785340314,
+ "grad_norm": 0.03937513753771782,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 4.273211479187012,
+ "step": 1221
+ },
+ {
+ "epoch": 16.977312390924958,
+ "grad_norm": 0.038478728383779526,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 4.281154155731201,
+ "step": 1222
+ },
+ {
+ "epoch": 16.991273996509598,
+ "grad_norm": 0.03913680836558342,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 4.269218444824219,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.02641749382019043,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 2.6905035972595215,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.5879416465759277,
+ "eval_runtime": 60.2807,
+ "eval_samples_per_second": 40.511,
+ "eval_steps_per_second": 0.647,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013961605584644,
+ "grad_norm": 0.048571839928627014,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 4.2008538246154785,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027923211169284,
+ "grad_norm": 0.0640900656580925,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 4.180826663970947,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041884816753928,
+ "grad_norm": 0.06458134204149246,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 4.199562072753906,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05584642233857,
+ "grad_norm": 0.0563119500875473,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 4.171557426452637,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069808027923212,
+ "grad_norm": 0.050914082676172256,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 4.183539390563965,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083769633507853,
+ "grad_norm": 0.053884491324424744,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 4.210450172424316,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097731239092496,
+ "grad_norm": 0.05301345884799957,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 4.1553192138671875,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111692844677137,
+ "grad_norm": 0.051138825714588165,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 4.158510208129883,
+ "step": 1232
+ },
+ {
+ "epoch": 17.12565445026178,
+ "grad_norm": 0.05355675891041756,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 4.167689323425293,
+ "step": 1233
+ },
+ {
+ "epoch": 17.13961605584642,
+ "grad_norm": 0.06062275916337967,
+ "learning_rate": 0.000459203665939781,
+ "loss": 4.170838356018066,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153577661431065,
+ "grad_norm": 0.058666616678237915,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 4.187441825866699,
+ "step": 1235
+ },
+ {
+ "epoch": 17.167539267015705,
+ "grad_norm": 0.057622168213129044,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 4.190746784210205,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18150087260035,
+ "grad_norm": 0.069090336561203,
+ "learning_rate": 0.000458522886595554,
+ "loss": 4.224946022033691,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195462478184993,
+ "grad_norm": 0.06552847474813461,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 4.204067230224609,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209424083769633,
+ "grad_norm": 0.06602177768945694,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 4.189286708831787,
+ "step": 1239
+ },
+ {
+ "epoch": 17.223385689354277,
+ "grad_norm": 0.06751416623592377,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 4.208927154541016,
+ "step": 1240
+ },
+ {
+ "epoch": 17.237347294938917,
+ "grad_norm": 0.0706794261932373,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 4.212923049926758,
+ "step": 1241
+ },
+ {
+ "epoch": 17.25130890052356,
+ "grad_norm": 0.05346836522221565,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 4.2221784591674805,
+ "step": 1242
+ },
+ {
+ "epoch": 17.2652705061082,
+ "grad_norm": 0.05395734682679176,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 4.196428298950195,
+ "step": 1243
+ },
+ {
+ "epoch": 17.279232111692846,
+ "grad_norm": 0.05496550351381302,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 4.208031177520752,
+ "step": 1244
+ },
+ {
+ "epoch": 17.293193717277486,
+ "grad_norm": 0.05320895463228226,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 4.232810974121094,
+ "step": 1245
+ },
+ {
+ "epoch": 17.30715532286213,
+ "grad_norm": 0.05064759403467178,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 4.207826614379883,
+ "step": 1246
+ },
+ {
+ "epoch": 17.32111692844677,
+ "grad_norm": 0.04287239536643028,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 4.195648670196533,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335078534031414,
+ "grad_norm": 0.04527301341295242,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 4.19954776763916,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349040139616054,
+ "grad_norm": 0.04421805590391159,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 4.248671054840088,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363001745200698,
+ "grad_norm": 0.04487175494432449,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 4.2072954177856445,
+ "step": 1250
+ },
+ {
+ "epoch": 17.376963350785342,
+ "grad_norm": 0.04415794834494591,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 4.222837448120117,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390924956369982,
+ "grad_norm": 0.04474013298749924,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 4.213104248046875,
+ "step": 1252
+ },
+ {
+ "epoch": 17.404886561954626,
+ "grad_norm": 0.042631037533283234,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 4.209726333618164,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418848167539267,
+ "grad_norm": 0.04424271360039711,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 4.202699661254883,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43280977312391,
+ "grad_norm": 0.04151812940835953,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 4.190732955932617,
+ "step": 1255
+ },
+ {
+ "epoch": 17.44677137870855,
+ "grad_norm": 0.041588131338357925,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 4.222360610961914,
+ "step": 1256
+ },
+ {
+ "epoch": 17.460732984293195,
+ "grad_norm": 0.042378414422273636,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 4.198424339294434,
+ "step": 1257
+ },
+ {
+ "epoch": 17.474694589877835,
+ "grad_norm": 0.041907574981451035,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 4.224804401397705,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48865619546248,
+ "grad_norm": 0.043774355202913284,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 4.2276611328125,
+ "step": 1259
+ },
+ {
+ "epoch": 17.50261780104712,
+ "grad_norm": 0.04238573834300041,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 4.228109359741211,
+ "step": 1260
+ },
+ {
+ "epoch": 17.516579406631763,
+ "grad_norm": 0.045351386070251465,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 4.208807468414307,
+ "step": 1261
+ },
+ {
+ "epoch": 17.530541012216403,
+ "grad_norm": 0.04300729185342789,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 4.257306098937988,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544502617801047,
+ "grad_norm": 0.041442908346652985,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 4.208004951477051,
+ "step": 1263
+ },
+ {
+ "epoch": 17.55846422338569,
+ "grad_norm": 0.04337337985634804,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 4.23393440246582,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57242582897033,
+ "grad_norm": 0.04335462674498558,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 4.228087425231934,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586387434554975,
+ "grad_norm": 0.03987223282456398,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 4.213971138000488,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600349040139616,
+ "grad_norm": 0.041760094463825226,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 4.248366355895996,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61431064572426,
+ "grad_norm": 0.04420867934823036,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 4.278865814208984,
+ "step": 1268
+ },
+ {
+ "epoch": 17.6282722513089,
+ "grad_norm": 0.04244982823729515,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 4.223642349243164,
+ "step": 1269
+ },
+ {
+ "epoch": 17.642233856893544,
+ "grad_norm": 0.04280664771795273,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 4.231023788452148,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656195462478184,
+ "grad_norm": 0.04294133558869362,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 4.222380638122559,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670157068062828,
+ "grad_norm": 0.04179249703884125,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 4.215641975402832,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68411867364747,
+ "grad_norm": 0.04038798063993454,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 4.228580474853516,
+ "step": 1273
+ },
+ {
+ "epoch": 17.698080279232112,
+ "grad_norm": 0.04211528226733208,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 4.245234489440918,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712041884816752,
+ "grad_norm": 0.04102989658713341,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 4.218495845794678,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726003490401396,
+ "grad_norm": 0.04038692265748978,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 4.2466511726379395,
+ "step": 1276
+ },
+ {
+ "epoch": 17.739965095986037,
+ "grad_norm": 0.03958791121840477,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 4.237762451171875,
+ "step": 1277
+ },
+ {
+ "epoch": 17.75392670157068,
+ "grad_norm": 0.04129907861351967,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 4.238739013671875,
+ "step": 1278
+ },
+ {
+ "epoch": 17.767888307155324,
+ "grad_norm": 0.042581796646118164,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 4.235558032989502,
+ "step": 1279
+ },
+ {
+ "epoch": 17.781849912739965,
+ "grad_norm": 0.04339978098869324,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 4.226593971252441,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79581151832461,
+ "grad_norm": 0.03808023780584335,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 4.238213062286377,
+ "step": 1281
+ },
+ {
+ "epoch": 17.80977312390925,
+ "grad_norm": 0.04022229090332985,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 4.214224815368652,
+ "step": 1282
+ },
+ {
+ "epoch": 17.823734729493893,
+ "grad_norm": 0.04379156231880188,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 4.253068447113037,
+ "step": 1283
+ },
+ {
+ "epoch": 17.837696335078533,
+ "grad_norm": 0.04232964664697647,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 4.249903202056885,
+ "step": 1284
+ },
+ {
+ "epoch": 17.851657940663177,
+ "grad_norm": 0.040980905294418335,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 4.252640724182129,
+ "step": 1285
+ },
+ {
+ "epoch": 17.865619546247817,
+ "grad_norm": 0.04015526548027992,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 4.234862327575684,
+ "step": 1286
+ },
+ {
+ "epoch": 17.87958115183246,
+ "grad_norm": 0.04273828864097595,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 4.216028213500977,
+ "step": 1287
+ },
+ {
+ "epoch": 17.8935427574171,
+ "grad_norm": 0.04210563376545906,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 4.23359489440918,
+ "step": 1288
+ },
+ {
+ "epoch": 17.907504363001745,
+ "grad_norm": 0.038766685873270035,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 4.220360279083252,
+ "step": 1289
+ },
+ {
+ "epoch": 17.921465968586386,
+ "grad_norm": 0.03921609744429588,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 4.253805160522461,
+ "step": 1290
+ },
+ {
+ "epoch": 17.93542757417103,
+ "grad_norm": 0.04071544483304024,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 4.211221694946289,
+ "step": 1291
+ },
+ {
+ "epoch": 17.949389179755673,
+ "grad_norm": 0.03996877372264862,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 4.241739273071289,
+ "step": 1292
+ },
+ {
+ "epoch": 17.963350785340314,
+ "grad_norm": 0.03878853842616081,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 4.237001419067383,
+ "step": 1293
+ },
+ {
+ "epoch": 17.977312390924958,
+ "grad_norm": 0.03889666125178337,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 4.244929790496826,
+ "step": 1294
+ },
+ {
+ "epoch": 17.991273996509598,
+ "grad_norm": 0.03876441717147827,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 4.236083030700684,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.027286000549793243,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 2.636739730834961,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.5896694660186768,
+ "eval_runtime": 60.4742,
+ "eval_samples_per_second": 40.381,
+ "eval_steps_per_second": 0.645,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013961605584644,
+ "grad_norm": 0.04279542714357376,
+ "learning_rate": 0.000444677103403854,
+ "loss": 4.136401653289795,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027923211169284,
+ "grad_norm": 0.051685452461242676,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 4.159584045410156,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041884816753928,
+ "grad_norm": 0.05411457270383835,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 4.160367965698242,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05584642233857,
+ "grad_norm": 0.048574138432741165,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 4.134189128875732,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069808027923212,
+ "grad_norm": 0.04528478533029556,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 4.1022725105285645,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083769633507853,
+ "grad_norm": 0.04474090039730072,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 4.16152286529541,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097731239092496,
+ "grad_norm": 0.04882791265845299,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 4.1553144454956055,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111692844677137,
+ "grad_norm": 0.04831654578447342,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 4.147022724151611,
+ "step": 1304
+ },
+ {
+ "epoch": 18.12565445026178,
+ "grad_norm": 0.047365427017211914,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 4.139405250549316,
+ "step": 1305
+ },
+ {
+ "epoch": 18.13961605584642,
+ "grad_norm": 0.0514012910425663,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 4.115568161010742,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153577661431065,
+ "grad_norm": 0.05225161090493202,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 4.166429042816162,
+ "step": 1307
+ },
+ {
+ "epoch": 18.167539267015705,
+ "grad_norm": 0.05575249716639519,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 4.1414289474487305,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18150087260035,
+ "grad_norm": 0.05405383184552193,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 4.138680934906006,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195462478184993,
+ "grad_norm": 0.05185031518340111,
+ "learning_rate": 0.000441621628895411,
+ "loss": 4.159243583679199,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209424083769633,
+ "grad_norm": 0.05070916935801506,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 4.1844892501831055,
+ "step": 1311
+ },
+ {
+ "epoch": 18.223385689354277,
+ "grad_norm": 0.05037331581115723,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 4.132781028747559,
+ "step": 1312
+ },
+ {
+ "epoch": 18.237347294938917,
+ "grad_norm": 0.05082111805677414,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 4.143031120300293,
+ "step": 1313
+ },
+ {
+ "epoch": 18.25130890052356,
+ "grad_norm": 0.04742169380187988,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 4.137580871582031,
+ "step": 1314
+ },
+ {
+ "epoch": 18.2652705061082,
+ "grad_norm": 0.049042731523513794,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 4.159721374511719,
+ "step": 1315
+ },
+ {
+ "epoch": 18.279232111692846,
+ "grad_norm": 0.045649368315935135,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 4.144008636474609,
+ "step": 1316
+ },
+ {
+ "epoch": 18.293193717277486,
+ "grad_norm": 0.0473787859082222,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 4.164243698120117,
+ "step": 1317
+ },
+ {
+ "epoch": 18.30715532286213,
+ "grad_norm": 0.047626055777072906,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 4.164441108703613,
+ "step": 1318
+ },
+ {
+ "epoch": 18.32111692844677,
+ "grad_norm": 0.04738573357462883,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 4.180338382720947,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335078534031414,
+ "grad_norm": 0.04712498560547829,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 4.146285533905029,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349040139616054,
+ "grad_norm": 0.045505691319704056,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 4.162572860717773,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363001745200698,
+ "grad_norm": 0.0489528588950634,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 4.160312175750732,
+ "step": 1322
+ },
+ {
+ "epoch": 18.376963350785342,
+ "grad_norm": 0.047517575323581696,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 4.164361953735352,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390924956369982,
+ "grad_norm": 0.044214535504579544,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 4.182154655456543,
+ "step": 1324
+ },
+ {
+ "epoch": 18.404886561954626,
+ "grad_norm": 0.043775733560323715,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 4.130522727966309,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418848167539267,
+ "grad_norm": 0.044364530593156815,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 4.151247024536133,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43280977312391,
+ "grad_norm": 0.046320103108882904,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 4.204159736633301,
+ "step": 1327
+ },
+ {
+ "epoch": 18.44677137870855,
+ "grad_norm": 0.04802565276622772,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 4.182550430297852,
+ "step": 1328
+ },
+ {
+ "epoch": 18.460732984293195,
+ "grad_norm": 0.049340926110744476,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 4.163658142089844,
+ "step": 1329
+ },
+ {
+ "epoch": 18.474694589877835,
+ "grad_norm": 0.046701837331056595,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 4.170445442199707,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48865619546248,
+ "grad_norm": 0.04746273159980774,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 4.220232963562012,
+ "step": 1331
+ },
+ {
+ "epoch": 18.50261780104712,
+ "grad_norm": 0.04644489288330078,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 4.190027236938477,
+ "step": 1332
+ },
+ {
+ "epoch": 18.516579406631763,
+ "grad_norm": 0.04394262284040451,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 4.159570693969727,
+ "step": 1333
+ },
+ {
+ "epoch": 18.530541012216403,
+ "grad_norm": 0.04609063267707825,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 4.1657586097717285,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544502617801047,
+ "grad_norm": 0.0450211800634861,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 4.184865474700928,
+ "step": 1335
+ },
+ {
+ "epoch": 18.55846422338569,
+ "grad_norm": 0.0475253090262413,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 4.185857772827148,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57242582897033,
+ "grad_norm": 0.04332203418016434,
+ "learning_rate": 0.000435215814386134,
+ "loss": 4.160165309906006,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586387434554975,
+ "grad_norm": 0.0427570678293705,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 4.1727495193481445,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600349040139616,
+ "grad_norm": 0.04651174694299698,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 4.204797744750977,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61431064572426,
+ "grad_norm": 0.04324701428413391,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 4.186931610107422,
+ "step": 1340
+ },
+ {
+ "epoch": 18.6282722513089,
+ "grad_norm": 0.044644393026828766,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 4.209554195404053,
+ "step": 1341
+ },
+ {
+ "epoch": 18.642233856893544,
+ "grad_norm": 0.042982228100299835,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 4.205570220947266,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656195462478184,
+ "grad_norm": 0.04385090991854668,
+ "learning_rate": 0.000433781621332752,
+ "loss": 4.177936553955078,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670157068062828,
+ "grad_norm": 0.04771510139107704,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 4.1795806884765625,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68411867364747,
+ "grad_norm": 0.04551433399319649,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 4.192194938659668,
+ "step": 1345
+ },
+ {
+ "epoch": 18.698080279232112,
+ "grad_norm": 0.043501488864421844,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 4.177781105041504,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712041884816752,
+ "grad_norm": 0.04394034296274185,
+ "learning_rate": 0.000432823382695327,
+ "loss": 4.206354141235352,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726003490401396,
+ "grad_norm": 0.04911153391003609,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 4.205874919891357,
+ "step": 1348
+ },
+ {
+ "epoch": 18.739965095986037,
+ "grad_norm": 0.04602045193314552,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 4.1947021484375,
+ "step": 1349
+ },
+ {
+ "epoch": 18.75392670157068,
+ "grad_norm": 0.04393904283642769,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 4.188337802886963,
+ "step": 1350
+ },
+ {
+ "epoch": 18.767888307155324,
+ "grad_norm": 0.04526393488049507,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 4.175235748291016,
+ "step": 1351
+ },
+ {
+ "epoch": 18.781849912739965,
+ "grad_norm": 0.041930247098207474,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 4.210986137390137,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79581151832461,
+ "grad_norm": 0.044262565672397614,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 4.176301002502441,
+ "step": 1353
+ },
+ {
+ "epoch": 18.80977312390925,
+ "grad_norm": 0.044190581887960434,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 4.1888837814331055,
+ "step": 1354
+ },
+ {
+ "epoch": 18.823734729493893,
+ "grad_norm": 0.045371655374765396,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 4.182244777679443,
+ "step": 1355
+ },
+ {
+ "epoch": 18.837696335078533,
+ "grad_norm": 0.04302282631397247,
+ "learning_rate": 0.000430661245733797,
+ "loss": 4.189350128173828,
+ "step": 1356
+ },
+ {
+ "epoch": 18.851657940663177,
+ "grad_norm": 0.04402372986078262,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 4.233372211456299,
+ "step": 1357
+ },
+ {
+ "epoch": 18.865619546247817,
+ "grad_norm": 0.04397045075893402,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 4.21682071685791,
+ "step": 1358
+ },
+ {
+ "epoch": 18.87958115183246,
+ "grad_norm": 0.04319734871387482,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 4.204183578491211,
+ "step": 1359
+ },
+ {
+ "epoch": 18.8935427574171,
+ "grad_norm": 0.04600509628653526,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 4.206524848937988,
+ "step": 1360
+ },
+ {
+ "epoch": 18.907504363001745,
+ "grad_norm": 0.04364705830812454,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 4.213656425476074,
+ "step": 1361
+ },
+ {
+ "epoch": 18.921465968586386,
+ "grad_norm": 0.044004205614328384,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 4.1841583251953125,
+ "step": 1362
+ },
+ {
+ "epoch": 18.93542757417103,
+ "grad_norm": 0.04480595886707306,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 4.166214942932129,
+ "step": 1363
+ },
+ {
+ "epoch": 18.949389179755673,
+ "grad_norm": 0.04260887950658798,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 4.198726654052734,
+ "step": 1364
+ },
+ {
+ "epoch": 18.963350785340314,
+ "grad_norm": 0.04513927176594734,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 4.2122392654418945,
+ "step": 1365
+ },
+ {
+ "epoch": 18.977312390924958,
+ "grad_norm": 0.04286042973399162,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 4.192383766174316,
+ "step": 1366
+ },
+ {
+ "epoch": 18.991273996509598,
+ "grad_norm": 0.04287346079945564,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 4.221860408782959,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.030670644715428352,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 2.624823570251465,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.5910831689834595,
+ "eval_runtime": 60.5517,
+ "eval_samples_per_second": 40.329,
+ "eval_steps_per_second": 0.644,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013961605584644,
+ "grad_norm": 0.046196095645427704,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 4.104501724243164,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027923211169284,
+ "grad_norm": 0.05348595231771469,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 4.109098434448242,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041884816753928,
+ "grad_norm": 0.05256905406713486,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 4.115832328796387,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05584642233857,
+ "grad_norm": 0.05241430550813675,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 4.079610824584961,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069808027923212,
+ "grad_norm": 0.04983370006084442,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 4.095335006713867,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083769633507853,
+ "grad_norm": 0.04900007322430611,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 4.093306541442871,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097731239092496,
+ "grad_norm": 0.04995666816830635,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 4.111114501953125,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111692844677137,
+ "grad_norm": 0.04875681549310684,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 4.091944694519043,
+ "step": 1376
+ },
+ {
+ "epoch": 19.12565445026178,
+ "grad_norm": 0.05283811315894127,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 4.112549304962158,
+ "step": 1377
+ },
+ {
+ "epoch": 19.13961605584642,
+ "grad_norm": 0.04801992326974869,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 4.108818054199219,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153577661431065,
+ "grad_norm": 0.04751930385828018,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 4.089441299438477,
+ "step": 1379
+ },
+ {
+ "epoch": 19.167539267015705,
+ "grad_norm": 0.045653726905584335,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 4.07722282409668,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18150087260035,
+ "grad_norm": 0.045616209506988525,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 4.097578048706055,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195462478184993,
+ "grad_norm": 0.04929365590214729,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 4.082221984863281,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209424083769633,
+ "grad_norm": 0.04915834963321686,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 4.123330116271973,
+ "step": 1383
+ },
+ {
+ "epoch": 19.223385689354277,
+ "grad_norm": 0.0470132976770401,
+ "learning_rate": 0.000423881957237287,
+ "loss": 4.113927841186523,
+ "step": 1384
+ },
+ {
+ "epoch": 19.237347294938917,
+ "grad_norm": 0.04695259779691696,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 4.096814155578613,
+ "step": 1385
+ },
+ {
+ "epoch": 19.25130890052356,
+ "grad_norm": 0.04730398207902908,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 4.122525215148926,
+ "step": 1386
+ },
+ {
+ "epoch": 19.2652705061082,
+ "grad_norm": 0.050295207649469376,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 4.117993354797363,
+ "step": 1387
+ },
+ {
+ "epoch": 19.279232111692846,
+ "grad_norm": 0.05082492157816887,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 4.14082145690918,
+ "step": 1388
+ },
+ {
+ "epoch": 19.293193717277486,
+ "grad_norm": 0.04960232973098755,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 4.139887809753418,
+ "step": 1389
+ },
+ {
+ "epoch": 19.30715532286213,
+ "grad_norm": 0.04851173236966133,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 4.138492584228516,
+ "step": 1390
+ },
+ {
+ "epoch": 19.32111692844677,
+ "grad_norm": 0.047838468104600906,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 4.118064880371094,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335078534031414,
+ "grad_norm": 0.05148220434784889,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 4.136533737182617,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349040139616054,
+ "grad_norm": 0.053177181631326675,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 4.1125168800354,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363001745200698,
+ "grad_norm": 0.04861171543598175,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 4.1437153816223145,
+ "step": 1394
+ },
+ {
+ "epoch": 19.376963350785342,
+ "grad_norm": 0.05194924771785736,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 4.097813606262207,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390924956369982,
+ "grad_norm": 0.049609847366809845,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 4.068940162658691,
+ "step": 1396
+ },
+ {
+ "epoch": 19.404886561954626,
+ "grad_norm": 0.047042325139045715,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 4.121333122253418,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418848167539267,
+ "grad_norm": 0.047646693885326385,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 4.114943981170654,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43280977312391,
+ "grad_norm": 0.04676460474729538,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 4.139252185821533,
+ "step": 1399
+ },
+ {
+ "epoch": 19.44677137870855,
+ "grad_norm": 0.04905956983566284,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 4.135970115661621,
+ "step": 1400
+ },
+ {
+ "epoch": 19.460732984293195,
+ "grad_norm": 0.050025034695863724,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 4.15662956237793,
+ "step": 1401
+ },
+ {
+ "epoch": 19.474694589877835,
+ "grad_norm": 0.04740425944328308,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 4.140295505523682,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48865619546248,
+ "grad_norm": 0.046904709190130234,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 4.14083194732666,
+ "step": 1403
+ },
+ {
+ "epoch": 19.50261780104712,
+ "grad_norm": 0.048393651843070984,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 4.161823272705078,
+ "step": 1404
+ },
+ {
+ "epoch": 19.516579406631763,
+ "grad_norm": 0.04766222462058067,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 4.151817798614502,
+ "step": 1405
+ },
+ {
+ "epoch": 19.530541012216403,
+ "grad_norm": 0.04797709733247757,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 4.173160552978516,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544502617801047,
+ "grad_norm": 0.05012711137533188,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 4.161989688873291,
+ "step": 1407
+ },
+ {
+ "epoch": 19.55846422338569,
+ "grad_norm": 0.04726000875234604,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 4.136838436126709,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57242582897033,
+ "grad_norm": 0.046251531690359116,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 4.1526994705200195,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586387434554975,
+ "grad_norm": 0.047958049923181534,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 4.13817024230957,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600349040139616,
+ "grad_norm": 0.046333350241184235,
+ "learning_rate": 0.000417272427439646,
+ "loss": 4.160187721252441,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61431064572426,
+ "grad_norm": 0.0488678440451622,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 4.166190147399902,
+ "step": 1412
+ },
+ {
+ "epoch": 19.6282722513089,
+ "grad_norm": 0.04784323647618294,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 4.164414405822754,
+ "step": 1413
+ },
+ {
+ "epoch": 19.642233856893544,
+ "grad_norm": 0.05064673349261284,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 4.149814128875732,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656195462478184,
+ "grad_norm": 0.05072085186839104,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 4.167759895324707,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670157068062828,
+ "grad_norm": 0.0499395951628685,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 4.162792205810547,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68411867364747,
+ "grad_norm": 0.046766918152570724,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 4.1416521072387695,
+ "step": 1417
+ },
+ {
+ "epoch": 19.698080279232112,
+ "grad_norm": 0.04523273929953575,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 4.161751747131348,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712041884816752,
+ "grad_norm": 0.046106480062007904,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 4.166394233703613,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726003490401396,
+ "grad_norm": 0.04666793718934059,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 4.124961853027344,
+ "step": 1420
+ },
+ {
+ "epoch": 19.739965095986037,
+ "grad_norm": 0.047407012432813644,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 4.1794328689575195,
+ "step": 1421
+ },
+ {
+ "epoch": 19.75392670157068,
+ "grad_norm": 0.04690873995423317,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 4.160114765167236,
+ "step": 1422
+ },
+ {
+ "epoch": 19.767888307155324,
+ "grad_norm": 0.04589436575770378,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 4.137969970703125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.781849912739965,
+ "grad_norm": 0.04641961678862572,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 4.161533355712891,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79581151832461,
+ "grad_norm": 0.045799050480127335,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 4.187039375305176,
+ "step": 1425
+ },
+ {
+ "epoch": 19.80977312390925,
+ "grad_norm": 0.04881894588470459,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 4.158107757568359,
+ "step": 1426
+ },
+ {
+ "epoch": 19.823734729493893,
+ "grad_norm": 0.04506978392601013,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 4.143360137939453,
+ "step": 1427
+ },
+ {
+ "epoch": 19.837696335078533,
+ "grad_norm": 0.04504861310124397,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 4.174564361572266,
+ "step": 1428
+ },
+ {
+ "epoch": 19.851657940663177,
+ "grad_norm": 0.04639774188399315,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 4.162899971008301,
+ "step": 1429
+ },
+ {
+ "epoch": 19.865619546247817,
+ "grad_norm": 0.047444432973861694,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 4.143390655517578,
+ "step": 1430
+ },
+ {
+ "epoch": 19.87958115183246,
+ "grad_norm": 0.04461613669991493,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 4.171284198760986,
+ "step": 1431
+ },
+ {
+ "epoch": 19.8935427574171,
+ "grad_norm": 0.046383798122406006,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 4.14809513092041,
+ "step": 1432
+ },
+ {
+ "epoch": 19.907504363001745,
+ "grad_norm": 0.04642366245388985,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 4.178903579711914,
+ "step": 1433
+ },
+ {
+ "epoch": 19.921465968586386,
+ "grad_norm": 0.0451771542429924,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 4.171825408935547,
+ "step": 1434
+ },
+ {
+ "epoch": 19.93542757417103,
+ "grad_norm": 0.04767893627285957,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 4.169270038604736,
+ "step": 1435
+ },
+ {
+ "epoch": 19.949389179755673,
+ "grad_norm": 0.04673412814736366,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 4.175506591796875,
+ "step": 1436
+ },
+ {
+ "epoch": 19.963350785340314,
+ "grad_norm": 0.04465145245194435,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 4.149707794189453,
+ "step": 1437
+ },
+ {
+ "epoch": 19.977312390924958,
+ "grad_norm": 0.04734131321310997,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 4.166044235229492,
+ "step": 1438
+ },
+ {
+ "epoch": 19.991273996509598,
+ "grad_norm": 0.04706178605556488,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 4.152081489562988,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.031556785106658936,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 2.60109543800354,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013961605584644,
+ "grad_norm": 0.05152537673711777,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 4.084830284118652,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027923211169284,
+ "grad_norm": 0.05918274074792862,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 4.045302391052246,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041884816753928,
+ "grad_norm": 0.05093073844909668,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 4.070842742919922,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05584642233857,
+ "grad_norm": 0.05095985531806946,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 4.083273410797119,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069808027923212,
+ "grad_norm": 0.05502687394618988,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 4.072626113891602,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083769633507853,
+ "grad_norm": 0.05037364363670349,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 4.075478553771973,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097731239092496,
+ "grad_norm": 0.04983789846301079,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 4.066207408905029,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111692844677137,
+ "grad_norm": 0.049478255212306976,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 4.069969654083252,
+ "step": 1448
+ },
+ {
+ "epoch": 20.12565445026178,
+ "grad_norm": 0.051651593297719955,
+ "learning_rate": 0.000407857329622967,
+ "loss": 4.050713062286377,
+ "step": 1449
+ },
+ {
+ "epoch": 20.13961605584642,
+ "grad_norm": 0.05436555668711662,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 4.069849967956543,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153577661431065,
+ "grad_norm": 0.055190350860357285,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 4.081624984741211,
+ "step": 1451
+ },
+ {
+ "epoch": 20.167539267015705,
+ "grad_norm": 0.053246211260557175,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 4.076076507568359,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18150087260035,
+ "grad_norm": 0.051757775247097015,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 4.049139022827148,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195462478184993,
+ "grad_norm": 0.05064203217625618,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 4.068076133728027,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209424083769633,
+ "grad_norm": 0.05114085227251053,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 4.092901229858398,
+ "step": 1455
+ },
+ {
+ "epoch": 20.223385689354277,
+ "grad_norm": 0.047997843474149704,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 4.077084541320801,
+ "step": 1456
+ },
+ {
+ "epoch": 20.237347294938917,
+ "grad_norm": 0.04880451038479805,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 4.059187889099121,
+ "step": 1457
+ },
+ {
+ "epoch": 20.25130890052356,
+ "grad_norm": 0.05036284029483795,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 4.102115154266357,
+ "step": 1458
+ },
+ {
+ "epoch": 20.2652705061082,
+ "grad_norm": 0.04916979745030403,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 4.098356246948242,
+ "step": 1459
+ },
+ {
+ "epoch": 20.279232111692846,
+ "grad_norm": 0.04713984206318855,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 4.057218074798584,
+ "step": 1460
+ },
+ {
+ "epoch": 20.293193717277486,
+ "grad_norm": 0.04982447996735573,
+ "learning_rate": 0.000404858275823277,
+ "loss": 4.070363998413086,
+ "step": 1461
+ },
+ {
+ "epoch": 20.30715532286213,
+ "grad_norm": 0.0475616492331028,
+ "learning_rate": 0.000404607816444578,
+ "loss": 4.074314117431641,
+ "step": 1462
+ },
+ {
+ "epoch": 20.32111692844677,
+ "grad_norm": 0.046537093818187714,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 4.079501152038574,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335078534031414,
+ "grad_norm": 0.04889644682407379,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 4.0731425285339355,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349040139616054,
+ "grad_norm": 0.05095166340470314,
+ "learning_rate": 0.000403855947934478,
+ "loss": 4.08765983581543,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363001745200698,
+ "grad_norm": 0.049677830189466476,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 4.128052711486816,
+ "step": 1466
+ },
+ {
+ "epoch": 20.376963350785342,
+ "grad_norm": 0.05221942439675331,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 4.094407081604004,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390924956369982,
+ "grad_norm": 0.052042894065380096,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 4.088282585144043,
+ "step": 1468
+ },
+ {
+ "epoch": 20.404886561954626,
+ "grad_norm": 0.04822041094303131,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 4.1070990562438965,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418848167539267,
+ "grad_norm": 0.047207772731781006,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 4.102565765380859,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43280977312391,
+ "grad_norm": 0.049633774906396866,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 4.097216606140137,
+ "step": 1471
+ },
+ {
+ "epoch": 20.44677137870855,
+ "grad_norm": 0.05135185644030571,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 4.095193862915039,
+ "step": 1472
+ },
+ {
+ "epoch": 20.460732984293195,
+ "grad_norm": 0.05215826258063316,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 4.085207939147949,
+ "step": 1473
+ },
+ {
+ "epoch": 20.474694589877835,
+ "grad_norm": 0.050653353333473206,
+ "learning_rate": 0.000401595976318565,
+ "loss": 4.085142612457275,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48865619546248,
+ "grad_norm": 0.04795852303504944,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 4.1158270835876465,
+ "step": 1475
+ },
+ {
+ "epoch": 20.50261780104712,
+ "grad_norm": 0.047416940331459045,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 4.096989154815674,
+ "step": 1476
+ },
+ {
+ "epoch": 20.516579406631763,
+ "grad_norm": 0.049753669649362564,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 4.118251800537109,
+ "step": 1477
+ },
+ {
+ "epoch": 20.530541012216403,
+ "grad_norm": 0.04889055714011192,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 4.081287384033203,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544502617801047,
+ "grad_norm": 0.04857790097594261,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 4.084896564483643,
+ "step": 1479
+ },
+ {
+ "epoch": 20.55846422338569,
+ "grad_norm": 0.047250498086214066,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 4.113656997680664,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57242582897033,
+ "grad_norm": 0.048608481884002686,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 4.113809108734131,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586387434554975,
+ "grad_norm": 0.048189081251621246,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 4.113949775695801,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600349040139616,
+ "grad_norm": 0.049929242581129074,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 4.104221820831299,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61431064572426,
+ "grad_norm": 0.048482850193977356,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 4.091039180755615,
+ "step": 1484
+ },
+ {
+ "epoch": 20.6282722513089,
+ "grad_norm": 0.04810645803809166,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 4.130088806152344,
+ "step": 1485
+ },
+ {
+ "epoch": 20.642233856893544,
+ "grad_norm": 0.04745674133300781,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 4.106161117553711,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656195462478184,
+ "grad_norm": 0.048513609915971756,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 4.119222164154053,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670157068062828,
+ "grad_norm": 0.04506325349211693,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 4.115543842315674,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68411867364747,
+ "grad_norm": 0.04726741835474968,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 4.084345817565918,
+ "step": 1489
+ },
+ {
+ "epoch": 20.698080279232112,
+ "grad_norm": 0.04873545467853546,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 4.116580009460449,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712041884816752,
+ "grad_norm": 0.051252465695142746,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 4.129733085632324,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726003490401396,
+ "grad_norm": 0.0489567406475544,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 4.103477954864502,
+ "step": 1492
+ },
+ {
+ "epoch": 20.739965095986037,
+ "grad_norm": 0.04539056867361069,
+ "learning_rate": 0.000396803974909638,
+ "loss": 4.1327667236328125,
+ "step": 1493
+ },
+ {
+ "epoch": 20.75392670157068,
+ "grad_norm": 0.04954095557332039,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 4.084133625030518,
+ "step": 1494
+ },
+ {
+ "epoch": 20.767888307155324,
+ "grad_norm": 0.05041591823101044,
+ "learning_rate": 0.000396297942393265,
+ "loss": 4.12288761138916,
+ "step": 1495
+ },
+ {
+ "epoch": 20.781849912739965,
+ "grad_norm": 0.046549517661333084,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 4.1259918212890625,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79581151832461,
+ "grad_norm": 0.04704999923706055,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 4.102563858032227,
+ "step": 1497
+ },
+ {
+ "epoch": 20.80977312390925,
+ "grad_norm": 0.05025320500135422,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 4.135492324829102,
+ "step": 1498
+ },
+ {
+ "epoch": 20.823734729493893,
+ "grad_norm": 0.04559441655874252,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 4.118548393249512,
+ "step": 1499
+ },
+ {
+ "epoch": 20.837696335078533,
+ "grad_norm": 0.04913092032074928,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 4.129302024841309,
+ "step": 1500
+ },
+ {
+ "epoch": 20.851657940663177,
+ "grad_norm": 0.04846089705824852,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 4.166059970855713,
+ "step": 1501
+ },
+ {
+ "epoch": 20.865619546247817,
+ "grad_norm": 0.04710553213953972,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 4.104712009429932,
+ "step": 1502
+ },
+ {
+ "epoch": 20.87958115183246,
+ "grad_norm": 0.04896444454789162,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 4.140282154083252,
+ "step": 1503
+ },
+ {
+ "epoch": 20.8935427574171,
+ "grad_norm": 0.0469740591943264,
+ "learning_rate": 0.000394017098798633,
+ "loss": 4.125306129455566,
+ "step": 1504
+ },
+ {
+ "epoch": 20.907504363001745,
+ "grad_norm": 0.04638523980975151,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 4.133161544799805,
+ "step": 1505
+ },
+ {
+ "epoch": 20.921465968586386,
+ "grad_norm": 0.04814442619681358,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 4.093092918395996,
+ "step": 1506
+ },
+ {
+ "epoch": 20.93542757417103,
+ "grad_norm": 0.049911532551050186,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 4.1001715660095215,
+ "step": 1507
+ },
+ {
+ "epoch": 20.949389179755673,
+ "grad_norm": 0.0446670837700367,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 4.140350341796875,
+ "step": 1508
+ },
+ {
+ "epoch": 20.963350785340314,
+ "grad_norm": 0.04789397493004799,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 4.163155555725098,
+ "step": 1509
+ },
+ {
+ "epoch": 20.977312390924958,
+ "grad_norm": 0.050148315727710724,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 4.118380069732666,
+ "step": 1510
+ },
+ {
+ "epoch": 20.991273996509598,
+ "grad_norm": 0.047446198761463165,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 4.116507530212402,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.03415544703602791,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 2.5738282203674316,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.5933796167373657,
+ "eval_runtime": 60.4011,
+ "eval_samples_per_second": 40.43,
+ "eval_steps_per_second": 0.646,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013961605584644,
+ "grad_norm": 0.052276045083999634,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 4.0301923751831055,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027923211169284,
+ "grad_norm": 0.06514342874288559,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 4.031280517578125,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041884816753928,
+ "grad_norm": 0.05948134884238243,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 4.060680389404297,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05584642233857,
+ "grad_norm": 0.05251640826463699,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 4.0218400955200195,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069808027923212,
+ "grad_norm": 0.05426032841205597,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 4.026187896728516,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083769633507853,
+ "grad_norm": 0.052583567798137665,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 4.016767501831055,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097731239092496,
+ "grad_norm": 0.04966135695576668,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 4.043508529663086,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111692844677137,
+ "grad_norm": 0.053838398307561874,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 4.054832458496094,
+ "step": 1520
+ },
+ {
+ "epoch": 21.12565445026178,
+ "grad_norm": 0.05304184556007385,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 4.03280782699585,
+ "step": 1521
+ },
+ {
+ "epoch": 21.13961605584642,
+ "grad_norm": 0.05252664163708687,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 4.039841175079346,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153577661431065,
+ "grad_norm": 0.054593686014413834,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 4.048308372497559,
+ "step": 1523
+ },
+ {
+ "epoch": 21.167539267015705,
+ "grad_norm": 0.050722211599349976,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 4.051860332489014,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18150087260035,
+ "grad_norm": 0.05248095095157623,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 4.056846618652344,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195462478184993,
+ "grad_norm": 0.05166107788681984,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 4.045178413391113,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209424083769633,
+ "grad_norm": 0.04987792298197746,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 4.014827728271484,
+ "step": 1527
+ },
+ {
+ "epoch": 21.223385689354277,
+ "grad_norm": 0.05316167324781418,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 4.033143997192383,
+ "step": 1528
+ },
+ {
+ "epoch": 21.237347294938917,
+ "grad_norm": 0.05228257551789284,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 4.033842086791992,
+ "step": 1529
+ },
+ {
+ "epoch": 21.25130890052356,
+ "grad_norm": 0.049349382519721985,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 4.025982856750488,
+ "step": 1530
+ },
+ {
+ "epoch": 21.2652705061082,
+ "grad_norm": 0.055104926228523254,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 4.059151649475098,
+ "step": 1531
+ },
+ {
+ "epoch": 21.279232111692846,
+ "grad_norm": 0.052629340440034866,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 4.032320976257324,
+ "step": 1532
+ },
+ {
+ "epoch": 21.293193717277486,
+ "grad_norm": 0.05396415665745735,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 4.0425567626953125,
+ "step": 1533
+ },
+ {
+ "epoch": 21.30715532286213,
+ "grad_norm": 0.054417286068201065,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 4.03759241104126,
+ "step": 1534
+ },
+ {
+ "epoch": 21.32111692844677,
+ "grad_norm": 0.061248697340488434,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 4.0441083908081055,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335078534031414,
+ "grad_norm": 0.05934970825910568,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 4.050718307495117,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349040139616054,
+ "grad_norm": 0.052805181592702866,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 4.069460868835449,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363001745200698,
+ "grad_norm": 0.052820611745119095,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 4.063657283782959,
+ "step": 1538
+ },
+ {
+ "epoch": 21.376963350785342,
+ "grad_norm": 0.05043052136898041,
+ "learning_rate": 0.00038509205478744,
+ "loss": 4.029552459716797,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390924956369982,
+ "grad_norm": 0.04828709363937378,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 4.05656623840332,
+ "step": 1540
+ },
+ {
+ "epoch": 21.404886561954626,
+ "grad_norm": 0.049760788679122925,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 4.066338539123535,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418848167539267,
+ "grad_norm": 0.05037159472703934,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 4.06976842880249,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43280977312391,
+ "grad_norm": 0.05276317894458771,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 4.062975883483887,
+ "step": 1543
+ },
+ {
+ "epoch": 21.44677137870855,
+ "grad_norm": 0.05571872368454933,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 4.065622329711914,
+ "step": 1544
+ },
+ {
+ "epoch": 21.460732984293195,
+ "grad_norm": 0.05438469722867012,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 4.074963569641113,
+ "step": 1545
+ },
+ {
+ "epoch": 21.474694589877835,
+ "grad_norm": 0.053415946662425995,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 4.080055236816406,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48865619546248,
+ "grad_norm": 0.05216296389698982,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 4.091355800628662,
+ "step": 1547
+ },
+ {
+ "epoch": 21.50261780104712,
+ "grad_norm": 0.051212459802627563,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 4.058528900146484,
+ "step": 1548
+ },
+ {
+ "epoch": 21.516579406631763,
+ "grad_norm": 0.05033959075808525,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 4.064615726470947,
+ "step": 1549
+ },
+ {
+ "epoch": 21.530541012216403,
+ "grad_norm": 0.04665188118815422,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 4.059698104858398,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544502617801047,
+ "grad_norm": 0.050221361219882965,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 4.068614959716797,
+ "step": 1551
+ },
+ {
+ "epoch": 21.55846422338569,
+ "grad_norm": 0.05089551955461502,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 4.0833892822265625,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57242582897033,
+ "grad_norm": 0.051966603845357895,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 4.083354949951172,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586387434554975,
+ "grad_norm": 0.04981927201151848,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 4.083340644836426,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600349040139616,
+ "grad_norm": 0.05141765624284744,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 4.066590309143066,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61431064572426,
+ "grad_norm": 0.048985838890075684,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 4.088237762451172,
+ "step": 1556
+ },
+ {
+ "epoch": 21.6282722513089,
+ "grad_norm": 0.05013054236769676,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 4.081377983093262,
+ "step": 1557
+ },
+ {
+ "epoch": 21.642233856893544,
+ "grad_norm": 0.05170975625514984,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 4.054303169250488,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656195462478184,
+ "grad_norm": 0.048257142305374146,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 4.103672504425049,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670157068062828,
+ "grad_norm": 0.049942679703235626,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 4.029951572418213,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68411867364747,
+ "grad_norm": 0.05083315819501877,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 4.048798084259033,
+ "step": 1561
+ },
+ {
+ "epoch": 21.698080279232112,
+ "grad_norm": 0.04984128847718239,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 4.079655647277832,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712041884816752,
+ "grad_norm": 0.04985307902097702,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 4.071877956390381,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726003490401396,
+ "grad_norm": 0.051137227565050125,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 4.076539993286133,
+ "step": 1564
+ },
+ {
+ "epoch": 21.739965095986037,
+ "grad_norm": 0.04968905821442604,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 4.084498882293701,
+ "step": 1565
+ },
+ {
+ "epoch": 21.75392670157068,
+ "grad_norm": 0.049119509756565094,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 4.066773414611816,
+ "step": 1566
+ },
+ {
+ "epoch": 21.767888307155324,
+ "grad_norm": 0.05107425898313522,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 4.065309047698975,
+ "step": 1567
+ },
+ {
+ "epoch": 21.781849912739965,
+ "grad_norm": 0.049012504518032074,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 4.062110900878906,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79581151832461,
+ "grad_norm": 0.04972011595964432,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 4.0630903244018555,
+ "step": 1569
+ },
+ {
+ "epoch": 21.80977312390925,
+ "grad_norm": 0.04822700098156929,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 4.106127738952637,
+ "step": 1570
+ },
+ {
+ "epoch": 21.823734729493893,
+ "grad_norm": 0.04808267205953598,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 4.085200309753418,
+ "step": 1571
+ },
+ {
+ "epoch": 21.837696335078533,
+ "grad_norm": 0.048644762486219406,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 4.0856170654296875,
+ "step": 1572
+ },
+ {
+ "epoch": 21.851657940663177,
+ "grad_norm": 0.05232283100485802,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 4.126387119293213,
+ "step": 1573
+ },
+ {
+ "epoch": 21.865619546247817,
+ "grad_norm": 0.05095643922686577,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 4.108254432678223,
+ "step": 1574
+ },
+ {
+ "epoch": 21.87958115183246,
+ "grad_norm": 0.045972101390361786,
+ "learning_rate": 0.000375827577564042,
+ "loss": 4.058515548706055,
+ "step": 1575
+ },
+ {
+ "epoch": 21.8935427574171,
+ "grad_norm": 0.04974910616874695,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 4.089197158813477,
+ "step": 1576
+ },
+ {
+ "epoch": 21.907504363001745,
+ "grad_norm": 0.04890159144997597,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 4.1078410148620605,
+ "step": 1577
+ },
+ {
+ "epoch": 21.921465968586386,
+ "grad_norm": 0.04738525673747063,
+ "learning_rate": 0.000375051971546236,
+ "loss": 4.1219282150268555,
+ "step": 1578
+ },
+ {
+ "epoch": 21.93542757417103,
+ "grad_norm": 0.04713096842169762,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 4.10295295715332,
+ "step": 1579
+ },
+ {
+ "epoch": 21.949389179755673,
+ "grad_norm": 0.04834369197487831,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 4.102747917175293,
+ "step": 1580
+ },
+ {
+ "epoch": 21.963350785340314,
+ "grad_norm": 0.047073688358068466,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 4.08024787902832,
+ "step": 1581
+ },
+ {
+ "epoch": 21.977312390924958,
+ "grad_norm": 0.04568164795637131,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 4.078930854797363,
+ "step": 1582
+ },
+ {
+ "epoch": 21.991273996509598,
+ "grad_norm": 0.04822581633925438,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 4.089601516723633,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.03507270663976669,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 2.5586814880371094,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.595272958278656,
+ "eval_runtime": 60.7797,
+ "eval_samples_per_second": 40.178,
+ "eval_steps_per_second": 0.642,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013961605584644,
+ "grad_norm": 0.05232533439993858,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 4.011311054229736,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027923211169284,
+ "grad_norm": 0.058515798300504684,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 4.0049638748168945,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041884816753928,
+ "grad_norm": 0.055082619190216064,
+ "learning_rate": 0.000372722041257003,
+ "loss": 3.9904184341430664,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05584642233857,
+ "grad_norm": 0.05213981494307518,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 4.003454685211182,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069808027923212,
+ "grad_norm": 0.051619648933410645,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 3.9900388717651367,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083769633507853,
+ "grad_norm": 0.055212799459695816,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 3.973494052886963,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097731239092496,
+ "grad_norm": 0.05262092128396034,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 4.00933837890625,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111692844677137,
+ "grad_norm": 0.05012505501508713,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 4.01861047744751,
+ "step": 1592
+ },
+ {
+ "epoch": 22.12565445026178,
+ "grad_norm": 0.049657098948955536,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 4.011143684387207,
+ "step": 1593
+ },
+ {
+ "epoch": 22.13961605584642,
+ "grad_norm": 0.05187768116593361,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 3.9922969341278076,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153577661431065,
+ "grad_norm": 0.05053112655878067,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 4.007021903991699,
+ "step": 1595
+ },
+ {
+ "epoch": 22.167539267015705,
+ "grad_norm": 0.048542529344558716,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 3.9966416358947754,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18150087260035,
+ "grad_norm": 0.050972841680049896,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 3.9803225994110107,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195462478184993,
+ "grad_norm": 0.05001165345311165,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 4.0140509605407715,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209424083769633,
+ "grad_norm": 0.05100536718964577,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 3.9906058311462402,
+ "step": 1599
+ },
+ {
+ "epoch": 22.223385689354277,
+ "grad_norm": 0.052977193146944046,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 4.012383460998535,
+ "step": 1600
+ },
+ {
+ "epoch": 22.237347294938917,
+ "grad_norm": 0.051219623535871506,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 3.982931613922119,
+ "step": 1601
+ },
+ {
+ "epoch": 22.25130890052356,
+ "grad_norm": 0.05275029316544533,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 4.032669544219971,
+ "step": 1602
+ },
+ {
+ "epoch": 22.2652705061082,
+ "grad_norm": 0.04781945049762726,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 3.9933102130889893,
+ "step": 1603
+ },
+ {
+ "epoch": 22.279232111692846,
+ "grad_norm": 0.054087597876787186,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 4.0075201988220215,
+ "step": 1604
+ },
+ {
+ "epoch": 22.293193717277486,
+ "grad_norm": 0.05485915020108223,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 4.004299640655518,
+ "step": 1605
+ },
+ {
+ "epoch": 22.30715532286213,
+ "grad_norm": 0.05148882791399956,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 4.044161796569824,
+ "step": 1606
+ },
+ {
+ "epoch": 22.32111692844677,
+ "grad_norm": 0.051912397146224976,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 3.9777069091796875,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335078534031414,
+ "grad_norm": 0.052630942314863205,
+ "learning_rate": 0.000367268105959126,
+ "loss": 4.029580116271973,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349040139616054,
+ "grad_norm": 0.0512809231877327,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 4.022092342376709,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363001745200698,
+ "grad_norm": 0.05094102770090103,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 4.018102169036865,
+ "step": 1610
+ },
+ {
+ "epoch": 22.376963350785342,
+ "grad_norm": 0.04972759634256363,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 4.027868270874023,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390924956369982,
+ "grad_norm": 0.05072534456849098,
+ "learning_rate": 0.000366226612810024,
+ "loss": 3.995988368988037,
+ "step": 1612
+ },
+ {
+ "epoch": 22.404886561954626,
+ "grad_norm": 0.050623781979084015,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 4.019711494445801,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418848167539267,
+ "grad_norm": 0.05035007372498512,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 4.007514953613281,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43280977312391,
+ "grad_norm": 0.054426416754722595,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 4.0167131423950195,
+ "step": 1615
+ },
+ {
+ "epoch": 22.44677137870855,
+ "grad_norm": 0.05365690588951111,
+ "learning_rate": 0.000365184304613104,
+ "loss": 4.0313920974731445,
+ "step": 1616
+ },
+ {
+ "epoch": 22.460732984293195,
+ "grad_norm": 0.0509161539375782,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 4.022063255310059,
+ "step": 1617
+ },
+ {
+ "epoch": 22.474694589877835,
+ "grad_norm": 0.05309660732746124,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 4.017317771911621,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48865619546248,
+ "grad_norm": 0.0488961897790432,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 4.028288841247559,
+ "step": 1619
+ },
+ {
+ "epoch": 22.50261780104712,
+ "grad_norm": 0.04982095956802368,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 4.007115840911865,
+ "step": 1620
+ },
+ {
+ "epoch": 22.516579406631763,
+ "grad_norm": 0.05136589705944061,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 4.028022766113281,
+ "step": 1621
+ },
+ {
+ "epoch": 22.530541012216403,
+ "grad_norm": 0.04865279048681259,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 4.015928268432617,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544502617801047,
+ "grad_norm": 0.05131891369819641,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 4.041997909545898,
+ "step": 1623
+ },
+ {
+ "epoch": 22.55846422338569,
+ "grad_norm": 0.05063068866729736,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 4.042117118835449,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57242582897033,
+ "grad_norm": 0.049348752945661545,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 4.012609481811523,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586387434554975,
+ "grad_norm": 0.05123874917626381,
+ "learning_rate": 0.000362575056487632,
+ "loss": 4.03358793258667,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600349040139616,
+ "grad_norm": 0.04921206086874008,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 4.039857864379883,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61431064572426,
+ "grad_norm": 0.05011922866106033,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 4.015835762023926,
+ "step": 1628
+ },
+ {
+ "epoch": 22.6282722513089,
+ "grad_norm": 0.049950532615184784,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 4.068299293518066,
+ "step": 1629
+ },
+ {
+ "epoch": 22.642233856893544,
+ "grad_norm": 0.04962172359228134,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 4.03205680847168,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656195462478184,
+ "grad_norm": 0.05059773847460747,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 4.065027236938477,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670157068062828,
+ "grad_norm": 0.04986768960952759,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 4.01540470123291,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68411867364747,
+ "grad_norm": 0.050884976983070374,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 4.060809135437012,
+ "step": 1633
+ },
+ {
+ "epoch": 22.698080279232112,
+ "grad_norm": 0.052478861063718796,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 4.088409423828125,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712041884816752,
+ "grad_norm": 0.05118529871106148,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 4.071645736694336,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726003490401396,
+ "grad_norm": 0.05059075728058815,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 4.028438568115234,
+ "step": 1636
+ },
+ {
+ "epoch": 22.739965095986037,
+ "grad_norm": 0.049222562462091446,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 4.02166223526001,
+ "step": 1637
+ },
+ {
+ "epoch": 22.75392670157068,
+ "grad_norm": 0.049673549830913544,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 4.0511860847473145,
+ "step": 1638
+ },
+ {
+ "epoch": 22.767888307155324,
+ "grad_norm": 0.04954935610294342,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 4.036063194274902,
+ "step": 1639
+ },
+ {
+ "epoch": 22.781849912739965,
+ "grad_norm": 0.05110877379775047,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 4.081101417541504,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79581151832461,
+ "grad_norm": 0.049009695649147034,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 4.042314529418945,
+ "step": 1641
+ },
+ {
+ "epoch": 22.80977312390925,
+ "grad_norm": 0.0496281161904335,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 4.028315544128418,
+ "step": 1642
+ },
+ {
+ "epoch": 22.823734729493893,
+ "grad_norm": 0.04931195452809334,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 4.018457412719727,
+ "step": 1643
+ },
+ {
+ "epoch": 22.837696335078533,
+ "grad_norm": 0.05006880313158035,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 4.0709991455078125,
+ "step": 1644
+ },
+ {
+ "epoch": 22.851657940663177,
+ "grad_norm": 0.04956946149468422,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 4.065585136413574,
+ "step": 1645
+ },
+ {
+ "epoch": 22.865619546247817,
+ "grad_norm": 0.048182398080825806,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 4.0522589683532715,
+ "step": 1646
+ },
+ {
+ "epoch": 22.87958115183246,
+ "grad_norm": 0.04964807257056236,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 4.06989860534668,
+ "step": 1647
+ },
+ {
+ "epoch": 22.8935427574171,
+ "grad_norm": 0.049299128353595734,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 4.082858085632324,
+ "step": 1648
+ },
+ {
+ "epoch": 22.907504363001745,
+ "grad_norm": 0.04893320053815842,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 4.047144889831543,
+ "step": 1649
+ },
+ {
+ "epoch": 22.921465968586386,
+ "grad_norm": 0.04772866889834404,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 4.043060302734375,
+ "step": 1650
+ },
+ {
+ "epoch": 22.93542757417103,
+ "grad_norm": 0.049417730420827866,
+ "learning_rate": 0.000356031397755195,
+ "loss": 4.031186580657959,
+ "step": 1651
+ },
+ {
+ "epoch": 22.949389179755673,
+ "grad_norm": 0.049944665282964706,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 4.074260711669922,
+ "step": 1652
+ },
+ {
+ "epoch": 22.963350785340314,
+ "grad_norm": 0.04856271669268608,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 4.054103851318359,
+ "step": 1653
+ },
+ {
+ "epoch": 22.977312390924958,
+ "grad_norm": 0.0470513179898262,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 4.041568756103516,
+ "step": 1654
+ },
+ {
+ "epoch": 22.991273996509598,
+ "grad_norm": 0.049018729478120804,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 4.066161632537842,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.034226156771183014,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 2.5147223472595215,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.5970026850700378,
+ "eval_runtime": 60.0787,
+ "eval_samples_per_second": 40.647,
+ "eval_steps_per_second": 0.649,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013961605584644,
+ "grad_norm": 0.05091085657477379,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 3.95741868019104,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027923211169284,
+ "grad_norm": 0.05874285474419594,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 3.966327667236328,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041884816753928,
+ "grad_norm": 0.05394266918301582,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.937232732772827,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05584642233857,
+ "grad_norm": 0.05120540037751198,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 3.9637370109558105,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069808027923212,
+ "grad_norm": 0.05310065299272537,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 3.998631477355957,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083769633507853,
+ "grad_norm": 0.05127466097474098,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.949101448059082,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097731239092496,
+ "grad_norm": 0.05201594531536102,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 3.986267566680908,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111692844677137,
+ "grad_norm": 0.051927369087934494,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 3.9604103565216064,
+ "step": 1664
+ },
+ {
+ "epoch": 23.12565445026178,
+ "grad_norm": 0.05125141143798828,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.9339547157287598,
+ "step": 1665
+ },
+ {
+ "epoch": 23.13961605584642,
+ "grad_norm": 0.05241088196635246,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 3.971261978149414,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153577661431065,
+ "grad_norm": 0.050563592463731766,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.9450082778930664,
+ "step": 1667
+ },
+ {
+ "epoch": 23.167539267015705,
+ "grad_norm": 0.05204763635993004,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 3.9696707725524902,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18150087260035,
+ "grad_norm": 0.05518610402941704,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 3.9816057682037354,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195462478184993,
+ "grad_norm": 0.05336211621761322,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.932252883911133,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209424083769633,
+ "grad_norm": 0.052204642444849014,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 3.9818131923675537,
+ "step": 1671
+ },
+ {
+ "epoch": 23.223385689354277,
+ "grad_norm": 0.05299545079469681,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.9488978385925293,
+ "step": 1672
+ },
+ {
+ "epoch": 23.237347294938917,
+ "grad_norm": 0.05143733322620392,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 4.00959587097168,
+ "step": 1673
+ },
+ {
+ "epoch": 23.25130890052356,
+ "grad_norm": 0.05186324194073677,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.9529542922973633,
+ "step": 1674
+ },
+ {
+ "epoch": 23.2652705061082,
+ "grad_norm": 0.05334410071372986,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 3.96693754196167,
+ "step": 1675
+ },
+ {
+ "epoch": 23.279232111692846,
+ "grad_norm": 0.05345796048641205,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 3.9674174785614014,
+ "step": 1676
+ },
+ {
+ "epoch": 23.293193717277486,
+ "grad_norm": 0.050988901406526566,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 3.9698638916015625,
+ "step": 1677
+ },
+ {
+ "epoch": 23.30715532286213,
+ "grad_norm": 0.056467678397893906,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 3.9645955562591553,
+ "step": 1678
+ },
+ {
+ "epoch": 23.32111692844677,
+ "grad_norm": 0.05554647371172905,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 4.0036091804504395,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335078534031414,
+ "grad_norm": 0.05607818067073822,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 3.973773717880249,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349040139616054,
+ "grad_norm": 0.05734477937221527,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 3.987692356109619,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363001745200698,
+ "grad_norm": 0.052175361663103104,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 3.9761803150177,
+ "step": 1682
+ },
+ {
+ "epoch": 23.376963350785342,
+ "grad_norm": 0.053368814289569855,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 3.965020179748535,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390924956369982,
+ "grad_norm": 0.05505364388227463,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 3.987494468688965,
+ "step": 1684
+ },
+ {
+ "epoch": 23.404886561954626,
+ "grad_norm": 0.0524514764547348,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 3.9624319076538086,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418848167539267,
+ "grad_norm": 0.05681568384170532,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 4.008373737335205,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43280977312391,
+ "grad_norm": 0.05344909429550171,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 3.9840354919433594,
+ "step": 1687
+ },
+ {
+ "epoch": 23.44677137870855,
+ "grad_norm": 0.052657466381788254,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 3.992262363433838,
+ "step": 1688
+ },
+ {
+ "epoch": 23.460732984293195,
+ "grad_norm": 0.05526747554540634,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 3.97529935836792,
+ "step": 1689
+ },
+ {
+ "epoch": 23.474694589877835,
+ "grad_norm": 0.05492551624774933,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 4.004708766937256,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48865619546248,
+ "grad_norm": 0.053742099553346634,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 3.994016647338867,
+ "step": 1691
+ },
+ {
+ "epoch": 23.50261780104712,
+ "grad_norm": 0.05363722890615463,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 3.9942378997802734,
+ "step": 1692
+ },
+ {
+ "epoch": 23.516579406631763,
+ "grad_norm": 0.05167972669005394,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 3.998253345489502,
+ "step": 1693
+ },
+ {
+ "epoch": 23.530541012216403,
+ "grad_norm": 0.0518590584397316,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 3.9850645065307617,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544502617801047,
+ "grad_norm": 0.05521136894822121,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 4.02669620513916,
+ "step": 1695
+ },
+ {
+ "epoch": 23.55846422338569,
+ "grad_norm": 0.054986029863357544,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 4.022308826446533,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57242582897033,
+ "grad_norm": 0.05405082181096077,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 4.021487236022949,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586387434554975,
+ "grad_norm": 0.051257841289043427,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 3.9724135398864746,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600349040139616,
+ "grad_norm": 0.05109119415283203,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 3.978914499282837,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61431064572426,
+ "grad_norm": 0.05303420498967171,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 4.027509689331055,
+ "step": 1700
+ },
+ {
+ "epoch": 23.6282722513089,
+ "grad_norm": 0.055543459951877594,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 3.9750852584838867,
+ "step": 1701
+ },
+ {
+ "epoch": 23.642233856893544,
+ "grad_norm": 0.05068540200591087,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 4.029600143432617,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656195462478184,
+ "grad_norm": 0.04991193488240242,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 3.9901509284973145,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670157068062828,
+ "grad_norm": 0.05403713881969452,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 4.017679214477539,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68411867364747,
+ "grad_norm": 0.05203867331147194,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 4.008406639099121,
+ "step": 1705
+ },
+ {
+ "epoch": 23.698080279232112,
+ "grad_norm": 0.05106597766280174,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 4.009983062744141,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712041884816752,
+ "grad_norm": 0.04976251348853111,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 4.007230281829834,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726003490401396,
+ "grad_norm": 0.05199650675058365,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 3.9902141094207764,
+ "step": 1708
+ },
+ {
+ "epoch": 23.739965095986037,
+ "grad_norm": 0.05285118520259857,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 3.9966492652893066,
+ "step": 1709
+ },
+ {
+ "epoch": 23.75392670157068,
+ "grad_norm": 0.051937371492385864,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 3.962397336959839,
+ "step": 1710
+ },
+ {
+ "epoch": 23.767888307155324,
+ "grad_norm": 0.05179062858223915,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 4.027141571044922,
+ "step": 1711
+ },
+ {
+ "epoch": 23.781849912739965,
+ "grad_norm": 0.049667056649923325,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 4.024991512298584,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79581151832461,
+ "grad_norm": 0.05204658955335617,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 4.002647876739502,
+ "step": 1713
+ },
+ {
+ "epoch": 23.80977312390925,
+ "grad_norm": 0.05069981515407562,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 4.035422325134277,
+ "step": 1714
+ },
+ {
+ "epoch": 23.823734729493893,
+ "grad_norm": 0.05044785887002945,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 4.042465686798096,
+ "step": 1715
+ },
+ {
+ "epoch": 23.837696335078533,
+ "grad_norm": 0.053949885070323944,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 4.002717971801758,
+ "step": 1716
+ },
+ {
+ "epoch": 23.851657940663177,
+ "grad_norm": 0.05139258876442909,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 4.026609897613525,
+ "step": 1717
+ },
+ {
+ "epoch": 23.865619546247817,
+ "grad_norm": 0.05316372960805893,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 4.019758224487305,
+ "step": 1718
+ },
+ {
+ "epoch": 23.87958115183246,
+ "grad_norm": 0.050791382789611816,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 3.9859633445739746,
+ "step": 1719
+ },
+ {
+ "epoch": 23.8935427574171,
+ "grad_norm": 0.05124847590923309,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 4.038055419921875,
+ "step": 1720
+ },
+ {
+ "epoch": 23.907504363001745,
+ "grad_norm": 0.0495096817612648,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 4.02764368057251,
+ "step": 1721
+ },
+ {
+ "epoch": 23.921465968586386,
+ "grad_norm": 0.05205567181110382,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 4.003867149353027,
+ "step": 1722
+ },
+ {
+ "epoch": 23.93542757417103,
+ "grad_norm": 0.05233347788453102,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 4.024157524108887,
+ "step": 1723
+ },
+ {
+ "epoch": 23.949389179755673,
+ "grad_norm": 0.05126349255442619,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 4.05349063873291,
+ "step": 1724
+ },
+ {
+ "epoch": 23.963350785340314,
+ "grad_norm": 0.050006967037916183,
+ "learning_rate": 0.000336518346307424,
+ "loss": 4.024704933166504,
+ "step": 1725
+ },
+ {
+ "epoch": 23.977312390924958,
+ "grad_norm": 0.052343279123306274,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 4.008549213409424,
+ "step": 1726
+ },
+ {
+ "epoch": 23.991273996509598,
+ "grad_norm": 0.05126652494072914,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 4.002162933349609,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.03560106083750725,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 2.511049270629883,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.5985628962516785,
+ "eval_runtime": 59.0712,
+ "eval_samples_per_second": 41.34,
+ "eval_steps_per_second": 0.66,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013961605584644,
+ "grad_norm": 0.05205906555056572,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.936516284942627,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027923211169284,
+ "grad_norm": 0.056368257850408554,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.95200777053833,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041884816753928,
+ "grad_norm": 0.05075085908174515,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.940810441970825,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05584642233857,
+ "grad_norm": 0.05119713768362999,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.90130615234375,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069808027923212,
+ "grad_norm": 0.05254847928881645,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.9491095542907715,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083769633507853,
+ "grad_norm": 0.053025662899017334,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.9371917247772217,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097731239092496,
+ "grad_norm": 0.05402665585279465,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 3.9651317596435547,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111692844677137,
+ "grad_norm": 0.0519571453332901,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.9189062118530273,
+ "step": 1736
+ },
+ {
+ "epoch": 24.12565445026178,
+ "grad_norm": 0.05436019226908684,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.948184013366699,
+ "step": 1737
+ },
+ {
+ "epoch": 24.13961605584642,
+ "grad_norm": 0.056572459638118744,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.9076738357543945,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153577661431065,
+ "grad_norm": 0.05553183704614639,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.933659076690674,
+ "step": 1739
+ },
+ {
+ "epoch": 24.167539267015705,
+ "grad_norm": 0.05554374307394028,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.941751480102539,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18150087260035,
+ "grad_norm": 0.05451614409685135,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.928593635559082,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195462478184993,
+ "grad_norm": 0.055458780378103256,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 3.961820602416992,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209424083769633,
+ "grad_norm": 0.054980456829071045,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 3.9736528396606445,
+ "step": 1743
+ },
+ {
+ "epoch": 24.223385689354277,
+ "grad_norm": 0.05538604408502579,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.935093641281128,
+ "step": 1744
+ },
+ {
+ "epoch": 24.237347294938917,
+ "grad_norm": 0.05580678582191467,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.9240689277648926,
+ "step": 1745
+ },
+ {
+ "epoch": 24.25130890052356,
+ "grad_norm": 0.052294690161943436,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.937718152999878,
+ "step": 1746
+ },
+ {
+ "epoch": 24.2652705061082,
+ "grad_norm": 0.052018795162439346,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.942049503326416,
+ "step": 1747
+ },
+ {
+ "epoch": 24.279232111692846,
+ "grad_norm": 0.0550091527402401,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.9215564727783203,
+ "step": 1748
+ },
+ {
+ "epoch": 24.293193717277486,
+ "grad_norm": 0.05447716638445854,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.9154953956604004,
+ "step": 1749
+ },
+ {
+ "epoch": 24.30715532286213,
+ "grad_norm": 0.05300029739737511,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 3.9642622470855713,
+ "step": 1750
+ },
+ {
+ "epoch": 24.32111692844677,
+ "grad_norm": 0.05312013998627663,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.938897132873535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335078534031414,
+ "grad_norm": 0.053737837821245193,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.94972562789917,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349040139616054,
+ "grad_norm": 0.05087469890713692,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.9301164150238037,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363001745200698,
+ "grad_norm": 0.05179893225431442,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.9393205642700195,
+ "step": 1754
+ },
+ {
+ "epoch": 24.376963350785342,
+ "grad_norm": 0.05277438461780548,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 3.960968494415283,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390924956369982,
+ "grad_norm": 0.05090474337339401,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 3.9516637325286865,
+ "step": 1756
+ },
+ {
+ "epoch": 24.404886561954626,
+ "grad_norm": 0.052786506712436676,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.9332518577575684,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418848167539267,
+ "grad_norm": 0.052095942199230194,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.9108662605285645,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43280977312391,
+ "grad_norm": 0.05310450494289398,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 3.9746041297912598,
+ "step": 1759
+ },
+ {
+ "epoch": 24.44677137870855,
+ "grad_norm": 0.05078566074371338,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.93316650390625,
+ "step": 1760
+ },
+ {
+ "epoch": 24.460732984293195,
+ "grad_norm": 0.0519949309527874,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.9489870071411133,
+ "step": 1761
+ },
+ {
+ "epoch": 24.474694589877835,
+ "grad_norm": 0.05239349976181984,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 3.955118417739868,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48865619546248,
+ "grad_norm": 0.05193360522389412,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 3.9707555770874023,
+ "step": 1763
+ },
+ {
+ "epoch": 24.50261780104712,
+ "grad_norm": 0.05397362262010574,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 3.973053455352783,
+ "step": 1764
+ },
+ {
+ "epoch": 24.516579406631763,
+ "grad_norm": 0.05379117652773857,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 3.957289695739746,
+ "step": 1765
+ },
+ {
+ "epoch": 24.530541012216403,
+ "grad_norm": 0.0526256300508976,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 3.9595577716827393,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544502617801047,
+ "grad_norm": 0.05377693101763725,
+ "learning_rate": 0.000325372061241796,
+ "loss": 3.97849178314209,
+ "step": 1767
+ },
+ {
+ "epoch": 24.55846422338569,
+ "grad_norm": 0.05182809755206108,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 3.9714043140411377,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57242582897033,
+ "grad_norm": 0.05364793911576271,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 3.9582200050354004,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586387434554975,
+ "grad_norm": 0.051558300852775574,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 3.9747936725616455,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600349040139616,
+ "grad_norm": 0.053487639874219894,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 3.9556500911712646,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61431064572426,
+ "grad_norm": 0.05146970972418785,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.94858455657959,
+ "step": 1772
+ },
+ {
+ "epoch": 24.6282722513089,
+ "grad_norm": 0.05099836364388466,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 3.9716954231262207,
+ "step": 1773
+ },
+ {
+ "epoch": 24.642233856893544,
+ "grad_norm": 0.054634734988212585,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 3.974881887435913,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656195462478184,
+ "grad_norm": 0.054601747542619705,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 3.9746241569519043,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670157068062828,
+ "grad_norm": 0.052188221365213394,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 3.957106113433838,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68411867364747,
+ "grad_norm": 0.055673446506261826,
+ "learning_rate": 0.000322712903929477,
+ "loss": 3.9803967475891113,
+ "step": 1777
+ },
+ {
+ "epoch": 24.698080279232112,
+ "grad_norm": 0.05397328734397888,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 3.967238426208496,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712041884816752,
+ "grad_norm": 0.05384642630815506,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 3.949392795562744,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726003490401396,
+ "grad_norm": 0.05452995374798775,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.9432482719421387,
+ "step": 1780
+ },
+ {
+ "epoch": 24.739965095986037,
+ "grad_norm": 0.052430473268032074,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 3.9944653511047363,
+ "step": 1781
+ },
+ {
+ "epoch": 24.75392670157068,
+ "grad_norm": 0.05514240264892578,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 3.97713565826416,
+ "step": 1782
+ },
+ {
+ "epoch": 24.767888307155324,
+ "grad_norm": 0.05189919099211693,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 3.983150005340576,
+ "step": 1783
+ },
+ {
+ "epoch": 24.781849912739965,
+ "grad_norm": 0.05590776726603508,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 3.9548301696777344,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79581151832461,
+ "grad_norm": 0.05564190074801445,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 3.9725170135498047,
+ "step": 1785
+ },
+ {
+ "epoch": 24.80977312390925,
+ "grad_norm": 0.05121326446533203,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 3.9722347259521484,
+ "step": 1786
+ },
+ {
+ "epoch": 24.823734729493893,
+ "grad_norm": 0.05444202199578285,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 3.9634029865264893,
+ "step": 1787
+ },
+ {
+ "epoch": 24.837696335078533,
+ "grad_norm": 0.05242442339658737,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 3.973270893096924,
+ "step": 1788
+ },
+ {
+ "epoch": 24.851657940663177,
+ "grad_norm": 0.05497463420033455,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 3.9851887226104736,
+ "step": 1789
+ },
+ {
+ "epoch": 24.865619546247817,
+ "grad_norm": 0.05482025071978569,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 4.001110076904297,
+ "step": 1790
+ },
+ {
+ "epoch": 24.87958115183246,
+ "grad_norm": 0.054101862013339996,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 3.9782891273498535,
+ "step": 1791
+ },
+ {
+ "epoch": 24.8935427574171,
+ "grad_norm": 0.05538944527506828,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 3.973541498184204,
+ "step": 1792
+ },
+ {
+ "epoch": 24.907504363001745,
+ "grad_norm": 0.0502893440425396,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 3.9875426292419434,
+ "step": 1793
+ },
+ {
+ "epoch": 24.921465968586386,
+ "grad_norm": 0.05576152354478836,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.947035551071167,
+ "step": 1794
+ },
+ {
+ "epoch": 24.93542757417103,
+ "grad_norm": 0.05356278270483017,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 3.9683666229248047,
+ "step": 1795
+ },
+ {
+ "epoch": 24.949389179755673,
+ "grad_norm": 0.05377693101763725,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 3.9973440170288086,
+ "step": 1796
+ },
+ {
+ "epoch": 24.963350785340314,
+ "grad_norm": 0.05334382876753807,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 3.9843404293060303,
+ "step": 1797
+ },
+ {
+ "epoch": 24.977312390924958,
+ "grad_norm": 0.05514024198055267,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 3.9920568466186523,
+ "step": 1798
+ },
+ {
+ "epoch": 24.991273996509598,
+ "grad_norm": 0.05631054937839508,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 4.005859375,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.03653077408671379,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 2.503864049911499,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 0.5995591878890991,
+ "eval_runtime": 59.5213,
+ "eval_samples_per_second": 41.027,
+ "eval_steps_per_second": 0.655,
+ "step": 1800
+ },
+ {
+ "epoch": 25.013961605584644,
+ "grad_norm": 0.0576021745800972,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 3.894834518432617,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027923211169284,
+ "grad_norm": 0.05842140316963196,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 3.8879270553588867,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041884816753928,
+ "grad_norm": 0.051821205765008926,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 3.9011340141296387,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05584642233857,
+ "grad_norm": 0.053872160613536835,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 3.877183437347412,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069808027923212,
+ "grad_norm": 0.05487574636936188,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 3.9126815795898438,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083769633507853,
+ "grad_norm": 0.054469428956508636,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 3.894768238067627,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097731239092496,
+ "grad_norm": 0.05124446377158165,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 3.916809320449829,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111692844677137,
+ "grad_norm": 0.05193263292312622,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 3.917140483856201,
+ "step": 1808
+ },
+ {
+ "epoch": 25.12565445026178,
+ "grad_norm": 0.0519125871360302,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 3.903704881668091,
+ "step": 1809
+ },
+ {
+ "epoch": 25.13961605584642,
+ "grad_norm": 0.052205223590135574,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 3.892639636993408,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153577661431065,
+ "grad_norm": 0.052036914974451065,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 3.895559787750244,
+ "step": 1811
+ },
+ {
+ "epoch": 25.167539267015705,
+ "grad_norm": 0.054616156965494156,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 3.8976142406463623,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18150087260035,
+ "grad_norm": 0.05603514611721039,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 3.914311170578003,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195462478184993,
+ "grad_norm": 0.05170242115855217,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 3.9141643047332764,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209424083769633,
+ "grad_norm": 0.05476640164852142,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 3.8952436447143555,
+ "step": 1815
+ },
+ {
+ "epoch": 25.223385689354277,
+ "grad_norm": 0.055114105343818665,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 3.872978687286377,
+ "step": 1816
+ },
+ {
+ "epoch": 25.237347294938917,
+ "grad_norm": 0.051593828946352005,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 3.893911600112915,
+ "step": 1817
+ },
+ {
+ "epoch": 25.25130890052356,
+ "grad_norm": 0.05363611876964569,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 3.9062986373901367,
+ "step": 1818
+ },
+ {
+ "epoch": 25.2652705061082,
+ "grad_norm": 0.054656051099300385,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 3.9145383834838867,
+ "step": 1819
+ },
+ {
+ "epoch": 25.279232111692846,
+ "grad_norm": 0.05251947045326233,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 3.912965774536133,
+ "step": 1820
+ },
+ {
+ "epoch": 25.293193717277486,
+ "grad_norm": 0.05575991049408913,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 3.901510238647461,
+ "step": 1821
+ },
+ {
+ "epoch": 25.30715532286213,
+ "grad_norm": 0.05540676787495613,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 3.8659887313842773,
+ "step": 1822
+ },
+ {
+ "epoch": 25.32111692844677,
+ "grad_norm": 0.0577131025493145,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 3.884077548980713,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335078534031414,
+ "grad_norm": 0.055947501212358475,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 3.9274377822875977,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349040139616054,
+ "grad_norm": 0.05328478664159775,
+ "learning_rate": 0.000309930621834911,
+ "loss": 3.8783111572265625,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363001745200698,
+ "grad_norm": 0.057696059346199036,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 3.903721570968628,
+ "step": 1826
+ },
+ {
+ "epoch": 25.376963350785342,
+ "grad_norm": 0.05717543140053749,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 3.9259400367736816,
+ "step": 1827
+ },
+ {
+ "epoch": 25.390924956369982,
+ "grad_norm": 0.05328287556767464,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 3.9450511932373047,
+ "step": 1828
+ },
+ {
+ "epoch": 25.404886561954626,
+ "grad_norm": 0.05414140224456787,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 3.9231765270233154,
+ "step": 1829
+ },
+ {
+ "epoch": 25.418848167539267,
+ "grad_norm": 0.05535200983285904,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 3.8928630352020264,
+ "step": 1830
+ },
+ {
+ "epoch": 25.43280977312391,
+ "grad_norm": 0.05468215420842171,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 3.933401584625244,
+ "step": 1831
+ },
+ {
+ "epoch": 25.44677137870855,
+ "grad_norm": 0.056624677032232285,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 3.922624111175537,
+ "step": 1832
+ },
+ {
+ "epoch": 25.460732984293195,
+ "grad_norm": 0.0545354001224041,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 3.916515827178955,
+ "step": 1833
+ },
+ {
+ "epoch": 25.474694589877835,
+ "grad_norm": 0.053979165852069855,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 3.891812324523926,
+ "step": 1834
+ },
+ {
+ "epoch": 25.48865619546248,
+ "grad_norm": 0.05619875714182854,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 3.959524631500244,
+ "step": 1835
+ },
+ {
+ "epoch": 25.50261780104712,
+ "grad_norm": 0.053561773151159286,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 3.9109320640563965,
+ "step": 1836
+ },
+ {
+ "epoch": 25.516579406631763,
+ "grad_norm": 0.05566735938191414,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 3.910492420196533,
+ "step": 1837
+ },
+ {
+ "epoch": 25.530541012216403,
+ "grad_norm": 0.05757030472159386,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 3.9716763496398926,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544502617801047,
+ "grad_norm": 0.05121897906064987,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 3.9114646911621094,
+ "step": 1839
+ },
+ {
+ "epoch": 25.55846422338569,
+ "grad_norm": 0.05664774775505066,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 3.9465560913085938,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57242582897033,
+ "grad_norm": 0.0569215826690197,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 3.908374309539795,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586387434554975,
+ "grad_norm": 0.05345458909869194,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 3.9336414337158203,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600349040139616,
+ "grad_norm": 0.0565933920443058,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 3.9478759765625,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61431064572426,
+ "grad_norm": 0.05887870490550995,
+ "learning_rate": 0.000304866093757771,
+ "loss": 3.9018349647521973,
+ "step": 1844
+ },
+ {
+ "epoch": 25.6282722513089,
+ "grad_norm": 0.05515308678150177,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 3.9446825981140137,
+ "step": 1845
+ },
+ {
+ "epoch": 25.642233856893544,
+ "grad_norm": 0.058602213859558105,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 3.9497885704040527,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656195462478184,
+ "grad_norm": 0.053806282579898834,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 3.9247193336486816,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670157068062828,
+ "grad_norm": 0.05495952442288399,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 3.912296772003174,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68411867364747,
+ "grad_norm": 0.05479473993182182,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 3.9471306800842285,
+ "step": 1849
+ },
+ {
+ "epoch": 25.698080279232112,
+ "grad_norm": 0.053319111466407776,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 3.955942153930664,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712041884816752,
+ "grad_norm": 0.05832696333527565,
+ "learning_rate": 0.000303,
+ "loss": 3.959169626235962,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726003490401396,
+ "grad_norm": 0.057019371539354324,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 3.9682188034057617,
+ "step": 1852
+ },
+ {
+ "epoch": 25.739965095986037,
+ "grad_norm": 0.05099530518054962,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 3.921377658843994,
+ "step": 1853
+ },
+ {
+ "epoch": 25.75392670157068,
+ "grad_norm": 0.05466439574956894,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 3.946341037750244,
+ "step": 1854
+ },
+ {
+ "epoch": 25.767888307155324,
+ "grad_norm": 0.05742308124899864,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 3.948031425476074,
+ "step": 1855
+ },
+ {
+ "epoch": 25.781849912739965,
+ "grad_norm": 0.057581741362810135,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 3.91162109375,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79581151832461,
+ "grad_norm": 0.05520833283662796,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 3.937495231628418,
+ "step": 1857
+ },
+ {
+ "epoch": 25.80977312390925,
+ "grad_norm": 0.05763327330350876,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 3.9565534591674805,
+ "step": 1858
+ },
+ {
+ "epoch": 25.823734729493893,
+ "grad_norm": 0.0544707253575325,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 3.965190887451172,
+ "step": 1859
+ },
+ {
+ "epoch": 25.837696335078533,
+ "grad_norm": 0.05445104092359543,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 3.938063859939575,
+ "step": 1860
+ },
+ {
+ "epoch": 25.851657940663177,
+ "grad_norm": 0.05505113676190376,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 3.928392171859741,
+ "step": 1861
+ },
+ {
+ "epoch": 25.865619546247817,
+ "grad_norm": 0.05574212223291397,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 3.959397077560425,
+ "step": 1862
+ },
+ {
+ "epoch": 25.87958115183246,
+ "grad_norm": 0.053950466215610504,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 3.9532437324523926,
+ "step": 1863
+ },
+ {
+ "epoch": 25.8935427574171,
+ "grad_norm": 0.05495546758174896,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 3.946746826171875,
+ "step": 1864
+ },
+ {
+ "epoch": 25.907504363001745,
+ "grad_norm": 0.05328625068068504,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 3.9494776725769043,
+ "step": 1865
+ },
+ {
+ "epoch": 25.921465968586386,
+ "grad_norm": 0.05593104660511017,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 3.959989547729492,
+ "step": 1866
+ },
+ {
+ "epoch": 25.93542757417103,
+ "grad_norm": 0.05442969873547554,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 3.952371597290039,
+ "step": 1867
+ },
+ {
+ "epoch": 25.949389179755673,
+ "grad_norm": 0.05446415767073631,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 3.9538888931274414,
+ "step": 1868
+ },
+ {
+ "epoch": 25.963350785340314,
+ "grad_norm": 0.054761797189712524,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 3.9386215209960938,
+ "step": 1869
+ },
+ {
+ "epoch": 25.977312390924958,
+ "grad_norm": 0.054201845079660416,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 3.9252963066101074,
+ "step": 1870
+ },
+ {
+ "epoch": 25.991273996509598,
+ "grad_norm": 0.05364517495036125,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 3.9733197689056396,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.038914378732442856,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 2.4261250495910645,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 0.6009622812271118,
+ "eval_runtime": 60.6864,
+ "eval_samples_per_second": 40.24,
+ "eval_steps_per_second": 0.643,
+ "step": 1872
+ },
+ {
+ "epoch": 26.013961605584644,
+ "grad_norm": 0.05345558747649193,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 3.8902673721313477,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027923211169284,
+ "grad_norm": 0.05672457069158554,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 3.8676247596740723,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041884816753928,
+ "grad_norm": 0.051981374621391296,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 3.872048854827881,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05584642233857,
+ "grad_norm": 0.05262405797839165,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 3.8582329750061035,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069808027923212,
+ "grad_norm": 0.054473262280225754,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 3.8352112770080566,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083769633507853,
+ "grad_norm": 0.053167302161455154,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 3.881760835647583,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097731239092496,
+ "grad_norm": 0.053720805794000626,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 3.891775608062744,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111692844677137,
+ "grad_norm": 0.055102936923503876,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 3.8872156143188477,
+ "step": 1880
+ },
+ {
+ "epoch": 26.12565445026178,
+ "grad_norm": 0.05108853057026863,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 3.8625407218933105,
+ "step": 1881
+ },
+ {
+ "epoch": 26.13961605584642,
+ "grad_norm": 0.05526706948876381,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 3.886127233505249,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153577661431065,
+ "grad_norm": 0.05458226427435875,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 3.8722105026245117,
+ "step": 1883
+ },
+ {
+ "epoch": 26.167539267015705,
+ "grad_norm": 0.05342113599181175,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 3.853677749633789,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18150087260035,
+ "grad_norm": 0.053665641695261,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 3.849315881729126,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195462478184993,
+ "grad_norm": 0.054922688752412796,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 3.8599777221679688,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209424083769633,
+ "grad_norm": 0.05412537604570389,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 3.8767571449279785,
+ "step": 1887
+ },
+ {
+ "epoch": 26.223385689354277,
+ "grad_norm": 0.05467838793992996,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 3.8436965942382812,
+ "step": 1888
+ },
+ {
+ "epoch": 26.237347294938917,
+ "grad_norm": 0.054886411875486374,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 3.8796300888061523,
+ "step": 1889
+ },
+ {
+ "epoch": 26.25130890052356,
+ "grad_norm": 0.05437029153108597,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 3.858133316040039,
+ "step": 1890
+ },
+ {
+ "epoch": 26.2652705061082,
+ "grad_norm": 0.05273836478590965,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 3.839015483856201,
+ "step": 1891
+ },
+ {
+ "epoch": 26.279232111692846,
+ "grad_norm": 0.0546451173722744,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 3.8723373413085938,
+ "step": 1892
+ },
+ {
+ "epoch": 26.293193717277486,
+ "grad_norm": 0.05240490287542343,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 3.877549886703491,
+ "step": 1893
+ },
+ {
+ "epoch": 26.30715532286213,
+ "grad_norm": 0.05420667678117752,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 3.8643171787261963,
+ "step": 1894
+ },
+ {
+ "epoch": 26.32111692844677,
+ "grad_norm": 0.053440142422914505,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 3.8681447505950928,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335078534031414,
+ "grad_norm": 0.052396632730960846,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 3.8691887855529785,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349040139616054,
+ "grad_norm": 0.05318925902247429,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 3.897003650665283,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363001745200698,
+ "grad_norm": 0.05328533053398132,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 3.9079437255859375,
+ "step": 1898
+ },
+ {
+ "epoch": 26.376963350785342,
+ "grad_norm": 0.05196928232908249,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 3.8392016887664795,
+ "step": 1899
+ },
+ {
+ "epoch": 26.390924956369982,
+ "grad_norm": 0.053326766937971115,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 3.8930301666259766,
+ "step": 1900
+ },
+ {
+ "epoch": 26.404886561954626,
+ "grad_norm": 0.052926234900951385,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 3.866997718811035,
+ "step": 1901
+ },
+ {
+ "epoch": 26.418848167539267,
+ "grad_norm": 0.05289287120103836,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 3.896690607070923,
+ "step": 1902
+ },
+ {
+ "epoch": 26.43280977312391,
+ "grad_norm": 0.052162300795316696,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 3.8642401695251465,
+ "step": 1903
+ },
+ {
+ "epoch": 26.44677137870855,
+ "grad_norm": 0.05495629832148552,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 3.8982696533203125,
+ "step": 1904
+ },
+ {
+ "epoch": 26.460732984293195,
+ "grad_norm": 0.055102866142988205,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 3.8725972175598145,
+ "step": 1905
+ },
+ {
+ "epoch": 26.474694589877835,
+ "grad_norm": 0.05537770688533783,
+ "learning_rate": 0.000288343693342466,
+ "loss": 3.9050111770629883,
+ "step": 1906
+ },
+ {
+ "epoch": 26.48865619546248,
+ "grad_norm": 0.05490661785006523,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 3.90586256980896,
+ "step": 1907
+ },
+ {
+ "epoch": 26.50261780104712,
+ "grad_norm": 0.054393451660871506,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 3.8841705322265625,
+ "step": 1908
+ },
+ {
+ "epoch": 26.516579406631763,
+ "grad_norm": 0.0553058385848999,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 3.883500576019287,
+ "step": 1909
+ },
+ {
+ "epoch": 26.530541012216403,
+ "grad_norm": 0.053976885974407196,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 3.894948959350586,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544502617801047,
+ "grad_norm": 0.05290692299604416,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 3.8880579471588135,
+ "step": 1911
+ },
+ {
+ "epoch": 26.55846422338569,
+ "grad_norm": 0.05570613220334053,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 3.9035487174987793,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57242582897033,
+ "grad_norm": 0.054136309772729874,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 3.9137563705444336,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586387434554975,
+ "grad_norm": 0.05328850820660591,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 3.87838077545166,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600349040139616,
+ "grad_norm": 0.05169905349612236,
+ "learning_rate": 0.000285947841605349,
+ "loss": 3.882963180541992,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61431064572426,
+ "grad_norm": 0.05304642394185066,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 3.9250073432922363,
+ "step": 1916
+ },
+ {
+ "epoch": 26.6282722513089,
+ "grad_norm": 0.051766350865364075,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 3.8915140628814697,
+ "step": 1917
+ },
+ {
+ "epoch": 26.642233856893544,
+ "grad_norm": 0.054485417902469635,
+ "learning_rate": 0.000285149463934261,
+ "loss": 3.8973851203918457,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656195462478184,
+ "grad_norm": 0.05437815189361572,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 3.9054274559020996,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670157068062828,
+ "grad_norm": 0.05402572825551033,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 3.9060683250427246,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68411867364747,
+ "grad_norm": 0.05332810431718826,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 3.9053144454956055,
+ "step": 1921
+ },
+ {
+ "epoch": 26.698080279232112,
+ "grad_norm": 0.0557575598359108,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 3.9026694297790527,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712041884816752,
+ "grad_norm": 0.05650033429265022,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 3.8795413970947266,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726003490401396,
+ "grad_norm": 0.054456304758787155,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 3.932950496673584,
+ "step": 1924
+ },
+ {
+ "epoch": 26.739965095986037,
+ "grad_norm": 0.05676668882369995,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 3.878704071044922,
+ "step": 1925
+ },
+ {
+ "epoch": 26.75392670157068,
+ "grad_norm": 0.05875959247350693,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 3.920717239379883,
+ "step": 1926
+ },
+ {
+ "epoch": 26.767888307155324,
+ "grad_norm": 0.054490379989147186,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 3.912592887878418,
+ "step": 1927
+ },
+ {
+ "epoch": 26.781849912739965,
+ "grad_norm": 0.055268481373786926,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 3.900418758392334,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79581151832461,
+ "grad_norm": 0.05617086961865425,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 3.899095296859741,
+ "step": 1929
+ },
+ {
+ "epoch": 26.80977312390925,
+ "grad_norm": 0.05310523882508278,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 3.917501449584961,
+ "step": 1930
+ },
+ {
+ "epoch": 26.823734729493893,
+ "grad_norm": 0.05476578325033188,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 3.867847442626953,
+ "step": 1931
+ },
+ {
+ "epoch": 26.837696335078533,
+ "grad_norm": 0.05456431582570076,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 3.8760366439819336,
+ "step": 1932
+ },
+ {
+ "epoch": 26.851657940663177,
+ "grad_norm": 0.055269334465265274,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 3.910797119140625,
+ "step": 1933
+ },
+ {
+ "epoch": 26.865619546247817,
+ "grad_norm": 0.054999858140945435,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 3.946951389312744,
+ "step": 1934
+ },
+ {
+ "epoch": 26.87958115183246,
+ "grad_norm": 0.05520344525575638,
+ "learning_rate": 0.000280627938758204,
+ "loss": 3.9161229133605957,
+ "step": 1935
+ },
+ {
+ "epoch": 26.8935427574171,
+ "grad_norm": 0.05451727285981178,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 3.908015727996826,
+ "step": 1936
+ },
+ {
+ "epoch": 26.907504363001745,
+ "grad_norm": 0.05561735853552818,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 3.944608449935913,
+ "step": 1937
+ },
+ {
+ "epoch": 26.921465968586386,
+ "grad_norm": 0.054743457585573196,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 3.906559467315674,
+ "step": 1938
+ },
+ {
+ "epoch": 26.93542757417103,
+ "grad_norm": 0.05497661978006363,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 3.900981903076172,
+ "step": 1939
+ },
+ {
+ "epoch": 26.949389179755673,
+ "grad_norm": 0.05452290549874306,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 3.910825252532959,
+ "step": 1940
+ },
+ {
+ "epoch": 26.963350785340314,
+ "grad_norm": 0.05409001186490059,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 3.921509265899658,
+ "step": 1941
+ },
+ {
+ "epoch": 26.977312390924958,
+ "grad_norm": 0.05467002093791962,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 3.9039864540100098,
+ "step": 1942
+ },
+ {
+ "epoch": 26.991273996509598,
+ "grad_norm": 0.05442725867033005,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 3.9153316020965576,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.039271868765354156,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 2.451029062271118,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 0.6021912097930908,
+ "eval_runtime": 61.3232,
+ "eval_samples_per_second": 39.822,
+ "eval_steps_per_second": 0.636,
+ "step": 1944
+ },
+ {
+ "epoch": 27.013961605584644,
+ "grad_norm": 0.05251244455575943,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 3.8143270015716553,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027923211169284,
+ "grad_norm": 0.056699808686971664,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 3.8331453800201416,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041884816753928,
+ "grad_norm": 0.055446673184633255,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 3.833841323852539,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05584642233857,
+ "grad_norm": 0.053186994045972824,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 3.82643985748291,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069808027923212,
+ "grad_norm": 0.05483868718147278,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 3.8209316730499268,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083769633507853,
+ "grad_norm": 0.05387664586305618,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 3.8402364253997803,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097731239092496,
+ "grad_norm": 0.05422234162688255,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 3.8125908374786377,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111692844677137,
+ "grad_norm": 0.05522442236542702,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 3.830599784851074,
+ "step": 1952
+ },
+ {
+ "epoch": 27.12565445026178,
+ "grad_norm": 0.05163128301501274,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 3.833247184753418,
+ "step": 1953
+ },
+ {
+ "epoch": 27.13961605584642,
+ "grad_norm": 0.053523529320955276,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 3.812887191772461,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153577661431065,
+ "grad_norm": 0.05362863466143608,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 3.8325610160827637,
+ "step": 1955
+ },
+ {
+ "epoch": 27.167539267015705,
+ "grad_norm": 0.054255858063697815,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 3.807152271270752,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18150087260035,
+ "grad_norm": 0.05259740352630615,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 3.843714475631714,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195462478184993,
+ "grad_norm": 0.05457160621881485,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 3.847249984741211,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209424083769633,
+ "grad_norm": 0.05517619848251343,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 3.864114284515381,
+ "step": 1959
+ },
+ {
+ "epoch": 27.223385689354277,
+ "grad_norm": 0.052946750074625015,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 3.8219656944274902,
+ "step": 1960
+ },
+ {
+ "epoch": 27.237347294938917,
+ "grad_norm": 0.053545452654361725,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 3.8502726554870605,
+ "step": 1961
+ },
+ {
+ "epoch": 27.25130890052356,
+ "grad_norm": 0.054657287895679474,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 3.837273120880127,
+ "step": 1962
+ },
+ {
+ "epoch": 27.2652705061082,
+ "grad_norm": 0.05385086312890053,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 3.837573528289795,
+ "step": 1963
+ },
+ {
+ "epoch": 27.279232111692846,
+ "grad_norm": 0.05512048304080963,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 3.842432737350464,
+ "step": 1964
+ },
+ {
+ "epoch": 27.293193717277486,
+ "grad_norm": 0.05286296457052231,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 3.845123767852783,
+ "step": 1965
+ },
+ {
+ "epoch": 27.30715532286213,
+ "grad_norm": 0.05396866053342819,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 3.836874485015869,
+ "step": 1966
+ },
+ {
+ "epoch": 27.32111692844677,
+ "grad_norm": 0.053341615945100784,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 3.8860960006713867,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335078534031414,
+ "grad_norm": 0.0531761609017849,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 3.828686237335205,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349040139616054,
+ "grad_norm": 0.054875023663043976,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 3.83782696723938,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363001745200698,
+ "grad_norm": 0.057710494846105576,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 3.8841819763183594,
+ "step": 1970
+ },
+ {
+ "epoch": 27.376963350785342,
+ "grad_norm": 0.05443769320845604,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 3.856564521789551,
+ "step": 1971
+ },
+ {
+ "epoch": 27.390924956369982,
+ "grad_norm": 0.05571551248431206,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 3.847541570663452,
+ "step": 1972
+ },
+ {
+ "epoch": 27.404886561954626,
+ "grad_norm": 0.05770037695765495,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 3.843390941619873,
+ "step": 1973
+ },
+ {
+ "epoch": 27.418848167539267,
+ "grad_norm": 0.05414504557847977,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 3.8786306381225586,
+ "step": 1974
+ },
+ {
+ "epoch": 27.43280977312391,
+ "grad_norm": 0.058058444410562515,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 3.8680999279022217,
+ "step": 1975
+ },
+ {
+ "epoch": 27.44677137870855,
+ "grad_norm": 0.056312717497348785,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 3.858879566192627,
+ "step": 1976
+ },
+ {
+ "epoch": 27.460732984293195,
+ "grad_norm": 0.055812690407037735,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 3.870741844177246,
+ "step": 1977
+ },
+ {
+ "epoch": 27.474694589877835,
+ "grad_norm": 0.057005349546670914,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 3.8850440979003906,
+ "step": 1978
+ },
+ {
+ "epoch": 27.48865619546248,
+ "grad_norm": 0.055214181542396545,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 3.8485918045043945,
+ "step": 1979
+ },
+ {
+ "epoch": 27.50261780104712,
+ "grad_norm": 0.054300058633089066,
+ "learning_rate": 0.00026868712586269,
+ "loss": 3.8759729862213135,
+ "step": 1980
+ },
+ {
+ "epoch": 27.516579406631763,
+ "grad_norm": 0.05370737239718437,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 3.8498623371124268,
+ "step": 1981
+ },
+ {
+ "epoch": 27.530541012216403,
+ "grad_norm": 0.05481775104999542,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 3.8408432006835938,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544502617801047,
+ "grad_norm": 0.05360998213291168,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 3.8800129890441895,
+ "step": 1983
+ },
+ {
+ "epoch": 27.55846422338569,
+ "grad_norm": 0.05510537698864937,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 3.8329358100891113,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57242582897033,
+ "grad_norm": 0.057950302958488464,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 3.870562791824341,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586387434554975,
+ "grad_norm": 0.0546487458050251,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 3.85231351852417,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600349040139616,
+ "grad_norm": 0.056359753012657166,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 3.8782615661621094,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61431064572426,
+ "grad_norm": 0.05409953370690346,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 3.8209736347198486,
+ "step": 1988
+ },
+ {
+ "epoch": 27.6282722513089,
+ "grad_norm": 0.05549091845750809,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 3.9041504859924316,
+ "step": 1989
+ },
+ {
+ "epoch": 27.642233856893544,
+ "grad_norm": 0.05437365546822548,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 3.8530654907226562,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656195462478184,
+ "grad_norm": 0.055456943809986115,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 3.862886428833008,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670157068062828,
+ "grad_norm": 0.05614335089921951,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 3.867595672607422,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68411867364747,
+ "grad_norm": 0.05374857038259506,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 3.8731391429901123,
+ "step": 1993
+ },
+ {
+ "epoch": 27.698080279232112,
+ "grad_norm": 0.05544714257121086,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 3.837608575820923,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712041884816752,
+ "grad_norm": 0.05495234578847885,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 3.8587915897369385,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726003490401396,
+ "grad_norm": 0.05354582890868187,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 3.8618052005767822,
+ "step": 1996
+ },
+ {
+ "epoch": 27.739965095986037,
+ "grad_norm": 0.05556262657046318,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 3.8630597591400146,
+ "step": 1997
+ },
+ {
+ "epoch": 27.75392670157068,
+ "grad_norm": 0.05576452240347862,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 3.8570914268493652,
+ "step": 1998
+ },
+ {
+ "epoch": 27.767888307155324,
+ "grad_norm": 0.053394027054309845,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 3.868783950805664,
+ "step": 1999
+ },
+ {
+ "epoch": 27.781849912739965,
+ "grad_norm": 0.05506644397974014,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 3.871555805206299,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79581151832461,
+ "grad_norm": 0.054367583245038986,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 3.8487513065338135,
+ "step": 2001
+ },
+ {
+ "epoch": 27.80977312390925,
+ "grad_norm": 0.05482787266373634,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 3.88199520111084,
+ "step": 2002
+ },
+ {
+ "epoch": 27.823734729493893,
+ "grad_norm": 0.05416601523756981,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 3.8876938819885254,
+ "step": 2003
+ },
+ {
+ "epoch": 27.837696335078533,
+ "grad_norm": 0.05612047761678696,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 3.8772940635681152,
+ "step": 2004
+ },
+ {
+ "epoch": 27.851657940663177,
+ "grad_norm": 0.05662519112229347,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 3.8429012298583984,
+ "step": 2005
+ },
+ {
+ "epoch": 27.865619546247817,
+ "grad_norm": 0.055557794868946075,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 3.8712010383605957,
+ "step": 2006
+ },
+ {
+ "epoch": 27.87958115183246,
+ "grad_norm": 0.054365914314985275,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 3.884854793548584,
+ "step": 2007
+ },
+ {
+ "epoch": 27.8935427574171,
+ "grad_norm": 0.05622334033250809,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 3.853909492492676,
+ "step": 2008
+ },
+ {
+ "epoch": 27.907504363001745,
+ "grad_norm": 0.05513857677578926,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 3.888460159301758,
+ "step": 2009
+ },
+ {
+ "epoch": 27.921465968586386,
+ "grad_norm": 0.05537945404648781,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 3.891343593597412,
+ "step": 2010
+ },
+ {
+ "epoch": 27.93542757417103,
+ "grad_norm": 0.05649508908390999,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 3.8807168006896973,
+ "step": 2011
+ },
+ {
+ "epoch": 27.949389179755673,
+ "grad_norm": 0.054261550307273865,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 3.8827662467956543,
+ "step": 2012
+ },
+ {
+ "epoch": 27.963350785340314,
+ "grad_norm": 0.05615224689245224,
+ "learning_rate": 0.00025996500713765,
+ "loss": 3.8720929622650146,
+ "step": 2013
+ },
+ {
+ "epoch": 27.977312390924958,
+ "grad_norm": 0.05571882426738739,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 3.8978285789489746,
+ "step": 2014
+ },
+ {
+ "epoch": 27.991273996509598,
+ "grad_norm": 0.05477411672472954,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 3.875284194946289,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.03992476686835289,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 2.404466152191162,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 0.6039577126502991,
+ "eval_runtime": 59.8223,
+ "eval_samples_per_second": 40.821,
+ "eval_steps_per_second": 0.652,
+ "step": 2016
+ },
+ {
+ "epoch": 28.013961605584644,
+ "grad_norm": 0.05395985022187233,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 3.78012752532959,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027923211169284,
+ "grad_norm": 0.05445432290434837,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 3.794297218322754,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041884816753928,
+ "grad_norm": 0.05048619955778122,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 3.8059635162353516,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05584642233857,
+ "grad_norm": 0.054649483412504196,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 3.8080105781555176,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069808027923212,
+ "grad_norm": 0.054571982473134995,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 3.8171639442443848,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083769633507853,
+ "grad_norm": 0.05476690083742142,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 3.7937488555908203,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097731239092496,
+ "grad_norm": 0.05586248263716698,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 3.804347515106201,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111692844677137,
+ "grad_norm": 0.05466577783226967,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 3.8114442825317383,
+ "step": 2024
+ },
+ {
+ "epoch": 28.12565445026178,
+ "grad_norm": 0.05342315137386322,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 3.8107991218566895,
+ "step": 2025
+ },
+ {
+ "epoch": 28.13961605584642,
+ "grad_norm": 0.05318617820739746,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 3.8390064239501953,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153577661431065,
+ "grad_norm": 0.05371803045272827,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 3.8021721839904785,
+ "step": 2027
+ },
+ {
+ "epoch": 28.167539267015705,
+ "grad_norm": 0.05319817736744881,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 3.844275712966919,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18150087260035,
+ "grad_norm": 0.05202893167734146,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 3.8008670806884766,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195462478184993,
+ "grad_norm": 0.054026514291763306,
+ "learning_rate": 0.000255486047794226,
+ "loss": 3.81990909576416,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209424083769633,
+ "grad_norm": 0.05195610597729683,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 3.8142457008361816,
+ "step": 2031
+ },
+ {
+ "epoch": 28.223385689354277,
+ "grad_norm": 0.05212428793311119,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 3.8139138221740723,
+ "step": 2032
+ },
+ {
+ "epoch": 28.237347294938917,
+ "grad_norm": 0.05503816530108452,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 3.8072667121887207,
+ "step": 2033
+ },
+ {
+ "epoch": 28.25130890052356,
+ "grad_norm": 0.05305452272295952,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 3.8146510124206543,
+ "step": 2034
+ },
+ {
+ "epoch": 28.2652705061082,
+ "grad_norm": 0.05301898717880249,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 3.7936806678771973,
+ "step": 2035
+ },
+ {
+ "epoch": 28.279232111692846,
+ "grad_norm": 0.0542520172894001,
+ "learning_rate": 0.000253907826333243,
+ "loss": 3.7920427322387695,
+ "step": 2036
+ },
+ {
+ "epoch": 28.293193717277486,
+ "grad_norm": 0.05349896103143692,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 3.819732189178467,
+ "step": 2037
+ },
+ {
+ "epoch": 28.30715532286213,
+ "grad_norm": 0.05463109537959099,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 3.804321765899658,
+ "step": 2038
+ },
+ {
+ "epoch": 28.32111692844677,
+ "grad_norm": 0.05445708706974983,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 3.8128511905670166,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335078534031414,
+ "grad_norm": 0.05655837804079056,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 3.798435688018799,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349040139616054,
+ "grad_norm": 0.05543001368641853,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 3.779137134552002,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363001745200698,
+ "grad_norm": 0.05458657816052437,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 3.833167552947998,
+ "step": 2042
+ },
+ {
+ "epoch": 28.376963350785342,
+ "grad_norm": 0.056820500642061234,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 3.8207151889801025,
+ "step": 2043
+ },
+ {
+ "epoch": 28.390924956369982,
+ "grad_norm": 0.05464114993810654,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 3.788215160369873,
+ "step": 2044
+ },
+ {
+ "epoch": 28.404886561954626,
+ "grad_norm": 0.053662411868572235,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 3.8059287071228027,
+ "step": 2045
+ },
+ {
+ "epoch": 28.418848167539267,
+ "grad_norm": 0.055799756199121475,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 3.8050360679626465,
+ "step": 2046
+ },
+ {
+ "epoch": 28.43280977312391,
+ "grad_norm": 0.05412062630057335,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 3.8110790252685547,
+ "step": 2047
+ },
+ {
+ "epoch": 28.44677137870855,
+ "grad_norm": 0.05672532692551613,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 3.841888904571533,
+ "step": 2048
+ },
+ {
+ "epoch": 28.460732984293195,
+ "grad_norm": 0.05622367933392525,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 3.8200669288635254,
+ "step": 2049
+ },
+ {
+ "epoch": 28.474694589877835,
+ "grad_norm": 0.0544997975230217,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 3.8031818866729736,
+ "step": 2050
+ },
+ {
+ "epoch": 28.48865619546248,
+ "grad_norm": 0.058867089450359344,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 3.856843948364258,
+ "step": 2051
+ },
+ {
+ "epoch": 28.50261780104712,
+ "grad_norm": 0.055324845016002655,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 3.807135820388794,
+ "step": 2052
+ },
+ {
+ "epoch": 28.516579406631763,
+ "grad_norm": 0.055414266884326935,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 3.8380215167999268,
+ "step": 2053
+ },
+ {
+ "epoch": 28.530541012216403,
+ "grad_norm": 0.05787660554051399,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 3.831721782684326,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544502617801047,
+ "grad_norm": 0.05551844462752342,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 3.8572306632995605,
+ "step": 2055
+ },
+ {
+ "epoch": 28.55846422338569,
+ "grad_norm": 0.05452537164092064,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 3.840963840484619,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57242582897033,
+ "grad_norm": 0.05530945956707001,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 3.840442419052124,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586387434554975,
+ "grad_norm": 0.05631886050105095,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 3.7851438522338867,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600349040139616,
+ "grad_norm": 0.05489915981888771,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 3.7870893478393555,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61431064572426,
+ "grad_norm": 0.055331867188215256,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 3.8441572189331055,
+ "step": 2060
+ },
+ {
+ "epoch": 28.6282722513089,
+ "grad_norm": 0.05580870062112808,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 3.834148645401001,
+ "step": 2061
+ },
+ {
+ "epoch": 28.642233856893544,
+ "grad_norm": 0.05681769549846649,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 3.8349759578704834,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656195462478184,
+ "grad_norm": 0.05484442040324211,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 3.798792600631714,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670157068062828,
+ "grad_norm": 0.05657925084233284,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 3.853915214538574,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68411867364747,
+ "grad_norm": 0.05419609695672989,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 3.8423995971679688,
+ "step": 2065
+ },
+ {
+ "epoch": 28.698080279232112,
+ "grad_norm": 0.05700303241610527,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 3.822547674179077,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712041884816752,
+ "grad_norm": 0.055846985429525375,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 3.830044984817505,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726003490401396,
+ "grad_norm": 0.05754954740405083,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 3.868316173553467,
+ "step": 2068
+ },
+ {
+ "epoch": 28.739965095986037,
+ "grad_norm": 0.056265417486429214,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 3.8297243118286133,
+ "step": 2069
+ },
+ {
+ "epoch": 28.75392670157068,
+ "grad_norm": 0.05429232493042946,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 3.872836112976074,
+ "step": 2070
+ },
+ {
+ "epoch": 28.767888307155324,
+ "grad_norm": 0.05589764565229416,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 3.838311195373535,
+ "step": 2071
+ },
+ {
+ "epoch": 28.781849912739965,
+ "grad_norm": 0.05460003390908241,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 3.8316597938537598,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79581151832461,
+ "grad_norm": 0.054542504251003265,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 3.816542625427246,
+ "step": 2073
+ },
+ {
+ "epoch": 28.80977312390925,
+ "grad_norm": 0.0549996979534626,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 3.796245574951172,
+ "step": 2074
+ },
+ {
+ "epoch": 28.823734729493893,
+ "grad_norm": 0.05521940439939499,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 3.8112473487854004,
+ "step": 2075
+ },
+ {
+ "epoch": 28.837696335078533,
+ "grad_norm": 0.05690402165055275,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 3.8242125511169434,
+ "step": 2076
+ },
+ {
+ "epoch": 28.851657940663177,
+ "grad_norm": 0.055383820086717606,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 3.84938383102417,
+ "step": 2077
+ },
+ {
+ "epoch": 28.865619546247817,
+ "grad_norm": 0.05612736567854881,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 3.8383970260620117,
+ "step": 2078
+ },
+ {
+ "epoch": 28.87958115183246,
+ "grad_norm": 0.05723752826452255,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 3.828171730041504,
+ "step": 2079
+ },
+ {
+ "epoch": 28.8935427574171,
+ "grad_norm": 0.0576927475631237,
+ "learning_rate": 0.000242380656502223,
+ "loss": 3.8756160736083984,
+ "step": 2080
+ },
+ {
+ "epoch": 28.907504363001745,
+ "grad_norm": 0.05609026178717613,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 3.8703267574310303,
+ "step": 2081
+ },
+ {
+ "epoch": 28.921465968586386,
+ "grad_norm": 0.057038433849811554,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 3.8566956520080566,
+ "step": 2082
+ },
+ {
+ "epoch": 28.93542757417103,
+ "grad_norm": 0.05683749541640282,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 3.8409035205841064,
+ "step": 2083
+ },
+ {
+ "epoch": 28.949389179755673,
+ "grad_norm": 0.05682830885052681,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 3.8435707092285156,
+ "step": 2084
+ },
+ {
+ "epoch": 28.963350785340314,
+ "grad_norm": 0.05645477771759033,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 3.8142364025115967,
+ "step": 2085
+ },
+ {
+ "epoch": 28.977312390924958,
+ "grad_norm": 0.05519694834947586,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 3.830415725708008,
+ "step": 2086
+ },
+ {
+ "epoch": 28.991273996509598,
+ "grad_norm": 0.05785970389842987,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 3.8413705825805664,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.04018386825919151,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 2.3790979385375977,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 0.6047304272651672,
+ "eval_runtime": 61.1049,
+ "eval_samples_per_second": 39.964,
+ "eval_steps_per_second": 0.638,
+ "step": 2088
+ },
+ {
+ "epoch": 29.013961605584644,
+ "grad_norm": 0.058019861578941345,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 3.793783664703369,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027923211169284,
+ "grad_norm": 0.058003105223178864,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 3.783708095550537,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041884816753928,
+ "grad_norm": 0.05485067516565323,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 3.7829298973083496,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05584642233857,
+ "grad_norm": 0.056009069085121155,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 3.7747392654418945,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069808027923212,
+ "grad_norm": 0.05480542406439781,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 3.774707317352295,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083769633507853,
+ "grad_norm": 0.05384726822376251,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 3.75443959236145,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097731239092496,
+ "grad_norm": 0.053818102926015854,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 3.7790050506591797,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111692844677137,
+ "grad_norm": 0.05595344305038452,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 3.74898624420166,
+ "step": 2096
+ },
+ {
+ "epoch": 29.12565445026178,
+ "grad_norm": 0.05394545570015907,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 3.7751331329345703,
+ "step": 2097
+ },
+ {
+ "epoch": 29.13961605584642,
+ "grad_norm": 0.05346408113837242,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 3.7671167850494385,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153577661431065,
+ "grad_norm": 0.0541236512362957,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 3.749098062515259,
+ "step": 2099
+ },
+ {
+ "epoch": 29.167539267015705,
+ "grad_norm": 0.05369208753108978,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 3.7778515815734863,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18150087260035,
+ "grad_norm": 0.055362965911626816,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 3.747697591781616,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195462478184993,
+ "grad_norm": 0.05389911308884621,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 3.7759058475494385,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209424083769633,
+ "grad_norm": 0.05315527692437172,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 3.7832705974578857,
+ "step": 2103
+ },
+ {
+ "epoch": 29.223385689354277,
+ "grad_norm": 0.05377589911222458,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 3.7743210792541504,
+ "step": 2104
+ },
+ {
+ "epoch": 29.237347294938917,
+ "grad_norm": 0.05307800695300102,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 3.7654497623443604,
+ "step": 2105
+ },
+ {
+ "epoch": 29.25130890052356,
+ "grad_norm": 0.056719157844781876,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 3.794179916381836,
+ "step": 2106
+ },
+ {
+ "epoch": 29.2652705061082,
+ "grad_norm": 0.053281862288713455,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 3.7620739936828613,
+ "step": 2107
+ },
+ {
+ "epoch": 29.279232111692846,
+ "grad_norm": 0.05470574274659157,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 3.788421869277954,
+ "step": 2108
+ },
+ {
+ "epoch": 29.293193717277486,
+ "grad_norm": 0.05293853580951691,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 3.7655067443847656,
+ "step": 2109
+ },
+ {
+ "epoch": 29.30715532286213,
+ "grad_norm": 0.05524440482258797,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 3.7701616287231445,
+ "step": 2110
+ },
+ {
+ "epoch": 29.32111692844677,
+ "grad_norm": 0.05508396402001381,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 3.7993319034576416,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335078534031414,
+ "grad_norm": 0.05559603124856949,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 3.7799668312072754,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349040139616054,
+ "grad_norm": 0.05530122295022011,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 3.751032829284668,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363001745200698,
+ "grad_norm": 0.05408896878361702,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 3.7761611938476562,
+ "step": 2114
+ },
+ {
+ "epoch": 29.376963350785342,
+ "grad_norm": 0.05530737340450287,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 3.745713233947754,
+ "step": 2115
+ },
+ {
+ "epoch": 29.390924956369982,
+ "grad_norm": 0.05424542352557182,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 3.8041834831237793,
+ "step": 2116
+ },
+ {
+ "epoch": 29.404886561954626,
+ "grad_norm": 0.054800380021333694,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 3.7995352745056152,
+ "step": 2117
+ },
+ {
+ "epoch": 29.418848167539267,
+ "grad_norm": 0.05685900151729584,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 3.7970776557922363,
+ "step": 2118
+ },
+ {
+ "epoch": 29.43280977312391,
+ "grad_norm": 0.05555475875735283,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 3.7950801849365234,
+ "step": 2119
+ },
+ {
+ "epoch": 29.44677137870855,
+ "grad_norm": 0.05670571327209473,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 3.7967746257781982,
+ "step": 2120
+ },
+ {
+ "epoch": 29.460732984293195,
+ "grad_norm": 0.05598503723740578,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 3.809349298477173,
+ "step": 2121
+ },
+ {
+ "epoch": 29.474694589877835,
+ "grad_norm": 0.05501085892319679,
+ "learning_rate": 0.000231465389734324,
+ "loss": 3.803968906402588,
+ "step": 2122
+ },
+ {
+ "epoch": 29.48865619546248,
+ "grad_norm": 0.05519728362560272,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 3.7615163326263428,
+ "step": 2123
+ },
+ {
+ "epoch": 29.50261780104712,
+ "grad_norm": 0.055866893380880356,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 3.7939207553863525,
+ "step": 2124
+ },
+ {
+ "epoch": 29.516579406631763,
+ "grad_norm": 0.05460986867547035,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 3.7740321159362793,
+ "step": 2125
+ },
+ {
+ "epoch": 29.530541012216403,
+ "grad_norm": 0.055752526968717575,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 3.82424259185791,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544502617801047,
+ "grad_norm": 0.056792002171278,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 3.8006062507629395,
+ "step": 2127
+ },
+ {
+ "epoch": 29.55846422338569,
+ "grad_norm": 0.054986681789159775,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 3.7713897228240967,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57242582897033,
+ "grad_norm": 0.054215673357248306,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 3.8005690574645996,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586387434554975,
+ "grad_norm": 0.05700439214706421,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 3.825161933898926,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600349040139616,
+ "grad_norm": 0.055078551173210144,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 3.786716938018799,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61431064572426,
+ "grad_norm": 0.055569086223840714,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 3.8020172119140625,
+ "step": 2132
+ },
+ {
+ "epoch": 29.6282722513089,
+ "grad_norm": 0.0548536442220211,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 3.8038570880889893,
+ "step": 2133
+ },
+ {
+ "epoch": 29.642233856893544,
+ "grad_norm": 0.05604325234889984,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 3.796006441116333,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656195462478184,
+ "grad_norm": 0.05614005774259567,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 3.7938876152038574,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670157068062828,
+ "grad_norm": 0.05761241167783737,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 3.813568592071533,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68411867364747,
+ "grad_norm": 0.05579913780093193,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 3.826862335205078,
+ "step": 2137
+ },
+ {
+ "epoch": 29.698080279232112,
+ "grad_norm": 0.05784283205866814,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 3.808281421661377,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712041884816752,
+ "grad_norm": 0.05596453696489334,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 3.8119921684265137,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726003490401396,
+ "grad_norm": 0.05699361115694046,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 3.822702407836914,
+ "step": 2140
+ },
+ {
+ "epoch": 29.739965095986037,
+ "grad_norm": 0.05375700071454048,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 3.78023099899292,
+ "step": 2141
+ },
+ {
+ "epoch": 29.75392670157068,
+ "grad_norm": 0.05588206648826599,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 3.791456699371338,
+ "step": 2142
+ },
+ {
+ "epoch": 29.767888307155324,
+ "grad_norm": 0.05579312890768051,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 3.799713134765625,
+ "step": 2143
+ },
+ {
+ "epoch": 29.781849912739965,
+ "grad_norm": 0.05713854730129242,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 3.8211703300476074,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79581151832461,
+ "grad_norm": 0.05645165592432022,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 3.7947046756744385,
+ "step": 2145
+ },
+ {
+ "epoch": 29.80977312390925,
+ "grad_norm": 0.056201402097940445,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 3.811009407043457,
+ "step": 2146
+ },
+ {
+ "epoch": 29.823734729493893,
+ "grad_norm": 0.05659479275345802,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 3.8009982109069824,
+ "step": 2147
+ },
+ {
+ "epoch": 29.837696335078533,
+ "grad_norm": 0.05797034874558449,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 3.7994565963745117,
+ "step": 2148
+ },
+ {
+ "epoch": 29.851657940663177,
+ "grad_norm": 0.05734949931502342,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 3.7696728706359863,
+ "step": 2149
+ },
+ {
+ "epoch": 29.865619546247817,
+ "grad_norm": 0.060396358370780945,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 3.8269689083099365,
+ "step": 2150
+ },
+ {
+ "epoch": 29.87958115183246,
+ "grad_norm": 0.057348065078258514,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 3.7925050258636475,
+ "step": 2151
+ },
+ {
+ "epoch": 29.8935427574171,
+ "grad_norm": 0.058563247323036194,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 3.82887864112854,
+ "step": 2152
+ },
+ {
+ "epoch": 29.907504363001745,
+ "grad_norm": 0.06011585891246796,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 3.8061296939849854,
+ "step": 2153
+ },
+ {
+ "epoch": 29.921465968586386,
+ "grad_norm": 0.058440592139959335,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 3.792586326599121,
+ "step": 2154
+ },
+ {
+ "epoch": 29.93542757417103,
+ "grad_norm": 0.057050157338380814,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 3.7885212898254395,
+ "step": 2155
+ },
+ {
+ "epoch": 29.949389179755673,
+ "grad_norm": 0.05920996889472008,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 3.813566207885742,
+ "step": 2156
+ },
+ {
+ "epoch": 29.963350785340314,
+ "grad_norm": 0.057791322469711304,
+ "learning_rate": 0.00022244633283095,
+ "loss": 3.8399810791015625,
+ "step": 2157
+ },
+ {
+ "epoch": 29.977312390924958,
+ "grad_norm": 0.05644001066684723,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 3.79464054107666,
+ "step": 2158
+ },
+ {
+ "epoch": 29.991273996509598,
+ "grad_norm": 0.06021320819854736,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 3.804086208343506,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.040621865540742874,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 2.386565923690796,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 0.6063579320907593,
+ "eval_runtime": 59.5345,
+ "eval_samples_per_second": 41.018,
+ "eval_steps_per_second": 0.655,
+ "step": 2160
+ },
+ {
+ "epoch": 30.013961605584644,
+ "grad_norm": 0.05961514264345169,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 3.756401777267456,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027923211169284,
+ "grad_norm": 0.057169776409864426,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 3.7327475547790527,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041884816753928,
+ "grad_norm": 0.056084226816892624,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 3.7365498542785645,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05584642233857,
+ "grad_norm": 0.059099674224853516,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 3.7383875846862793,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069808027923212,
+ "grad_norm": 0.05632825568318367,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 3.7642762660980225,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083769633507853,
+ "grad_norm": 0.056937988847494125,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 3.7614428997039795,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097731239092496,
+ "grad_norm": 0.05722609534859657,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 3.7394845485687256,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111692844677137,
+ "grad_norm": 0.056204989552497864,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 3.7473866939544678,
+ "step": 2168
+ },
+ {
+ "epoch": 30.12565445026178,
+ "grad_norm": 0.05640130862593651,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 3.7504067420959473,
+ "step": 2169
+ },
+ {
+ "epoch": 30.13961605584642,
+ "grad_norm": 0.05625637248158455,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 3.754812717437744,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153577661431065,
+ "grad_norm": 0.05596087500452995,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 3.741854429244995,
+ "step": 2171
+ },
+ {
+ "epoch": 30.167539267015705,
+ "grad_norm": 0.054819896817207336,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 3.724705934524536,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18150087260035,
+ "grad_norm": 0.05644344910979271,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 3.7609610557556152,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195462478184993,
+ "grad_norm": 0.05366845801472664,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 3.7557997703552246,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209424083769633,
+ "grad_norm": 0.05517663434147835,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 3.7494120597839355,
+ "step": 2175
+ },
+ {
+ "epoch": 30.223385689354277,
+ "grad_norm": 0.05463553965091705,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 3.7434799671173096,
+ "step": 2176
+ },
+ {
+ "epoch": 30.237347294938917,
+ "grad_norm": 0.05352703109383583,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 3.7435178756713867,
+ "step": 2177
+ },
+ {
+ "epoch": 30.25130890052356,
+ "grad_norm": 0.053156204521656036,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 3.7648849487304688,
+ "step": 2178
+ },
+ {
+ "epoch": 30.2652705061082,
+ "grad_norm": 0.05557673051953316,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 3.746366024017334,
+ "step": 2179
+ },
+ {
+ "epoch": 30.279232111692846,
+ "grad_norm": 0.052989061921834946,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 3.7453460693359375,
+ "step": 2180
+ },
+ {
+ "epoch": 30.293193717277486,
+ "grad_norm": 0.055707018822431564,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 3.800962209701538,
+ "step": 2181
+ },
+ {
+ "epoch": 30.30715532286213,
+ "grad_norm": 0.05560754984617233,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 3.7416906356811523,
+ "step": 2182
+ },
+ {
+ "epoch": 30.32111692844677,
+ "grad_norm": 0.055344413965940475,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 3.7672080993652344,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335078534031414,
+ "grad_norm": 0.05490704998373985,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 3.7664341926574707,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349040139616054,
+ "grad_norm": 0.0578596331179142,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 3.7632641792297363,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363001745200698,
+ "grad_norm": 0.055299609899520874,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 3.7502024173736572,
+ "step": 2186
+ },
+ {
+ "epoch": 30.376963350785342,
+ "grad_norm": 0.055016499012708664,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 3.732128143310547,
+ "step": 2187
+ },
+ {
+ "epoch": 30.390924956369982,
+ "grad_norm": 0.057352419942617416,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 3.723008871078491,
+ "step": 2188
+ },
+ {
+ "epoch": 30.404886561954626,
+ "grad_norm": 0.05596115067601204,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 3.786574363708496,
+ "step": 2189
+ },
+ {
+ "epoch": 30.418848167539267,
+ "grad_norm": 0.05596112087368965,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 3.770382881164551,
+ "step": 2190
+ },
+ {
+ "epoch": 30.43280977312391,
+ "grad_norm": 0.0562058687210083,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 3.7605714797973633,
+ "step": 2191
+ },
+ {
+ "epoch": 30.44677137870855,
+ "grad_norm": 0.056407757103443146,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 3.7485427856445312,
+ "step": 2192
+ },
+ {
+ "epoch": 30.460732984293195,
+ "grad_norm": 0.05596652254462242,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 3.7630670070648193,
+ "step": 2193
+ },
+ {
+ "epoch": 30.474694589877835,
+ "grad_norm": 0.056551866233348846,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 3.7981796264648438,
+ "step": 2194
+ },
+ {
+ "epoch": 30.48865619546248,
+ "grad_norm": 0.057354725897312164,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 3.7544898986816406,
+ "step": 2195
+ },
+ {
+ "epoch": 30.50261780104712,
+ "grad_norm": 0.05791718140244484,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 3.7241907119750977,
+ "step": 2196
+ },
+ {
+ "epoch": 30.516579406631763,
+ "grad_norm": 0.05675847455859184,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 3.794506549835205,
+ "step": 2197
+ },
+ {
+ "epoch": 30.530541012216403,
+ "grad_norm": 0.060349952429533005,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 3.75946044921875,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544502617801047,
+ "grad_norm": 0.05547260120511055,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 3.7278571128845215,
+ "step": 2199
+ },
+ {
+ "epoch": 30.55846422338569,
+ "grad_norm": 0.058400996029376984,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 3.756891965866089,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57242582897033,
+ "grad_norm": 0.057871121913194656,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 3.783863067626953,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586387434554975,
+ "grad_norm": 0.05899912118911743,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 3.7787115573883057,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600349040139616,
+ "grad_norm": 0.056164853274822235,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 3.7599334716796875,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61431064572426,
+ "grad_norm": 0.05775163322687149,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 3.7457070350646973,
+ "step": 2204
+ },
+ {
+ "epoch": 30.6282722513089,
+ "grad_norm": 0.057221125811338425,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 3.7663235664367676,
+ "step": 2205
+ },
+ {
+ "epoch": 30.642233856893544,
+ "grad_norm": 0.05788435414433479,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 3.7676830291748047,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656195462478184,
+ "grad_norm": 0.057100843638181686,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 3.7586870193481445,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670157068062828,
+ "grad_norm": 0.05703200399875641,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 3.7561769485473633,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68411867364747,
+ "grad_norm": 0.05664575845003128,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 3.761331081390381,
+ "step": 2209
+ },
+ {
+ "epoch": 30.698080279232112,
+ "grad_norm": 0.056622982025146484,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 3.737915515899658,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712041884816752,
+ "grad_norm": 0.056481845676898956,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 3.7413077354431152,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726003490401396,
+ "grad_norm": 0.057718146592378616,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 3.751267433166504,
+ "step": 2212
+ },
+ {
+ "epoch": 30.739965095986037,
+ "grad_norm": 0.05772862955927849,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 3.753678321838379,
+ "step": 2213
+ },
+ {
+ "epoch": 30.75392670157068,
+ "grad_norm": 0.056571926921606064,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 3.751033306121826,
+ "step": 2214
+ },
+ {
+ "epoch": 30.767888307155324,
+ "grad_norm": 0.05682666599750519,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 3.763153553009033,
+ "step": 2215
+ },
+ {
+ "epoch": 30.781849912739965,
+ "grad_norm": 0.05993897467851639,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 3.787890911102295,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79581151832461,
+ "grad_norm": 0.05659458413720131,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 3.7655086517333984,
+ "step": 2217
+ },
+ {
+ "epoch": 30.80977312390925,
+ "grad_norm": 0.057996224611997604,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 3.774881601333618,
+ "step": 2218
+ },
+ {
+ "epoch": 30.823734729493893,
+ "grad_norm": 0.05677317455410957,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 3.7600479125976562,
+ "step": 2219
+ },
+ {
+ "epoch": 30.837696335078533,
+ "grad_norm": 0.05664737895131111,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 3.787637233734131,
+ "step": 2220
+ },
+ {
+ "epoch": 30.851657940663177,
+ "grad_norm": 0.05808616429567337,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 3.7529220581054688,
+ "step": 2221
+ },
+ {
+ "epoch": 30.865619546247817,
+ "grad_norm": 0.05666997656226158,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 3.7790794372558594,
+ "step": 2222
+ },
+ {
+ "epoch": 30.87958115183246,
+ "grad_norm": 0.05963682755827904,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 3.762615203857422,
+ "step": 2223
+ },
+ {
+ "epoch": 30.8935427574171,
+ "grad_norm": 0.05558215081691742,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 3.7871851921081543,
+ "step": 2224
+ },
+ {
+ "epoch": 30.907504363001745,
+ "grad_norm": 0.05762261897325516,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 3.7605397701263428,
+ "step": 2225
+ },
+ {
+ "epoch": 30.921465968586386,
+ "grad_norm": 0.0565858818590641,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 3.7693653106689453,
+ "step": 2226
+ },
+ {
+ "epoch": 30.93542757417103,
+ "grad_norm": 0.05821581557393074,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 3.766758918762207,
+ "step": 2227
+ },
+ {
+ "epoch": 30.949389179755673,
+ "grad_norm": 0.055186714977025986,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 3.79498553276062,
+ "step": 2228
+ },
+ {
+ "epoch": 30.963350785340314,
+ "grad_norm": 0.05850847810506821,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 3.799769163131714,
+ "step": 2229
+ },
+ {
+ "epoch": 30.977312390924958,
+ "grad_norm": 0.05700397863984108,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 3.7555625438690186,
+ "step": 2230
+ },
+ {
+ "epoch": 30.991273996509598,
+ "grad_norm": 0.057813212275505066,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 3.777778148651123,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.041730619966983795,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 2.3564491271972656,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 0.6080329418182373,
+ "eval_runtime": 60.4607,
+ "eval_samples_per_second": 40.39,
+ "eval_steps_per_second": 0.645,
+ "step": 2232
+ },
+ {
+ "epoch": 31.013961605584644,
+ "grad_norm": 0.0563117116689682,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 3.719982624053955,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027923211169284,
+ "grad_norm": 0.05564197897911072,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 3.719735622406006,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041884816753928,
+ "grad_norm": 0.05562874674797058,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 3.739713430404663,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05584642233857,
+ "grad_norm": 0.057146716862916946,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 3.736401081085205,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069808027923212,
+ "grad_norm": 0.053163520991802216,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 3.7049455642700195,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083769633507853,
+ "grad_norm": 0.057067256420850754,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 3.723662853240967,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097731239092496,
+ "grad_norm": 0.05276816710829735,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 3.710062026977539,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111692844677137,
+ "grad_norm": 0.055126190185546875,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 3.696173667907715,
+ "step": 2240
+ },
+ {
+ "epoch": 31.12565445026178,
+ "grad_norm": 0.05533793941140175,
+ "learning_rate": 0.000201141724176723,
+ "loss": 3.710416793823242,
+ "step": 2241
+ },
+ {
+ "epoch": 31.13961605584642,
+ "grad_norm": 0.055686868727207184,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 3.7188220024108887,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153577661431065,
+ "grad_norm": 0.055188920348882675,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 3.7026562690734863,
+ "step": 2243
+ },
+ {
+ "epoch": 31.167539267015705,
+ "grad_norm": 0.053627416491508484,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 3.7224509716033936,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18150087260035,
+ "grad_norm": 0.05535290017724037,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 3.710698366165161,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195462478184993,
+ "grad_norm": 0.0528891459107399,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 3.6979012489318848,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209424083769633,
+ "grad_norm": 0.055201176553964615,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 3.7324817180633545,
+ "step": 2247
+ },
+ {
+ "epoch": 31.223385689354277,
+ "grad_norm": 0.05338778346776962,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 3.722564697265625,
+ "step": 2248
+ },
+ {
+ "epoch": 31.237347294938917,
+ "grad_norm": 0.05368843302130699,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 3.7118425369262695,
+ "step": 2249
+ },
+ {
+ "epoch": 31.25130890052356,
+ "grad_norm": 0.052788376808166504,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 3.6875901222229004,
+ "step": 2250
+ },
+ {
+ "epoch": 31.2652705061082,
+ "grad_norm": 0.05427199602127075,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 3.7042768001556396,
+ "step": 2251
+ },
+ {
+ "epoch": 31.279232111692846,
+ "grad_norm": 0.05434224382042885,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 3.7138795852661133,
+ "step": 2252
+ },
+ {
+ "epoch": 31.293193717277486,
+ "grad_norm": 0.054364923387765884,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 3.735142946243286,
+ "step": 2253
+ },
+ {
+ "epoch": 31.30715532286213,
+ "grad_norm": 0.05535430088639259,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 3.7198615074157715,
+ "step": 2254
+ },
+ {
+ "epoch": 31.32111692844677,
+ "grad_norm": 0.05463665351271629,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 3.733370542526245,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335078534031414,
+ "grad_norm": 0.0553361251950264,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 3.702805519104004,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349040139616054,
+ "grad_norm": 0.05423365905880928,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 3.7315120697021484,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363001745200698,
+ "grad_norm": 0.05499394237995148,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 3.731891632080078,
+ "step": 2258
+ },
+ {
+ "epoch": 31.376963350785342,
+ "grad_norm": 0.054968126118183136,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 3.7355899810791016,
+ "step": 2259
+ },
+ {
+ "epoch": 31.390924956369982,
+ "grad_norm": 0.05478367581963539,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 3.7143867015838623,
+ "step": 2260
+ },
+ {
+ "epoch": 31.404886561954626,
+ "grad_norm": 0.05554179847240448,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 3.7227137088775635,
+ "step": 2261
+ },
+ {
+ "epoch": 31.418848167539267,
+ "grad_norm": 0.05399509146809578,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 3.750650405883789,
+ "step": 2262
+ },
+ {
+ "epoch": 31.43280977312391,
+ "grad_norm": 0.056201111525297165,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 3.7478344440460205,
+ "step": 2263
+ },
+ {
+ "epoch": 31.44677137870855,
+ "grad_norm": 0.05437091737985611,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 3.7283718585968018,
+ "step": 2264
+ },
+ {
+ "epoch": 31.460732984293195,
+ "grad_norm": 0.057144906371831894,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 3.7482407093048096,
+ "step": 2265
+ },
+ {
+ "epoch": 31.474694589877835,
+ "grad_norm": 0.05441201478242874,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 3.726378917694092,
+ "step": 2266
+ },
+ {
+ "epoch": 31.48865619546248,
+ "grad_norm": 0.056569576263427734,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 3.7211437225341797,
+ "step": 2267
+ },
+ {
+ "epoch": 31.50261780104712,
+ "grad_norm": 0.05561831593513489,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 3.7278850078582764,
+ "step": 2268
+ },
+ {
+ "epoch": 31.516579406631763,
+ "grad_norm": 0.0569809265434742,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 3.713228702545166,
+ "step": 2269
+ },
+ {
+ "epoch": 31.530541012216403,
+ "grad_norm": 0.05654345825314522,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 3.7354423999786377,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544502617801047,
+ "grad_norm": 0.05482077598571777,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 3.7043380737304688,
+ "step": 2271
+ },
+ {
+ "epoch": 31.55846422338569,
+ "grad_norm": 0.055417660623788834,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 3.7199621200561523,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57242582897033,
+ "grad_norm": 0.054656337946653366,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 3.734508514404297,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586387434554975,
+ "grad_norm": 0.05609512701630592,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 3.7341227531433105,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600349040139616,
+ "grad_norm": 0.05602060630917549,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 3.719949960708618,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61431064572426,
+ "grad_norm": 0.05535270646214485,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 3.7124414443969727,
+ "step": 2276
+ },
+ {
+ "epoch": 31.6282722513089,
+ "grad_norm": 0.05577492713928223,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 3.706301212310791,
+ "step": 2277
+ },
+ {
+ "epoch": 31.642233856893544,
+ "grad_norm": 0.057382550090551376,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 3.7411818504333496,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656195462478184,
+ "grad_norm": 0.056383028626441956,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 3.751235008239746,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670157068062828,
+ "grad_norm": 0.05667940899729729,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 3.7379674911499023,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68411867364747,
+ "grad_norm": 0.05635855346918106,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 3.728001594543457,
+ "step": 2281
+ },
+ {
+ "epoch": 31.698080279232112,
+ "grad_norm": 0.05649873986840248,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 3.740757465362549,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712041884816752,
+ "grad_norm": 0.055871494114398956,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 3.7373616695404053,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726003490401396,
+ "grad_norm": 0.057106468826532364,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 3.7572178840637207,
+ "step": 2284
+ },
+ {
+ "epoch": 31.739965095986037,
+ "grad_norm": 0.05788225308060646,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 3.7474923133850098,
+ "step": 2285
+ },
+ {
+ "epoch": 31.75392670157068,
+ "grad_norm": 0.057152729481458664,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 3.7224936485290527,
+ "step": 2286
+ },
+ {
+ "epoch": 31.767888307155324,
+ "grad_norm": 0.0560687892138958,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 3.7336764335632324,
+ "step": 2287
+ },
+ {
+ "epoch": 31.781849912739965,
+ "grad_norm": 0.05890286713838577,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 3.754326343536377,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79581151832461,
+ "grad_norm": 0.058598410338163376,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 3.7265028953552246,
+ "step": 2289
+ },
+ {
+ "epoch": 31.80977312390925,
+ "grad_norm": 0.05728591978549957,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 3.715481758117676,
+ "step": 2290
+ },
+ {
+ "epoch": 31.823734729493893,
+ "grad_norm": 0.05874952673912048,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 3.7139101028442383,
+ "step": 2291
+ },
+ {
+ "epoch": 31.837696335078533,
+ "grad_norm": 0.056882575154304504,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 3.728553533554077,
+ "step": 2292
+ },
+ {
+ "epoch": 31.851657940663177,
+ "grad_norm": 0.057810861617326736,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 3.757908582687378,
+ "step": 2293
+ },
+ {
+ "epoch": 31.865619546247817,
+ "grad_norm": 0.057366516441106796,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 3.7265639305114746,
+ "step": 2294
+ },
+ {
+ "epoch": 31.87958115183246,
+ "grad_norm": 0.059405453503131866,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 3.740511417388916,
+ "step": 2295
+ },
+ {
+ "epoch": 31.8935427574171,
+ "grad_norm": 0.05881514027714729,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 3.751497268676758,
+ "step": 2296
+ },
+ {
+ "epoch": 31.907504363001745,
+ "grad_norm": 0.06030813232064247,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 3.75439715385437,
+ "step": 2297
+ },
+ {
+ "epoch": 31.921465968586386,
+ "grad_norm": 0.06062643602490425,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 3.7215723991394043,
+ "step": 2298
+ },
+ {
+ "epoch": 31.93542757417103,
+ "grad_norm": 0.05953266844153404,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 3.7635340690612793,
+ "step": 2299
+ },
+ {
+ "epoch": 31.949389179755673,
+ "grad_norm": 0.06019049137830734,
+ "learning_rate": 0.000186516746349841,
+ "loss": 3.7750091552734375,
+ "step": 2300
+ },
+ {
+ "epoch": 31.963350785340314,
+ "grad_norm": 0.058755744248628616,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 3.7646124362945557,
+ "step": 2301
+ },
+ {
+ "epoch": 31.977312390924958,
+ "grad_norm": 0.05986425653100014,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 3.7566630840301514,
+ "step": 2302
+ },
+ {
+ "epoch": 31.991273996509598,
+ "grad_norm": 0.057212650775909424,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 3.714637279510498,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.04401736706495285,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 2.33223032951355,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 0.6091130375862122,
+ "eval_runtime": 60.7334,
+ "eval_samples_per_second": 40.209,
+ "eval_steps_per_second": 0.642,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01396160558464,
+ "grad_norm": 0.05650010704994202,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 3.702322244644165,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02792321116929,
+ "grad_norm": 0.054223258048295975,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 3.699272632598877,
+ "step": 2306
+ },
+ {
+ "epoch": 32.04188481675393,
+ "grad_norm": 0.05616949126124382,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 3.6740047931671143,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05584642233857,
+ "grad_norm": 0.05479196831583977,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 3.6942710876464844,
+ "step": 2308
+ },
+ {
+ "epoch": 32.06980802792321,
+ "grad_norm": 0.05735877528786659,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 3.6724581718444824,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083769633507856,
+ "grad_norm": 0.056006450206041336,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 3.6736817359924316,
+ "step": 2310
+ },
+ {
+ "epoch": 32.0977312390925,
+ "grad_norm": 0.055535659193992615,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 3.710303783416748,
+ "step": 2311
+ },
+ {
+ "epoch": 32.11169284467714,
+ "grad_norm": 0.05510629341006279,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 3.669536590576172,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12565445026178,
+ "grad_norm": 0.05426035821437836,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 3.670957088470459,
+ "step": 2313
+ },
+ {
+ "epoch": 32.139616055846425,
+ "grad_norm": 0.05547798424959183,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 3.7050318717956543,
+ "step": 2314
+ },
+ {
+ "epoch": 32.153577661431065,
+ "grad_norm": 0.05412966012954712,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 3.6849310398101807,
+ "step": 2315
+ },
+ {
+ "epoch": 32.167539267015705,
+ "grad_norm": 0.0554192028939724,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 3.702563762664795,
+ "step": 2316
+ },
+ {
+ "epoch": 32.181500872600346,
+ "grad_norm": 0.054564788937568665,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 3.707547187805176,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19546247818499,
+ "grad_norm": 0.05424094945192337,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 3.650768280029297,
+ "step": 2318
+ },
+ {
+ "epoch": 32.20942408376963,
+ "grad_norm": 0.05562194436788559,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 3.6763458251953125,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223385689354274,
+ "grad_norm": 0.053762275725603104,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 3.690493583679199,
+ "step": 2320
+ },
+ {
+ "epoch": 32.23734729493892,
+ "grad_norm": 0.05513778328895569,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 3.6990034580230713,
+ "step": 2321
+ },
+ {
+ "epoch": 32.25130890052356,
+ "grad_norm": 0.05438566580414772,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 3.702357292175293,
+ "step": 2322
+ },
+ {
+ "epoch": 32.2652705061082,
+ "grad_norm": 0.054521508514881134,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 3.701066255569458,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27923211169284,
+ "grad_norm": 0.05416509136557579,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 3.673935651779175,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29319371727749,
+ "grad_norm": 0.05350432172417641,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 3.6740450859069824,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30715532286213,
+ "grad_norm": 0.054736558347940445,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 3.710815668106079,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32111692844677,
+ "grad_norm": 0.055077534168958664,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 3.6732773780822754,
+ "step": 2327
+ },
+ {
+ "epoch": 32.33507853403141,
+ "grad_norm": 0.054749298840761185,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 3.6978728771209717,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34904013961606,
+ "grad_norm": 0.05378924310207367,
+ "learning_rate": 0.000179445406945268,
+ "loss": 3.6783430576324463,
+ "step": 2329
+ },
+ {
+ "epoch": 32.3630017452007,
+ "grad_norm": 0.0553874596953392,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 3.6876044273376465,
+ "step": 2330
+ },
+ {
+ "epoch": 32.37696335078534,
+ "grad_norm": 0.053926583379507065,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 3.712519884109497,
+ "step": 2331
+ },
+ {
+ "epoch": 32.390924956369986,
+ "grad_norm": 0.055778294801712036,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 3.686643123626709,
+ "step": 2332
+ },
+ {
+ "epoch": 32.404886561954626,
+ "grad_norm": 0.05676745995879173,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 3.685737371444702,
+ "step": 2333
+ },
+ {
+ "epoch": 32.41884816753927,
+ "grad_norm": 0.054717857390642166,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 3.6684136390686035,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43280977312391,
+ "grad_norm": 0.0552208311855793,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 3.696580410003662,
+ "step": 2335
+ },
+ {
+ "epoch": 32.446771378708554,
+ "grad_norm": 0.0549275167286396,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 3.7011008262634277,
+ "step": 2336
+ },
+ {
+ "epoch": 32.460732984293195,
+ "grad_norm": 0.05518808960914612,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 3.7204928398132324,
+ "step": 2337
+ },
+ {
+ "epoch": 32.474694589877835,
+ "grad_norm": 0.05621159076690674,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 3.68092679977417,
+ "step": 2338
+ },
+ {
+ "epoch": 32.488656195462475,
+ "grad_norm": 0.054848797619342804,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 3.677091598510742,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50261780104712,
+ "grad_norm": 0.0562039278447628,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 3.7283501625061035,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51657940663176,
+ "grad_norm": 0.05434577167034149,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 3.710252285003662,
+ "step": 2341
+ },
+ {
+ "epoch": 32.5305410122164,
+ "grad_norm": 0.05737415328621864,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 3.7027816772460938,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544502617801044,
+ "grad_norm": 0.057093650102615356,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 3.7027580738067627,
+ "step": 2343
+ },
+ {
+ "epoch": 32.55846422338569,
+ "grad_norm": 0.0550924576818943,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 3.68161678314209,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57242582897033,
+ "grad_norm": 0.057579416781663895,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 3.67604923248291,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58638743455497,
+ "grad_norm": 0.05515696108341217,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 3.693655490875244,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60034904013962,
+ "grad_norm": 0.05788129195570946,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 3.725567102432251,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61431064572426,
+ "grad_norm": 0.05617021024227142,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 3.7027430534362793,
+ "step": 2348
+ },
+ {
+ "epoch": 32.6282722513089,
+ "grad_norm": 0.05649486556649208,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 3.717970371246338,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64223385689354,
+ "grad_norm": 0.05582305043935776,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 3.6795706748962402,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65619546247819,
+ "grad_norm": 0.058177400380373,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 3.6926321983337402,
+ "step": 2351
+ },
+ {
+ "epoch": 32.67015706806283,
+ "grad_norm": 0.05563648045063019,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 3.7227001190185547,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68411867364747,
+ "grad_norm": 0.058799829334020615,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 3.706979513168335,
+ "step": 2353
+ },
+ {
+ "epoch": 32.69808027923211,
+ "grad_norm": 0.05645359680056572,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 3.6872620582580566,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712041884816756,
+ "grad_norm": 0.058064281940460205,
+ "learning_rate": 0.000173176617304673,
+ "loss": 3.695530414581299,
+ "step": 2355
+ },
+ {
+ "epoch": 32.726003490401396,
+ "grad_norm": 0.05756758525967598,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 3.7229819297790527,
+ "step": 2356
+ },
+ {
+ "epoch": 32.73996509598604,
+ "grad_norm": 0.057592105120420456,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 3.7098472118377686,
+ "step": 2357
+ },
+ {
+ "epoch": 32.753926701570684,
+ "grad_norm": 0.05855923518538475,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 3.704477310180664,
+ "step": 2358
+ },
+ {
+ "epoch": 32.767888307155324,
+ "grad_norm": 0.05809807404875755,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 3.7231266498565674,
+ "step": 2359
+ },
+ {
+ "epoch": 32.781849912739965,
+ "grad_norm": 0.05663153529167175,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 3.708165168762207,
+ "step": 2360
+ },
+ {
+ "epoch": 32.795811518324605,
+ "grad_norm": 0.05867684260010719,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 3.70230770111084,
+ "step": 2361
+ },
+ {
+ "epoch": 32.80977312390925,
+ "grad_norm": 0.05850773677229881,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 3.7339577674865723,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82373472949389,
+ "grad_norm": 0.05651238560676575,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 3.7185606956481934,
+ "step": 2363
+ },
+ {
+ "epoch": 32.83769633507853,
+ "grad_norm": 0.05758710578083992,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 3.677966356277466,
+ "step": 2364
+ },
+ {
+ "epoch": 32.85165794066317,
+ "grad_norm": 0.057175327092409134,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 3.720883846282959,
+ "step": 2365
+ },
+ {
+ "epoch": 32.86561954624782,
+ "grad_norm": 0.05690391734242439,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 3.706535577774048,
+ "step": 2366
+ },
+ {
+ "epoch": 32.87958115183246,
+ "grad_norm": 0.05758208781480789,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 3.743523120880127,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8935427574171,
+ "grad_norm": 0.05831100419163704,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 3.726163387298584,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90750436300174,
+ "grad_norm": 0.05664364993572235,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 3.691837787628174,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92146596858639,
+ "grad_norm": 0.05847351998090744,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 3.7386794090270996,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93542757417103,
+ "grad_norm": 0.05673616752028465,
+ "learning_rate": 0.00016935382741164,
+ "loss": 3.72623348236084,
+ "step": 2371
+ },
+ {
+ "epoch": 32.94938917975567,
+ "grad_norm": 0.058967288583517075,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 3.733576536178589,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96335078534032,
+ "grad_norm": 0.056366659700870514,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 3.7006077766418457,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97731239092496,
+ "grad_norm": 0.05691685527563095,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 3.7104835510253906,
+ "step": 2374
+ },
+ {
+ "epoch": 32.9912739965096,
+ "grad_norm": 0.05809671804308891,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 3.7175397872924805,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.0423070453107357,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 2.3232760429382324,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 0.6105995178222656,
+ "eval_runtime": 61.6632,
+ "eval_samples_per_second": 39.602,
+ "eval_steps_per_second": 0.632,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01396160558464,
+ "grad_norm": 0.05560882017016411,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 3.6668033599853516,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02792321116929,
+ "grad_norm": 0.053790636360645294,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 3.6379659175872803,
+ "step": 2378
+ },
+ {
+ "epoch": 33.04188481675393,
+ "grad_norm": 0.056180961430072784,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 3.6418814659118652,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05584642233857,
+ "grad_norm": 0.054904621094465256,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 3.651149034500122,
+ "step": 2380
+ },
+ {
+ "epoch": 33.06980802792321,
+ "grad_norm": 0.05298716947436333,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 3.686598777770996,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083769633507856,
+ "grad_norm": 0.053739335387945175,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 3.6473655700683594,
+ "step": 2382
+ },
+ {
+ "epoch": 33.0977312390925,
+ "grad_norm": 0.05417626351118088,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 3.6908068656921387,
+ "step": 2383
+ },
+ {
+ "epoch": 33.11169284467714,
+ "grad_norm": 0.0533599928021431,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 3.6613330841064453,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12565445026178,
+ "grad_norm": 0.05309661850333214,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 3.6534783840179443,
+ "step": 2385
+ },
+ {
+ "epoch": 33.139616055846425,
+ "grad_norm": 0.05432313680648804,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 3.664217948913574,
+ "step": 2386
+ },
+ {
+ "epoch": 33.153577661431065,
+ "grad_norm": 0.054764073342084885,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 3.660922050476074,
+ "step": 2387
+ },
+ {
+ "epoch": 33.167539267015705,
+ "grad_norm": 0.05396303907036781,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 3.662687063217163,
+ "step": 2388
+ },
+ {
+ "epoch": 33.181500872600346,
+ "grad_norm": 0.05540551245212555,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 3.6671347618103027,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19546247818499,
+ "grad_norm": 0.05510672926902771,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 3.6410186290740967,
+ "step": 2390
+ },
+ {
+ "epoch": 33.20942408376963,
+ "grad_norm": 0.055180903524160385,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 3.653822422027588,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223385689354274,
+ "grad_norm": 0.054320868104696274,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 3.6835601329803467,
+ "step": 2392
+ },
+ {
+ "epoch": 33.23734729493892,
+ "grad_norm": 0.05568838492035866,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 3.671053171157837,
+ "step": 2393
+ },
+ {
+ "epoch": 33.25130890052356,
+ "grad_norm": 0.056060485541820526,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 3.682260751724243,
+ "step": 2394
+ },
+ {
+ "epoch": 33.2652705061082,
+ "grad_norm": 0.054811324924230576,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 3.6742751598358154,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27923211169284,
+ "grad_norm": 0.05570835620164871,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 3.6490235328674316,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29319371727749,
+ "grad_norm": 0.05547381564974785,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 3.6795434951782227,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30715532286213,
+ "grad_norm": 0.05463161692023277,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 3.655003547668457,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32111692844677,
+ "grad_norm": 0.05583210662007332,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 3.680175304412842,
+ "step": 2399
+ },
+ {
+ "epoch": 33.33507853403141,
+ "grad_norm": 0.056169651448726654,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 3.667175054550171,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34904013961606,
+ "grad_norm": 0.055788591504096985,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 3.6673946380615234,
+ "step": 2401
+ },
+ {
+ "epoch": 33.3630017452007,
+ "grad_norm": 0.05556197836995125,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 3.650700569152832,
+ "step": 2402
+ },
+ {
+ "epoch": 33.37696335078534,
+ "grad_norm": 0.05749715119600296,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 3.6977670192718506,
+ "step": 2403
+ },
+ {
+ "epoch": 33.390924956369986,
+ "grad_norm": 0.05672239512205124,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 3.6604905128479004,
+ "step": 2404
+ },
+ {
+ "epoch": 33.404886561954626,
+ "grad_norm": 0.05661430582404137,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 3.64540958404541,
+ "step": 2405
+ },
+ {
+ "epoch": 33.41884816753927,
+ "grad_norm": 0.05520932003855705,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 3.6648027896881104,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43280977312391,
+ "grad_norm": 0.0557708777487278,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 3.6694397926330566,
+ "step": 2407
+ },
+ {
+ "epoch": 33.446771378708554,
+ "grad_norm": 0.05510152503848076,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 3.6723294258117676,
+ "step": 2408
+ },
+ {
+ "epoch": 33.460732984293195,
+ "grad_norm": 0.05714292824268341,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 3.6610422134399414,
+ "step": 2409
+ },
+ {
+ "epoch": 33.474694589877835,
+ "grad_norm": 0.05525747686624527,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 3.655020236968994,
+ "step": 2410
+ },
+ {
+ "epoch": 33.488656195462475,
+ "grad_norm": 0.05610140040516853,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 3.6707541942596436,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50261780104712,
+ "grad_norm": 0.05564311146736145,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 3.699244976043701,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51657940663176,
+ "grad_norm": 0.05607445910573006,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 3.6682703495025635,
+ "step": 2413
+ },
+ {
+ "epoch": 33.5305410122164,
+ "grad_norm": 0.05434789881110191,
+ "learning_rate": 0.000159218843594243,
+ "loss": 3.6799025535583496,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544502617801044,
+ "grad_norm": 0.055795323103666306,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 3.6433563232421875,
+ "step": 2415
+ },
+ {
+ "epoch": 33.55846422338569,
+ "grad_norm": 0.05676727741956711,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 3.665454864501953,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57242582897033,
+ "grad_norm": 0.05657052621245384,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 3.6618518829345703,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58638743455497,
+ "grad_norm": 0.056275662034749985,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 3.6946463584899902,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60034904013962,
+ "grad_norm": 0.056533295661211014,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 3.6695780754089355,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61431064572426,
+ "grad_norm": 0.05692996829748154,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 3.6850693225860596,
+ "step": 2420
+ },
+ {
+ "epoch": 33.6282722513089,
+ "grad_norm": 0.0577891543507576,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 3.668393611907959,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64223385689354,
+ "grad_norm": 0.05549122393131256,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 3.6772003173828125,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65619546247819,
+ "grad_norm": 0.05636308714747429,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 3.6818785667419434,
+ "step": 2423
+ },
+ {
+ "epoch": 33.67015706806283,
+ "grad_norm": 0.05681503564119339,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 3.6955409049987793,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68411867364747,
+ "grad_norm": 0.05746293440461159,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 3.6917951107025146,
+ "step": 2425
+ },
+ {
+ "epoch": 33.69808027923211,
+ "grad_norm": 0.056621894240379333,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 3.635894775390625,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712041884816756,
+ "grad_norm": 0.05611368268728256,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 3.679736375808716,
+ "step": 2427
+ },
+ {
+ "epoch": 33.726003490401396,
+ "grad_norm": 0.057769183069467545,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 3.6751227378845215,
+ "step": 2428
+ },
+ {
+ "epoch": 33.73996509598604,
+ "grad_norm": 0.05695759132504463,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 3.685178756713867,
+ "step": 2429
+ },
+ {
+ "epoch": 33.753926701570684,
+ "grad_norm": 0.05776882544159889,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 3.6355032920837402,
+ "step": 2430
+ },
+ {
+ "epoch": 33.767888307155324,
+ "grad_norm": 0.056043099611997604,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 3.668668270111084,
+ "step": 2431
+ },
+ {
+ "epoch": 33.781849912739965,
+ "grad_norm": 0.05634639039635658,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 3.6788253784179688,
+ "step": 2432
+ },
+ {
+ "epoch": 33.795811518324605,
+ "grad_norm": 0.05703914538025856,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 3.676163911819458,
+ "step": 2433
+ },
+ {
+ "epoch": 33.80977312390925,
+ "grad_norm": 0.05846793204545975,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 3.657705307006836,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82373472949389,
+ "grad_norm": 0.059191398322582245,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 3.70918607711792,
+ "step": 2435
+ },
+ {
+ "epoch": 33.83769633507853,
+ "grad_norm": 0.056207913905382156,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 3.690638542175293,
+ "step": 2436
+ },
+ {
+ "epoch": 33.85165794066317,
+ "grad_norm": 0.0590900182723999,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 3.6839122772216797,
+ "step": 2437
+ },
+ {
+ "epoch": 33.86561954624782,
+ "grad_norm": 0.05649874731898308,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 3.6689414978027344,
+ "step": 2438
+ },
+ {
+ "epoch": 33.87958115183246,
+ "grad_norm": 0.05947998911142349,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 3.653146743774414,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8935427574171,
+ "grad_norm": 0.05769696831703186,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 3.685544967651367,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90750436300174,
+ "grad_norm": 0.05837882682681084,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 3.6648831367492676,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92146596858639,
+ "grad_norm": 0.05676107108592987,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 3.7186074256896973,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93542757417103,
+ "grad_norm": 0.05771078169345856,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 3.678924083709717,
+ "step": 2443
+ },
+ {
+ "epoch": 33.94938917975567,
+ "grad_norm": 0.05686977133154869,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 3.7026078701019287,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96335078534032,
+ "grad_norm": 0.0571015290915966,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 3.6880578994750977,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97731239092496,
+ "grad_norm": 0.056974101811647415,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 3.6521928310394287,
+ "step": 2446
+ },
+ {
+ "epoch": 33.9912739965096,
+ "grad_norm": 0.05962125584483147,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 3.686126708984375,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.043099287897348404,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 2.308753490447998,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 0.6120397448539734,
+ "eval_runtime": 60.5299,
+ "eval_samples_per_second": 40.344,
+ "eval_steps_per_second": 0.644,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01396160558464,
+ "grad_norm": 0.05469869077205658,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 3.6253180503845215,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02792321116929,
+ "grad_norm": 0.05455945432186127,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 3.651854991912842,
+ "step": 2450
+ },
+ {
+ "epoch": 34.04188481675393,
+ "grad_norm": 0.05369517207145691,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 3.630152702331543,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05584642233857,
+ "grad_norm": 0.05498119816184044,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 3.6293938159942627,
+ "step": 2452
+ },
+ {
+ "epoch": 34.06980802792321,
+ "grad_norm": 0.05530959740281105,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 3.6548690795898438,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083769633507856,
+ "grad_norm": 0.055240705609321594,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 3.6408982276916504,
+ "step": 2454
+ },
+ {
+ "epoch": 34.0977312390925,
+ "grad_norm": 0.05358482152223587,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 3.6274750232696533,
+ "step": 2455
+ },
+ {
+ "epoch": 34.11169284467714,
+ "grad_norm": 0.05580627918243408,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 3.6357412338256836,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12565445026178,
+ "grad_norm": 0.05458527058362961,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 3.6637980937957764,
+ "step": 2457
+ },
+ {
+ "epoch": 34.139616055846425,
+ "grad_norm": 0.05548330023884773,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 3.627671718597412,
+ "step": 2458
+ },
+ {
+ "epoch": 34.153577661431065,
+ "grad_norm": 0.05482204258441925,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 3.642627239227295,
+ "step": 2459
+ },
+ {
+ "epoch": 34.167539267015705,
+ "grad_norm": 0.05435148626565933,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 3.6398377418518066,
+ "step": 2460
+ },
+ {
+ "epoch": 34.181500872600346,
+ "grad_norm": 0.05556599423289299,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 3.62994384765625,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19546247818499,
+ "grad_norm": 0.05394179746508598,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 3.593183994293213,
+ "step": 2462
+ },
+ {
+ "epoch": 34.20942408376963,
+ "grad_norm": 0.05472215637564659,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 3.645197868347168,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223385689354274,
+ "grad_norm": 0.05505116656422615,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 3.655001163482666,
+ "step": 2464
+ },
+ {
+ "epoch": 34.23734729493892,
+ "grad_norm": 0.05473910644650459,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 3.636812210083008,
+ "step": 2465
+ },
+ {
+ "epoch": 34.25130890052356,
+ "grad_norm": 0.05537406727671623,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 3.620631217956543,
+ "step": 2466
+ },
+ {
+ "epoch": 34.2652705061082,
+ "grad_norm": 0.054202884435653687,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 3.6500911712646484,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27923211169284,
+ "grad_norm": 0.0536428838968277,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 3.6268038749694824,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29319371727749,
+ "grad_norm": 0.053046368062496185,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 3.646955966949463,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30715532286213,
+ "grad_norm": 0.05279967561364174,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 3.6142497062683105,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32111692844677,
+ "grad_norm": 0.05348867550492287,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 3.643517017364502,
+ "step": 2471
+ },
+ {
+ "epoch": 34.33507853403141,
+ "grad_norm": 0.05406436324119568,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 3.6528172492980957,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34904013961606,
+ "grad_norm": 0.05249491706490517,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 3.6664018630981445,
+ "step": 2473
+ },
+ {
+ "epoch": 34.3630017452007,
+ "grad_norm": 0.053803686052560806,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 3.609346389770508,
+ "step": 2474
+ },
+ {
+ "epoch": 34.37696335078534,
+ "grad_norm": 0.054445188492536545,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 3.6647889614105225,
+ "step": 2475
+ },
+ {
+ "epoch": 34.390924956369986,
+ "grad_norm": 0.0550406239926815,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 3.656118154525757,
+ "step": 2476
+ },
+ {
+ "epoch": 34.404886561954626,
+ "grad_norm": 0.054776549339294434,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 3.6493523120880127,
+ "step": 2477
+ },
+ {
+ "epoch": 34.41884816753927,
+ "grad_norm": 0.054829373955726624,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 3.634632110595703,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43280977312391,
+ "grad_norm": 0.054575446993112564,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 3.636216878890991,
+ "step": 2479
+ },
+ {
+ "epoch": 34.446771378708554,
+ "grad_norm": 0.056146737188100815,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 3.6584529876708984,
+ "step": 2480
+ },
+ {
+ "epoch": 34.460732984293195,
+ "grad_norm": 0.054963648319244385,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 3.6363425254821777,
+ "step": 2481
+ },
+ {
+ "epoch": 34.474694589877835,
+ "grad_norm": 0.05659656599164009,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 3.6401073932647705,
+ "step": 2482
+ },
+ {
+ "epoch": 34.488656195462475,
+ "grad_norm": 0.05448099970817566,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 3.6161837577819824,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50261780104712,
+ "grad_norm": 0.057917334139347076,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 3.6802477836608887,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51657940663176,
+ "grad_norm": 0.056171346455812454,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 3.652339220046997,
+ "step": 2485
+ },
+ {
+ "epoch": 34.5305410122164,
+ "grad_norm": 0.05800231546163559,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 3.6437389850616455,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544502617801044,
+ "grad_norm": 0.05693773180246353,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 3.6268343925476074,
+ "step": 2487
+ },
+ {
+ "epoch": 34.55846422338569,
+ "grad_norm": 0.05424632877111435,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 3.6246471405029297,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57242582897033,
+ "grad_norm": 0.05645301938056946,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 3.649033308029175,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58638743455497,
+ "grad_norm": 0.056913044303655624,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 3.658982515335083,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60034904013962,
+ "grad_norm": 0.05770258232951164,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 3.621516704559326,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61431064572426,
+ "grad_norm": 0.05483509227633476,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 3.659385919570923,
+ "step": 2492
+ },
+ {
+ "epoch": 34.6282722513089,
+ "grad_norm": 0.057591512799263,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 3.6512298583984375,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64223385689354,
+ "grad_norm": 0.055053744465112686,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 3.630310297012329,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65619546247819,
+ "grad_norm": 0.056070517748594284,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 3.6766881942749023,
+ "step": 2495
+ },
+ {
+ "epoch": 34.67015706806283,
+ "grad_norm": 0.056556008756160736,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 3.6454544067382812,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68411867364747,
+ "grad_norm": 0.05624367296695709,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 3.6482372283935547,
+ "step": 2497
+ },
+ {
+ "epoch": 34.69808027923211,
+ "grad_norm": 0.056266799569129944,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 3.6528916358947754,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712041884816756,
+ "grad_norm": 0.05540672317147255,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 3.636371374130249,
+ "step": 2499
+ },
+ {
+ "epoch": 34.726003490401396,
+ "grad_norm": 0.05527781695127487,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 3.6440672874450684,
+ "step": 2500
+ },
+ {
+ "epoch": 34.73996509598604,
+ "grad_norm": 0.05645914003252983,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 3.6747167110443115,
+ "step": 2501
+ },
+ {
+ "epoch": 34.753926701570684,
+ "grad_norm": 0.055584684014320374,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 3.637603282928467,
+ "step": 2502
+ },
+ {
+ "epoch": 34.767888307155324,
+ "grad_norm": 0.0554841123521328,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 3.6546664237976074,
+ "step": 2503
+ },
+ {
+ "epoch": 34.781849912739965,
+ "grad_norm": 0.0557134784758091,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 3.6366615295410156,
+ "step": 2504
+ },
+ {
+ "epoch": 34.795811518324605,
+ "grad_norm": 0.056569263339042664,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 3.657125949859619,
+ "step": 2505
+ },
+ {
+ "epoch": 34.80977312390925,
+ "grad_norm": 0.05737468600273132,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 3.6409783363342285,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82373472949389,
+ "grad_norm": 0.05602499470114708,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 3.6670620441436768,
+ "step": 2507
+ },
+ {
+ "epoch": 34.83769633507853,
+ "grad_norm": 0.059535399079322815,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 3.6585612297058105,
+ "step": 2508
+ },
+ {
+ "epoch": 34.85165794066317,
+ "grad_norm": 0.05610577017068863,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 3.6544179916381836,
+ "step": 2509
+ },
+ {
+ "epoch": 34.86561954624782,
+ "grad_norm": 0.057512786239385605,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 3.658529281616211,
+ "step": 2510
+ },
+ {
+ "epoch": 34.87958115183246,
+ "grad_norm": 0.05605233088135719,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 3.622344493865967,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8935427574171,
+ "grad_norm": 0.055069468915462494,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 3.6812562942504883,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90750436300174,
+ "grad_norm": 0.05756533145904541,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 3.651491165161133,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92146596858639,
+ "grad_norm": 0.05596763640642166,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 3.633408784866333,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93542757417103,
+ "grad_norm": 0.057164326310157776,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 3.632655620574951,
+ "step": 2515
+ },
+ {
+ "epoch": 34.94938917975567,
+ "grad_norm": 0.059363316744565964,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 3.637542486190796,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96335078534032,
+ "grad_norm": 0.05572650581598282,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 3.666663885116577,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97731239092496,
+ "grad_norm": 0.05825378745794296,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 3.654174327850342,
+ "step": 2518
+ },
+ {
+ "epoch": 34.9912739965096,
+ "grad_norm": 0.058226607739925385,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 3.630509376525879,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.041674062609672546,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 2.2687463760375977,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 0.6132010817527771,
+ "eval_runtime": 58.9507,
+ "eval_samples_per_second": 41.424,
+ "eval_steps_per_second": 0.662,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01396160558464,
+ "grad_norm": 0.053901202976703644,
+ "learning_rate": 0.0001349604788260823,
+ "loss": 3.6262593269348145,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02792321116929,
+ "grad_norm": 0.051058750599622726,
+ "learning_rate": 0.00013474073268599152,
+ "loss": 3.596529483795166,
+ "step": 2522
+ },
+ {
+ "epoch": 35.04188481675393,
+ "grad_norm": 0.05482964590191841,
+ "learning_rate": 0.00013452112210935339,
+ "loss": 3.6197128295898438,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05584642233857,
+ "grad_norm": 0.05166293680667877,
+ "learning_rate": 0.00013430164727310399,
+ "loss": 3.6150989532470703,
+ "step": 2524
+ },
+ {
+ "epoch": 35.06980802792321,
+ "grad_norm": 0.051854558289051056,
+ "learning_rate": 0.00013408230835407027,
+ "loss": 3.594954013824463,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083769633507856,
+ "grad_norm": 0.05247301980853081,
+ "learning_rate": 0.00013386310552896977,
+ "loss": 3.6349754333496094,
+ "step": 2526
+ },
+ {
+ "epoch": 35.0977312390925,
+ "grad_norm": 0.05379469692707062,
+ "learning_rate": 0.00013364403897441016,
+ "loss": 3.611581802368164,
+ "step": 2527
+ },
+ {
+ "epoch": 35.11169284467714,
+ "grad_norm": 0.053294215351343155,
+ "learning_rate": 0.00013342510886688944,
+ "loss": 3.5933470726013184,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12565445026178,
+ "grad_norm": 0.05274510383605957,
+ "learning_rate": 0.00013320631538279548,
+ "loss": 3.6055941581726074,
+ "step": 2529
+ },
+ {
+ "epoch": 35.139616055846425,
+ "grad_norm": 0.05449166148900986,
+ "learning_rate": 0.00013298765869840658,
+ "loss": 3.5891900062561035,
+ "step": 2530
+ },
+ {
+ "epoch": 35.153577661431065,
+ "grad_norm": 0.05245492234826088,
+ "learning_rate": 0.00013276913898989013,
+ "loss": 3.614109992980957,
+ "step": 2531
+ },
+ {
+ "epoch": 35.167539267015705,
+ "grad_norm": 0.05323590710759163,
+ "learning_rate": 0.00013255075643330369,
+ "loss": 3.6126725673675537,
+ "step": 2532
+ },
+ {
+ "epoch": 35.181500872600346,
+ "grad_norm": 0.05381206050515175,
+ "learning_rate": 0.00013233251120459407,
+ "loss": 3.5820865631103516,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19546247818499,
+ "grad_norm": 0.054380469024181366,
+ "learning_rate": 0.0001321144034795976,
+ "loss": 3.6311044692993164,
+ "step": 2534
+ },
+ {
+ "epoch": 35.20942408376963,
+ "grad_norm": 0.053128909319639206,
+ "learning_rate": 0.00013189643343403977,
+ "loss": 3.5665154457092285,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223385689354274,
+ "grad_norm": 0.052764274179935455,
+ "learning_rate": 0.00013167860124353476,
+ "loss": 3.609287738800049,
+ "step": 2536
+ },
+ {
+ "epoch": 35.23734729493892,
+ "grad_norm": 0.05363981053233147,
+ "learning_rate": 0.00013146090708358657,
+ "loss": 3.6260581016540527,
+ "step": 2537
+ },
+ {
+ "epoch": 35.25130890052356,
+ "grad_norm": 0.05316983535885811,
+ "learning_rate": 0.00013124335112958704,
+ "loss": 3.589477300643921,
+ "step": 2538
+ },
+ {
+ "epoch": 35.2652705061082,
+ "grad_norm": 0.05451323837041855,
+ "learning_rate": 0.0001310259335568172,
+ "loss": 3.621201515197754,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27923211169284,
+ "grad_norm": 0.053846996277570724,
+ "learning_rate": 0.00013080865454044647,
+ "loss": 3.610521078109741,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29319371727749,
+ "grad_norm": 0.05316411331295967,
+ "learning_rate": 0.00013059151425553263,
+ "loss": 3.6240267753601074,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30715532286213,
+ "grad_norm": 0.053650014102458954,
+ "learning_rate": 0.00013037451287702183,
+ "loss": 3.6098155975341797,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32111692844677,
+ "grad_norm": 0.054066307842731476,
+ "learning_rate": 0.0001301576505797479,
+ "loss": 3.619396686553955,
+ "step": 2543
+ },
+ {
+ "epoch": 35.33507853403141,
+ "grad_norm": 0.05257592722773552,
+ "learning_rate": 0.00012994092753843293,
+ "loss": 3.655679225921631,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34904013961606,
+ "grad_norm": 0.05437186732888222,
+ "learning_rate": 0.00012972434392768687,
+ "loss": 3.6167373657226562,
+ "step": 2545
+ },
+ {
+ "epoch": 35.3630017452007,
+ "grad_norm": 0.05342245474457741,
+ "learning_rate": 0.00012950789992200714,
+ "loss": 3.5970587730407715,
+ "step": 2546
+ },
+ {
+ "epoch": 35.37696335078534,
+ "grad_norm": 0.054295122623443604,
+ "learning_rate": 0.00012929159569577886,
+ "loss": 3.5975115299224854,
+ "step": 2547
+ },
+ {
+ "epoch": 35.390924956369986,
+ "grad_norm": 0.05349617451429367,
+ "learning_rate": 0.0001290754314232743,
+ "loss": 3.603315591812134,
+ "step": 2548
+ },
+ {
+ "epoch": 35.404886561954626,
+ "grad_norm": 0.05614254251122475,
+ "learning_rate": 0.00012885940727865334,
+ "loss": 3.6229186058044434,
+ "step": 2549
+ },
+ {
+ "epoch": 35.41884816753927,
+ "grad_norm": 0.054123859852552414,
+ "learning_rate": 0.00012864352343596247,
+ "loss": 3.641162395477295,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43280977312391,
+ "grad_norm": 0.05414072424173355,
+ "learning_rate": 0.0001284277800691355,
+ "loss": 3.614049196243286,
+ "step": 2551
+ },
+ {
+ "epoch": 35.446771378708554,
+ "grad_norm": 0.05422188341617584,
+ "learning_rate": 0.00012821217735199298,
+ "loss": 3.5817785263061523,
+ "step": 2552
+ },
+ {
+ "epoch": 35.460732984293195,
+ "grad_norm": 0.05528900399804115,
+ "learning_rate": 0.00012799671545824212,
+ "loss": 3.618821859359741,
+ "step": 2553
+ },
+ {
+ "epoch": 35.474694589877835,
+ "grad_norm": 0.0527317114174366,
+ "learning_rate": 0.0001277813945614768,
+ "loss": 3.6165270805358887,
+ "step": 2554
+ },
+ {
+ "epoch": 35.488656195462475,
+ "grad_norm": 0.05585351586341858,
+ "learning_rate": 0.00012756621483517682,
+ "loss": 3.6158246994018555,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50261780104712,
+ "grad_norm": 0.05375446751713753,
+ "learning_rate": 0.00012735117645270905,
+ "loss": 3.610105037689209,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51657940663176,
+ "grad_norm": 0.055324699729681015,
+ "learning_rate": 0.00012713627958732567,
+ "loss": 3.6106624603271484,
+ "step": 2557
+ },
+ {
+ "epoch": 35.5305410122164,
+ "grad_norm": 0.0544339083135128,
+ "learning_rate": 0.00012692152441216539,
+ "loss": 3.605295181274414,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544502617801044,
+ "grad_norm": 0.05522064492106438,
+ "learning_rate": 0.0001267069111002525,
+ "loss": 3.607207775115967,
+ "step": 2559
+ },
+ {
+ "epoch": 35.55846422338569,
+ "grad_norm": 0.055271610617637634,
+ "learning_rate": 0.00012649243982449718,
+ "loss": 3.610219955444336,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57242582897033,
+ "grad_norm": 0.05532364174723625,
+ "learning_rate": 0.00012627811075769502,
+ "loss": 3.6226587295532227,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58638743455497,
+ "grad_norm": 0.05356820300221443,
+ "learning_rate": 0.00012606392407252687,
+ "loss": 3.628182888031006,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60034904013962,
+ "grad_norm": 0.054853104054927826,
+ "learning_rate": 0.00012584987994155943,
+ "loss": 3.596072196960449,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61431064572426,
+ "grad_norm": 0.05462052673101425,
+ "learning_rate": 0.00012563597853724388,
+ "loss": 3.6201977729797363,
+ "step": 2564
+ },
+ {
+ "epoch": 35.6282722513089,
+ "grad_norm": 0.05532895773649216,
+ "learning_rate": 0.000125422220031917,
+ "loss": 3.642716407775879,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64223385689354,
+ "grad_norm": 0.055425263941287994,
+ "learning_rate": 0.00012520860459779972,
+ "loss": 3.5959243774414062,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65619546247819,
+ "grad_norm": 0.056545358151197433,
+ "learning_rate": 0.0001249951324069986,
+ "loss": 3.5796356201171875,
+ "step": 2567
+ },
+ {
+ "epoch": 35.67015706806283,
+ "grad_norm": 0.055357806384563446,
+ "learning_rate": 0.0001247818036315042,
+ "loss": 3.6106321811676025,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68411867364747,
+ "grad_norm": 0.05461106821894646,
+ "learning_rate": 0.00012456861844319155,
+ "loss": 3.591688632965088,
+ "step": 2569
+ },
+ {
+ "epoch": 35.69808027923211,
+ "grad_norm": 0.055133529007434845,
+ "learning_rate": 0.00012435557701382012,
+ "loss": 3.6243996620178223,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712041884816756,
+ "grad_norm": 0.055249396711587906,
+ "learning_rate": 0.0001241426795150336,
+ "loss": 3.608017683029175,
+ "step": 2571
+ },
+ {
+ "epoch": 35.726003490401396,
+ "grad_norm": 0.056398529559373856,
+ "learning_rate": 0.00012392992611835973,
+ "loss": 3.6282615661621094,
+ "step": 2572
+ },
+ {
+ "epoch": 35.73996509598604,
+ "grad_norm": 0.0552498884499073,
+ "learning_rate": 0.0001237173169952098,
+ "loss": 3.6407179832458496,
+ "step": 2573
+ },
+ {
+ "epoch": 35.753926701570684,
+ "grad_norm": 0.056712064892053604,
+ "learning_rate": 0.00012350485231687954,
+ "loss": 3.616793632507324,
+ "step": 2574
+ },
+ {
+ "epoch": 35.767888307155324,
+ "grad_norm": 0.055920615792274475,
+ "learning_rate": 0.0001232925322545476,
+ "loss": 3.6330060958862305,
+ "step": 2575
+ },
+ {
+ "epoch": 35.781849912739965,
+ "grad_norm": 0.05519915744662285,
+ "learning_rate": 0.0001230803569792765,
+ "loss": 3.61584210395813,
+ "step": 2576
+ },
+ {
+ "epoch": 35.795811518324605,
+ "grad_norm": 0.05472308024764061,
+ "learning_rate": 0.0001228683266620121,
+ "loss": 3.6116104125976562,
+ "step": 2577
+ },
+ {
+ "epoch": 35.80977312390925,
+ "grad_norm": 0.0556313656270504,
+ "learning_rate": 0.00012265644147358326,
+ "loss": 3.634107828140259,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82373472949389,
+ "grad_norm": 0.057370636612176895,
+ "learning_rate": 0.00012244470158470226,
+ "loss": 3.6127190589904785,
+ "step": 2579
+ },
+ {
+ "epoch": 35.83769633507853,
+ "grad_norm": 0.05609549954533577,
+ "learning_rate": 0.00012223310716596387,
+ "loss": 3.6323940753936768,
+ "step": 2580
+ },
+ {
+ "epoch": 35.85165794066317,
+ "grad_norm": 0.054692406207323074,
+ "learning_rate": 0.00012202165838784601,
+ "loss": 3.6339163780212402,
+ "step": 2581
+ },
+ {
+ "epoch": 35.86561954624782,
+ "grad_norm": 0.05703910440206528,
+ "learning_rate": 0.00012181035542070913,
+ "loss": 3.6473350524902344,
+ "step": 2582
+ },
+ {
+ "epoch": 35.87958115183246,
+ "grad_norm": 0.055098868906497955,
+ "learning_rate": 0.00012159919843479617,
+ "loss": 3.6110334396362305,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8935427574171,
+ "grad_norm": 0.0570925697684288,
+ "learning_rate": 0.00012138818760023257,
+ "loss": 3.630516767501831,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90750436300174,
+ "grad_norm": 0.05486847832798958,
+ "learning_rate": 0.00012117732308702592,
+ "loss": 3.622447967529297,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92146596858639,
+ "grad_norm": 0.057239681482315063,
+ "learning_rate": 0.00012096660506506605,
+ "loss": 3.6436092853546143,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93542757417103,
+ "grad_norm": 0.05756063386797905,
+ "learning_rate": 0.00012075603370412443,
+ "loss": 3.6313462257385254,
+ "step": 2587
+ },
+ {
+ "epoch": 35.94938917975567,
+ "grad_norm": 0.05791015550494194,
+ "learning_rate": 0.00012054560917385476,
+ "loss": 3.617950677871704,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96335078534032,
+ "grad_norm": 0.05538728088140488,
+ "learning_rate": 0.00012033533164379225,
+ "loss": 3.665337085723877,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97731239092496,
+ "grad_norm": 0.05843896046280861,
+ "learning_rate": 0.0001201252012833537,
+ "loss": 3.6643996238708496,
+ "step": 2590
+ },
+ {
+ "epoch": 35.9912739965096,
+ "grad_norm": 0.05609741806983948,
+ "learning_rate": 0.00011991521826183747,
+ "loss": 3.6212263107299805,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.04305510222911835,
+ "learning_rate": 0.00011970538274842282,
+ "loss": 2.270846128463745,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 0.6145148873329163,
+ "eval_runtime": 60.714,
+ "eval_samples_per_second": 40.221,
+ "eval_steps_per_second": 0.642,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01396160558464,
+ "grad_norm": 0.054062873125076294,
+ "learning_rate": 0.00011949569491217073,
+ "loss": 3.581265926361084,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02792321116929,
+ "grad_norm": 0.051643744111061096,
+ "learning_rate": 0.00011928615492202269,
+ "loss": 3.586836814880371,
+ "step": 2594
+ },
+ {
+ "epoch": 36.04188481675393,
+ "grad_norm": 0.054067034274339676,
+ "learning_rate": 0.0001190767629468014,
+ "loss": 3.6008591651916504,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05584642233857,
+ "grad_norm": 0.05220582336187363,
+ "learning_rate": 0.00011886751915521007,
+ "loss": 3.5975303649902344,
+ "step": 2596
+ },
+ {
+ "epoch": 36.06980802792321,
+ "grad_norm": 0.053417500108480453,
+ "learning_rate": 0.00011865842371583278,
+ "loss": 3.5652084350585938,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083769633507856,
+ "grad_norm": 0.05344022065401077,
+ "learning_rate": 0.00011844947679713396,
+ "loss": 3.561967372894287,
+ "step": 2598
+ },
+ {
+ "epoch": 36.0977312390925,
+ "grad_norm": 0.05316327139735222,
+ "learning_rate": 0.00011824067856745812,
+ "loss": 3.590935707092285,
+ "step": 2599
+ },
+ {
+ "epoch": 36.11169284467714,
+ "grad_norm": 0.054267000406980515,
+ "learning_rate": 0.00011803202919503064,
+ "loss": 3.5788426399230957,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12565445026178,
+ "grad_norm": 0.05376334860920906,
+ "learning_rate": 0.00011782352884795615,
+ "loss": 3.589162826538086,
+ "step": 2601
+ },
+ {
+ "epoch": 36.139616055846425,
+ "grad_norm": 0.05369187891483307,
+ "learning_rate": 0.00011761517769421983,
+ "loss": 3.558194160461426,
+ "step": 2602
+ },
+ {
+ "epoch": 36.153577661431065,
+ "grad_norm": 0.05407107621431351,
+ "learning_rate": 0.00011740697590168635,
+ "loss": 3.571460247039795,
+ "step": 2603
+ },
+ {
+ "epoch": 36.167539267015705,
+ "grad_norm": 0.05279739201068878,
+ "learning_rate": 0.00011719892363810018,
+ "loss": 3.588043689727783,
+ "step": 2604
+ },
+ {
+ "epoch": 36.181500872600346,
+ "grad_norm": 0.05608730763196945,
+ "learning_rate": 0.00011699102107108539,
+ "loss": 3.5704660415649414,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19546247818499,
+ "grad_norm": 0.05290020629763603,
+ "learning_rate": 0.00011678326836814507,
+ "loss": 3.5612759590148926,
+ "step": 2606
+ },
+ {
+ "epoch": 36.20942408376963,
+ "grad_norm": 0.05398614704608917,
+ "learning_rate": 0.00011657566569666198,
+ "loss": 3.579409122467041,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223385689354274,
+ "grad_norm": 0.051791902631521225,
+ "learning_rate": 0.00011636821322389777,
+ "loss": 3.5976991653442383,
+ "step": 2608
+ },
+ {
+ "epoch": 36.23734729493892,
+ "grad_norm": 0.05329076945781708,
+ "learning_rate": 0.00011616091111699333,
+ "loss": 3.6209723949432373,
+ "step": 2609
+ },
+ {
+ "epoch": 36.25130890052356,
+ "grad_norm": 0.052824556827545166,
+ "learning_rate": 0.00011595375954296786,
+ "loss": 3.5926363468170166,
+ "step": 2610
+ },
+ {
+ "epoch": 36.2652705061082,
+ "grad_norm": 0.05411917343735695,
+ "learning_rate": 0.00011574675866871997,
+ "loss": 3.607084274291992,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27923211169284,
+ "grad_norm": 0.0527656190097332,
+ "learning_rate": 0.00011553990866102651,
+ "loss": 3.579601526260376,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29319371727749,
+ "grad_norm": 0.05329589545726776,
+ "learning_rate": 0.00011533320968654265,
+ "loss": 3.5800671577453613,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30715532286213,
+ "grad_norm": 0.05175339803099632,
+ "learning_rate": 0.00011512666191180204,
+ "loss": 3.594729423522949,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32111692844677,
+ "grad_norm": 0.05224638432264328,
+ "learning_rate": 0.00011492026550321653,
+ "loss": 3.567016839981079,
+ "step": 2615
+ },
+ {
+ "epoch": 36.33507853403141,
+ "grad_norm": 0.053442589938640594,
+ "learning_rate": 0.00011471402062707607,
+ "loss": 3.5964465141296387,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34904013961606,
+ "grad_norm": 0.05282003805041313,
+ "learning_rate": 0.00011450792744954827,
+ "loss": 3.5932140350341797,
+ "step": 2617
+ },
+ {
+ "epoch": 36.3630017452007,
+ "grad_norm": 0.054600607603788376,
+ "learning_rate": 0.00011430198613667861,
+ "loss": 3.5916247367858887,
+ "step": 2618
+ },
+ {
+ "epoch": 36.37696335078534,
+ "grad_norm": 0.053483184427022934,
+ "learning_rate": 0.00011409619685439064,
+ "loss": 3.588195323944092,
+ "step": 2619
+ },
+ {
+ "epoch": 36.390924956369986,
+ "grad_norm": 0.052354127168655396,
+ "learning_rate": 0.00011389055976848477,
+ "loss": 3.6031956672668457,
+ "step": 2620
+ },
+ {
+ "epoch": 36.404886561954626,
+ "grad_norm": 0.05411195009946823,
+ "learning_rate": 0.00011368507504463914,
+ "loss": 3.61525559425354,
+ "step": 2621
+ },
+ {
+ "epoch": 36.41884816753927,
+ "grad_norm": 0.055229391902685165,
+ "learning_rate": 0.00011347974284840914,
+ "loss": 3.58837890625,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43280977312391,
+ "grad_norm": 0.05329221487045288,
+ "learning_rate": 0.00011327456334522725,
+ "loss": 3.565725803375244,
+ "step": 2623
+ },
+ {
+ "epoch": 36.446771378708554,
+ "grad_norm": 0.05464653670787811,
+ "learning_rate": 0.00011306953670040275,
+ "loss": 3.6044445037841797,
+ "step": 2624
+ },
+ {
+ "epoch": 36.460732984293195,
+ "grad_norm": 0.05323687940835953,
+ "learning_rate": 0.00011286466307912184,
+ "loss": 3.5857181549072266,
+ "step": 2625
+ },
+ {
+ "epoch": 36.474694589877835,
+ "grad_norm": 0.054866403341293335,
+ "learning_rate": 0.00011265994264644783,
+ "loss": 3.592473030090332,
+ "step": 2626
+ },
+ {
+ "epoch": 36.488656195462475,
+ "grad_norm": 0.05508954077959061,
+ "learning_rate": 0.00011245537556732001,
+ "loss": 3.5929830074310303,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50261780104712,
+ "grad_norm": 0.057712286710739136,
+ "learning_rate": 0.00011225096200655442,
+ "loss": 3.613492488861084,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51657940663176,
+ "grad_norm": 0.05367693305015564,
+ "learning_rate": 0.00011204670212884342,
+ "loss": 3.5849246978759766,
+ "step": 2629
+ },
+ {
+ "epoch": 36.5305410122164,
+ "grad_norm": 0.055529702454805374,
+ "learning_rate": 0.00011184259609875557,
+ "loss": 3.60062313079834,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544502617801044,
+ "grad_norm": 0.05650486797094345,
+ "learning_rate": 0.00011163864408073522,
+ "loss": 3.570746898651123,
+ "step": 2631
+ },
+ {
+ "epoch": 36.55846422338569,
+ "grad_norm": 0.055417656898498535,
+ "learning_rate": 0.00011143484623910293,
+ "loss": 3.5839993953704834,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57242582897033,
+ "grad_norm": 0.055162668228149414,
+ "learning_rate": 0.00011123120273805496,
+ "loss": 3.5858545303344727,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58638743455497,
+ "grad_norm": 0.054991088807582855,
+ "learning_rate": 0.0001110277137416632,
+ "loss": 3.5967633724212646,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60034904013962,
+ "grad_norm": 0.055452100932598114,
+ "learning_rate": 0.00011082437941387512,
+ "loss": 3.6110477447509766,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61431064572426,
+ "grad_norm": 0.05554793030023575,
+ "learning_rate": 0.00011062119991851322,
+ "loss": 3.5871753692626953,
+ "step": 2636
+ },
+ {
+ "epoch": 36.6282722513089,
+ "grad_norm": 0.054689038544893265,
+ "learning_rate": 0.000110418175419276,
+ "loss": 3.590219020843506,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64223385689354,
+ "grad_norm": 0.05510031804442406,
+ "learning_rate": 0.00011021530607973626,
+ "loss": 3.5875062942504883,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65619546247819,
+ "grad_norm": 0.05421333760023117,
+ "learning_rate": 0.00011001259206334235,
+ "loss": 3.602912187576294,
+ "step": 2639
+ },
+ {
+ "epoch": 36.67015706806283,
+ "grad_norm": 0.05532202124595642,
+ "learning_rate": 0.00010981003353341721,
+ "loss": 3.5950570106506348,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68411867364747,
+ "grad_norm": 0.05355961248278618,
+ "learning_rate": 0.00010960763065315864,
+ "loss": 3.5744097232818604,
+ "step": 2641
+ },
+ {
+ "epoch": 36.69808027923211,
+ "grad_norm": 0.05564635619521141,
+ "learning_rate": 0.00010940538358563906,
+ "loss": 3.5986971855163574,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712041884816756,
+ "grad_norm": 0.054115377366542816,
+ "learning_rate": 0.00010920329249380509,
+ "loss": 3.596283197402954,
+ "step": 2643
+ },
+ {
+ "epoch": 36.726003490401396,
+ "grad_norm": 0.05481656640768051,
+ "learning_rate": 0.00010900135754047799,
+ "loss": 3.5836915969848633,
+ "step": 2644
+ },
+ {
+ "epoch": 36.73996509598604,
+ "grad_norm": 0.05519521236419678,
+ "learning_rate": 0.00010879957888835307,
+ "loss": 3.6295974254608154,
+ "step": 2645
+ },
+ {
+ "epoch": 36.753926701570684,
+ "grad_norm": 0.05460044741630554,
+ "learning_rate": 0.00010859795669999981,
+ "loss": 3.5932111740112305,
+ "step": 2646
+ },
+ {
+ "epoch": 36.767888307155324,
+ "grad_norm": 0.054085709154605865,
+ "learning_rate": 0.00010839649113786137,
+ "loss": 3.584069013595581,
+ "step": 2647
+ },
+ {
+ "epoch": 36.781849912739965,
+ "grad_norm": 0.05407894030213356,
+ "learning_rate": 0.00010819518236425514,
+ "loss": 3.5967845916748047,
+ "step": 2648
+ },
+ {
+ "epoch": 36.795811518324605,
+ "grad_norm": 0.05513349175453186,
+ "learning_rate": 0.00010799403054137197,
+ "loss": 3.598283290863037,
+ "step": 2649
+ },
+ {
+ "epoch": 36.80977312390925,
+ "grad_norm": 0.0558769516646862,
+ "learning_rate": 0.00010779303583127609,
+ "loss": 3.60675048828125,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82373472949389,
+ "grad_norm": 0.05440812557935715,
+ "learning_rate": 0.00010759219839590535,
+ "loss": 3.5989885330200195,
+ "step": 2651
+ },
+ {
+ "epoch": 36.83769633507853,
+ "grad_norm": 0.05533481389284134,
+ "learning_rate": 0.00010739151839707089,
+ "loss": 3.6296310424804688,
+ "step": 2652
+ },
+ {
+ "epoch": 36.85165794066317,
+ "grad_norm": 0.055396221578121185,
+ "learning_rate": 0.00010719099599645706,
+ "loss": 3.6243319511413574,
+ "step": 2653
+ },
+ {
+ "epoch": 36.86561954624782,
+ "grad_norm": 0.05608600378036499,
+ "learning_rate": 0.0001069906313556208,
+ "loss": 3.6124401092529297,
+ "step": 2654
+ },
+ {
+ "epoch": 36.87958115183246,
+ "grad_norm": 0.05493833124637604,
+ "learning_rate": 0.00010679042463599263,
+ "loss": 3.6008474826812744,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8935427574171,
+ "grad_norm": 0.055183470249176025,
+ "learning_rate": 0.00010659037599887545,
+ "loss": 3.5892064571380615,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90750436300174,
+ "grad_norm": 0.05589289590716362,
+ "learning_rate": 0.0001063904856054446,
+ "loss": 3.594189405441284,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92146596858639,
+ "grad_norm": 0.05483611673116684,
+ "learning_rate": 0.00010619075361674836,
+ "loss": 3.586012601852417,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93542757417103,
+ "grad_norm": 0.055471133440732956,
+ "learning_rate": 0.0001059911801937071,
+ "loss": 3.602313995361328,
+ "step": 2659
+ },
+ {
+ "epoch": 36.94938917975567,
+ "grad_norm": 0.0559491291642189,
+ "learning_rate": 0.00010579176549711365,
+ "loss": 3.6010072231292725,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96335078534032,
+ "grad_norm": 0.055661238729953766,
+ "learning_rate": 0.00010559250968763265,
+ "loss": 3.6093027591705322,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97731239092496,
+ "grad_norm": 0.05550997704267502,
+ "learning_rate": 0.00010539341292580086,
+ "loss": 3.614647150039673,
+ "step": 2662
+ },
+ {
+ "epoch": 36.9912739965096,
+ "grad_norm": 0.056113287806510925,
+ "learning_rate": 0.00010519447537202729,
+ "loss": 3.5756256580352783,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.04090592637658119,
+ "learning_rate": 0.00010499569718659201,
+ "loss": 2.237126588821411,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 0.6159548759460449,
+ "eval_runtime": 60.4892,
+ "eval_samples_per_second": 40.371,
+ "eval_steps_per_second": 0.645,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01396160558464,
+ "grad_norm": 0.05119108036160469,
+ "learning_rate": 0.00010479707852964713,
+ "loss": 3.5844168663024902,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02792321116929,
+ "grad_norm": 0.050482187420129776,
+ "learning_rate": 0.00010459861956121611,
+ "loss": 3.5596923828125,
+ "step": 2666
+ },
+ {
+ "epoch": 37.04188481675393,
+ "grad_norm": 0.05054338648915291,
+ "learning_rate": 0.00010440032044119383,
+ "loss": 3.533599376678467,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05584642233857,
+ "grad_norm": 0.050228655338287354,
+ "learning_rate": 0.00010420218132934614,
+ "loss": 3.557422637939453,
+ "step": 2668
+ },
+ {
+ "epoch": 37.06980802792321,
+ "grad_norm": 0.04899684339761734,
+ "learning_rate": 0.00010400420238531023,
+ "loss": 3.5452675819396973,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083769633507856,
+ "grad_norm": 0.049607474356889725,
+ "learning_rate": 0.00010380638376859416,
+ "loss": 3.571603298187256,
+ "step": 2670
+ },
+ {
+ "epoch": 37.0977312390925,
+ "grad_norm": 0.04938557744026184,
+ "learning_rate": 0.00010360872563857682,
+ "loss": 3.5608606338500977,
+ "step": 2671
+ },
+ {
+ "epoch": 37.11169284467714,
+ "grad_norm": 0.05051542446017265,
+ "learning_rate": 0.0001034112281545079,
+ "loss": 3.5389974117279053,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12565445026178,
+ "grad_norm": 0.04959443211555481,
+ "learning_rate": 0.00010321389147550725,
+ "loss": 3.5671193599700928,
+ "step": 2673
+ },
+ {
+ "epoch": 37.139616055846425,
+ "grad_norm": 0.0502617284655571,
+ "learning_rate": 0.00010301671576056588,
+ "loss": 3.5699734687805176,
+ "step": 2674
+ },
+ {
+ "epoch": 37.153577661431065,
+ "grad_norm": 0.051576778292655945,
+ "learning_rate": 0.00010281970116854436,
+ "loss": 3.562732219696045,
+ "step": 2675
+ },
+ {
+ "epoch": 37.167539267015705,
+ "grad_norm": 0.05000219866633415,
+ "learning_rate": 0.00010262284785817392,
+ "loss": 3.545559883117676,
+ "step": 2676
+ },
+ {
+ "epoch": 37.181500872600346,
+ "grad_norm": 0.049805279821157455,
+ "learning_rate": 0.00010242615598805574,
+ "loss": 3.5819478034973145,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19546247818499,
+ "grad_norm": 0.05044523999094963,
+ "learning_rate": 0.00010222962571666088,
+ "loss": 3.57983136177063,
+ "step": 2678
+ },
+ {
+ "epoch": 37.20942408376963,
+ "grad_norm": 0.05141332373023033,
+ "learning_rate": 0.00010203325720233032,
+ "loss": 3.5591063499450684,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223385689354274,
+ "grad_norm": 0.05136275291442871,
+ "learning_rate": 0.00010183705060327433,
+ "loss": 3.5776171684265137,
+ "step": 2680
+ },
+ {
+ "epoch": 37.23734729493892,
+ "grad_norm": 0.05119960755109787,
+ "learning_rate": 0.00010164100607757346,
+ "loss": 3.5721652507781982,
+ "step": 2681
+ },
+ {
+ "epoch": 37.25130890052356,
+ "grad_norm": 0.05113738030195236,
+ "learning_rate": 0.00010144512378317687,
+ "loss": 3.5741496086120605,
+ "step": 2682
+ },
+ {
+ "epoch": 37.2652705061082,
+ "grad_norm": 0.05131804198026657,
+ "learning_rate": 0.00010124940387790354,
+ "loss": 3.577519416809082,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27923211169284,
+ "grad_norm": 0.05085485801100731,
+ "learning_rate": 0.00010105384651944148,
+ "loss": 3.5454201698303223,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29319371727749,
+ "grad_norm": 0.05160171166062355,
+ "learning_rate": 0.00010085845186534769,
+ "loss": 3.574143886566162,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30715532286213,
+ "grad_norm": 0.05132637917995453,
+ "learning_rate": 0.00010066322007304819,
+ "loss": 3.5619382858276367,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32111692844677,
+ "grad_norm": 0.05075863003730774,
+ "learning_rate": 0.00010046815129983757,
+ "loss": 3.5536012649536133,
+ "step": 2687
+ },
+ {
+ "epoch": 37.33507853403141,
+ "grad_norm": 0.052456241101026535,
+ "learning_rate": 0.00010027324570287925,
+ "loss": 3.586211681365967,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34904013961606,
+ "grad_norm": 0.051105350255966187,
+ "learning_rate": 0.00010007850343920519,
+ "loss": 3.5617852210998535,
+ "step": 2689
+ },
+ {
+ "epoch": 37.3630017452007,
+ "grad_norm": 0.052098311483860016,
+ "learning_rate": 9.988392466571572e-05,
+ "loss": 3.5693583488464355,
+ "step": 2690
+ },
+ {
+ "epoch": 37.37696335078534,
+ "grad_norm": 0.05069877579808235,
+ "learning_rate": 9.968950953917945e-05,
+ "loss": 3.547999620437622,
+ "step": 2691
+ },
+ {
+ "epoch": 37.390924956369986,
+ "grad_norm": 0.05264188349246979,
+ "learning_rate": 9.949525821623309e-05,
+ "loss": 3.554316520690918,
+ "step": 2692
+ },
+ {
+ "epoch": 37.404886561954626,
+ "grad_norm": 0.05224690958857536,
+ "learning_rate": 9.930117085338156e-05,
+ "loss": 3.531437397003174,
+ "step": 2693
+ },
+ {
+ "epoch": 37.41884816753927,
+ "grad_norm": 0.05134997144341469,
+ "learning_rate": 9.910724760699731e-05,
+ "loss": 3.5924055576324463,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43280977312391,
+ "grad_norm": 0.05177623778581619,
+ "learning_rate": 9.891348863332093e-05,
+ "loss": 3.580904006958008,
+ "step": 2695
+ },
+ {
+ "epoch": 37.446771378708554,
+ "grad_norm": 0.052150946110486984,
+ "learning_rate": 9.871989408846051e-05,
+ "loss": 3.568195343017578,
+ "step": 2696
+ },
+ {
+ "epoch": 37.460732984293195,
+ "grad_norm": 0.05220150575041771,
+ "learning_rate": 9.852646412839178e-05,
+ "loss": 3.5756545066833496,
+ "step": 2697
+ },
+ {
+ "epoch": 37.474694589877835,
+ "grad_norm": 0.053251419216394424,
+ "learning_rate": 9.833319890895756e-05,
+ "loss": 3.5686798095703125,
+ "step": 2698
+ },
+ {
+ "epoch": 37.488656195462475,
+ "grad_norm": 0.05219034478068352,
+ "learning_rate": 9.814009858586817e-05,
+ "loss": 3.5607855319976807,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50261780104712,
+ "grad_norm": 0.05180756375193596,
+ "learning_rate": 9.794716331470131e-05,
+ "loss": 3.567056179046631,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51657940663176,
+ "grad_norm": 0.0524761863052845,
+ "learning_rate": 9.77543932509012e-05,
+ "loss": 3.5638341903686523,
+ "step": 2701
+ },
+ {
+ "epoch": 37.5305410122164,
+ "grad_norm": 0.05237184837460518,
+ "learning_rate": 9.756178854977925e-05,
+ "loss": 3.575751781463623,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544502617801044,
+ "grad_norm": 0.05274934321641922,
+ "learning_rate": 9.736934936651362e-05,
+ "loss": 3.5704338550567627,
+ "step": 2703
+ },
+ {
+ "epoch": 37.55846422338569,
+ "grad_norm": 0.053565170615911484,
+ "learning_rate": 9.717707585614916e-05,
+ "loss": 3.566765785217285,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57242582897033,
+ "grad_norm": 0.05329638719558716,
+ "learning_rate": 9.698496817359698e-05,
+ "loss": 3.588357448577881,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58638743455497,
+ "grad_norm": 0.05382629856467247,
+ "learning_rate": 9.679302647363476e-05,
+ "loss": 3.578054428100586,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60034904013962,
+ "grad_norm": 0.05376645550131798,
+ "learning_rate": 9.660125091090675e-05,
+ "loss": 3.5657711029052734,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61431064572426,
+ "grad_norm": 0.05371774733066559,
+ "learning_rate": 9.64096416399228e-05,
+ "loss": 3.591853141784668,
+ "step": 2708
+ },
+ {
+ "epoch": 37.6282722513089,
+ "grad_norm": 0.055276691913604736,
+ "learning_rate": 9.621819881505918e-05,
+ "loss": 3.5756590366363525,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64223385689354,
+ "grad_norm": 0.05425848439335823,
+ "learning_rate": 9.602692259055767e-05,
+ "loss": 3.5689821243286133,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65619546247819,
+ "grad_norm": 0.0549544058740139,
+ "learning_rate": 9.583581312052646e-05,
+ "loss": 3.5869078636169434,
+ "step": 2711
+ },
+ {
+ "epoch": 37.67015706806283,
+ "grad_norm": 0.05412036553025246,
+ "learning_rate": 9.564487055893867e-05,
+ "loss": 3.57331919670105,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68411867364747,
+ "grad_norm": 0.05490295961499214,
+ "learning_rate": 9.545409505963338e-05,
+ "loss": 3.581331729888916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.69808027923211,
+ "grad_norm": 0.05448489263653755,
+ "learning_rate": 9.526348677631499e-05,
+ "loss": 3.605595588684082,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712041884816756,
+ "grad_norm": 0.05397019162774086,
+ "learning_rate": 9.507304586255311e-05,
+ "loss": 3.5643043518066406,
+ "step": 2715
+ },
+ {
+ "epoch": 37.726003490401396,
+ "grad_norm": 0.05399048700928688,
+ "learning_rate": 9.488277247178267e-05,
+ "loss": 3.5841126441955566,
+ "step": 2716
+ },
+ {
+ "epoch": 37.73996509598604,
+ "grad_norm": 0.055010151118040085,
+ "learning_rate": 9.469266675730319e-05,
+ "loss": 3.5840625762939453,
+ "step": 2717
+ },
+ {
+ "epoch": 37.753926701570684,
+ "grad_norm": 0.05478813126683235,
+ "learning_rate": 9.450272887227983e-05,
+ "loss": 3.56626033782959,
+ "step": 2718
+ },
+ {
+ "epoch": 37.767888307155324,
+ "grad_norm": 0.05396709218621254,
+ "learning_rate": 9.431295896974182e-05,
+ "loss": 3.5726799964904785,
+ "step": 2719
+ },
+ {
+ "epoch": 37.781849912739965,
+ "grad_norm": 0.05348159372806549,
+ "learning_rate": 9.412335720258341e-05,
+ "loss": 3.572139263153076,
+ "step": 2720
+ },
+ {
+ "epoch": 37.795811518324605,
+ "grad_norm": 0.055067192763090134,
+ "learning_rate": 9.393392372356335e-05,
+ "loss": 3.5794320106506348,
+ "step": 2721
+ },
+ {
+ "epoch": 37.80977312390925,
+ "grad_norm": 0.054990511387586594,
+ "learning_rate": 9.374465868530477e-05,
+ "loss": 3.573702335357666,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82373472949389,
+ "grad_norm": 0.05417406186461449,
+ "learning_rate": 9.355556224029515e-05,
+ "loss": 3.574542999267578,
+ "step": 2723
+ },
+ {
+ "epoch": 37.83769633507853,
+ "grad_norm": 0.05412792041897774,
+ "learning_rate": 9.336663454088593e-05,
+ "loss": 3.5837464332580566,
+ "step": 2724
+ },
+ {
+ "epoch": 37.85165794066317,
+ "grad_norm": 0.055187419056892395,
+ "learning_rate": 9.317787573929282e-05,
+ "loss": 3.5841941833496094,
+ "step": 2725
+ },
+ {
+ "epoch": 37.86561954624782,
+ "grad_norm": 0.05523138865828514,
+ "learning_rate": 9.298928598759541e-05,
+ "loss": 3.582552909851074,
+ "step": 2726
+ },
+ {
+ "epoch": 37.87958115183246,
+ "grad_norm": 0.05365348979830742,
+ "learning_rate": 9.280086543773698e-05,
+ "loss": 3.5693717002868652,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8935427574171,
+ "grad_norm": 0.0547465942800045,
+ "learning_rate": 9.26126142415247e-05,
+ "loss": 3.551072597503662,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90750436300174,
+ "grad_norm": 0.05491182208061218,
+ "learning_rate": 9.2424532550629e-05,
+ "loss": 3.573578357696533,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92146596858639,
+ "grad_norm": 0.055363260209560394,
+ "learning_rate": 9.223662051658408e-05,
+ "loss": 3.5630593299865723,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93542757417103,
+ "grad_norm": 0.05482657626271248,
+ "learning_rate": 9.204887829078709e-05,
+ "loss": 3.5579466819763184,
+ "step": 2731
+ },
+ {
+ "epoch": 37.94938917975567,
+ "grad_norm": 0.05555591359734535,
+ "learning_rate": 9.186130602449861e-05,
+ "loss": 3.572491407394409,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96335078534032,
+ "grad_norm": 0.0549536868929863,
+ "learning_rate": 9.167390386884224e-05,
+ "loss": 3.576451301574707,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97731239092496,
+ "grad_norm": 0.0547570176422596,
+ "learning_rate": 9.148667197480455e-05,
+ "loss": 3.5588786602020264,
+ "step": 2734
+ },
+ {
+ "epoch": 37.9912739965096,
+ "grad_norm": 0.055114034563302994,
+ "learning_rate": 9.129961049323494e-05,
+ "loss": 3.578408718109131,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.04000422731041908,
+ "learning_rate": 9.111271957484519e-05,
+ "loss": 2.212750196456909,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 0.6171724796295166,
+ "eval_runtime": 60.0184,
+ "eval_samples_per_second": 40.687,
+ "eval_steps_per_second": 0.65,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01396160558464,
+ "grad_norm": 0.05178803205490112,
+ "learning_rate": 9.092599937021031e-05,
+ "loss": 3.5207724571228027,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02792321116929,
+ "grad_norm": 0.048786960542201996,
+ "learning_rate": 9.073945002976715e-05,
+ "loss": 3.524625778198242,
+ "step": 2738
+ },
+ {
+ "epoch": 38.04188481675393,
+ "grad_norm": 0.05165952071547508,
+ "learning_rate": 9.055307170381522e-05,
+ "loss": 3.5370655059814453,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05584642233857,
+ "grad_norm": 0.05022848770022392,
+ "learning_rate": 9.036686454251613e-05,
+ "loss": 3.52911376953125,
+ "step": 2740
+ },
+ {
+ "epoch": 38.06980802792321,
+ "grad_norm": 0.04977148398756981,
+ "learning_rate": 9.018082869589369e-05,
+ "loss": 3.548173666000366,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083769633507856,
+ "grad_norm": 0.04955532029271126,
+ "learning_rate": 8.999496431383368e-05,
+ "loss": 3.5462260246276855,
+ "step": 2742
+ },
+ {
+ "epoch": 38.0977312390925,
+ "grad_norm": 0.051178816705942154,
+ "learning_rate": 8.980927154608341e-05,
+ "loss": 3.539468288421631,
+ "step": 2743
+ },
+ {
+ "epoch": 38.11169284467714,
+ "grad_norm": 0.05035502463579178,
+ "learning_rate": 8.962375054225255e-05,
+ "loss": 3.5545785427093506,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12565445026178,
+ "grad_norm": 0.04895094409584999,
+ "learning_rate": 8.943840145181177e-05,
+ "loss": 3.535142421722412,
+ "step": 2745
+ },
+ {
+ "epoch": 38.139616055846425,
+ "grad_norm": 0.050836317241191864,
+ "learning_rate": 8.925322442409361e-05,
+ "loss": 3.5562942028045654,
+ "step": 2746
+ },
+ {
+ "epoch": 38.153577661431065,
+ "grad_norm": 0.04898427426815033,
+ "learning_rate": 8.906821960829163e-05,
+ "loss": 3.533984661102295,
+ "step": 2747
+ },
+ {
+ "epoch": 38.167539267015705,
+ "grad_norm": 0.05146025866270065,
+ "learning_rate": 8.888338715346113e-05,
+ "loss": 3.5528769493103027,
+ "step": 2748
+ },
+ {
+ "epoch": 38.181500872600346,
+ "grad_norm": 0.049640901386737823,
+ "learning_rate": 8.869872720851831e-05,
+ "loss": 3.5450353622436523,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19546247818499,
+ "grad_norm": 0.04954954609274864,
+ "learning_rate": 8.851423992224012e-05,
+ "loss": 3.566117763519287,
+ "step": 2750
+ },
+ {
+ "epoch": 38.20942408376963,
+ "grad_norm": 0.04950375854969025,
+ "learning_rate": 8.832992544326473e-05,
+ "loss": 3.549238443374634,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223385689354274,
+ "grad_norm": 0.05026113614439964,
+ "learning_rate": 8.814578392009104e-05,
+ "loss": 3.5252106189727783,
+ "step": 2752
+ },
+ {
+ "epoch": 38.23734729493892,
+ "grad_norm": 0.050842445343732834,
+ "learning_rate": 8.796181550107857e-05,
+ "loss": 3.5351614952087402,
+ "step": 2753
+ },
+ {
+ "epoch": 38.25130890052356,
+ "grad_norm": 0.050328366458415985,
+ "learning_rate": 8.777802033444712e-05,
+ "loss": 3.5556483268737793,
+ "step": 2754
+ },
+ {
+ "epoch": 38.2652705061082,
+ "grad_norm": 0.05114470049738884,
+ "learning_rate": 8.75943985682774e-05,
+ "loss": 3.5427448749542236,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27923211169284,
+ "grad_norm": 0.05121666193008423,
+ "learning_rate": 8.741095035051017e-05,
+ "loss": 3.5361504554748535,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29319371727749,
+ "grad_norm": 0.0501880943775177,
+ "learning_rate": 8.722767582894613e-05,
+ "loss": 3.5314698219299316,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30715532286213,
+ "grad_norm": 0.05276653543114662,
+ "learning_rate": 8.704457515124636e-05,
+ "loss": 3.5526633262634277,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32111692844677,
+ "grad_norm": 0.05037686228752136,
+ "learning_rate": 8.686164846493176e-05,
+ "loss": 3.537220001220703,
+ "step": 2759
+ },
+ {
+ "epoch": 38.33507853403141,
+ "grad_norm": 0.051579032093286514,
+ "learning_rate": 8.667889591738317e-05,
+ "loss": 3.5302658081054688,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34904013961606,
+ "grad_norm": 0.05040337145328522,
+ "learning_rate": 8.649631765584083e-05,
+ "loss": 3.528028964996338,
+ "step": 2761
+ },
+ {
+ "epoch": 38.3630017452007,
+ "grad_norm": 0.05162221938371658,
+ "learning_rate": 8.631391382740482e-05,
+ "loss": 3.5771780014038086,
+ "step": 2762
+ },
+ {
+ "epoch": 38.37696335078534,
+ "grad_norm": 0.05134383216500282,
+ "learning_rate": 8.613168457903459e-05,
+ "loss": 3.532721996307373,
+ "step": 2763
+ },
+ {
+ "epoch": 38.390924956369986,
+ "grad_norm": 0.05137534812092781,
+ "learning_rate": 8.594963005754901e-05,
+ "loss": 3.550847291946411,
+ "step": 2764
+ },
+ {
+ "epoch": 38.404886561954626,
+ "grad_norm": 0.05198913812637329,
+ "learning_rate": 8.57677504096261e-05,
+ "loss": 3.556427478790283,
+ "step": 2765
+ },
+ {
+ "epoch": 38.41884816753927,
+ "grad_norm": 0.05211595818400383,
+ "learning_rate": 8.558604578180301e-05,
+ "loss": 3.579143524169922,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43280977312391,
+ "grad_norm": 0.05178852006793022,
+ "learning_rate": 8.540451632047593e-05,
+ "loss": 3.548065423965454,
+ "step": 2767
+ },
+ {
+ "epoch": 38.446771378708554,
+ "grad_norm": 0.05172461271286011,
+ "learning_rate": 8.522316217189972e-05,
+ "loss": 3.523592948913574,
+ "step": 2768
+ },
+ {
+ "epoch": 38.460732984293195,
+ "grad_norm": 0.05164046585559845,
+ "learning_rate": 8.504198348218821e-05,
+ "loss": 3.5595390796661377,
+ "step": 2769
+ },
+ {
+ "epoch": 38.474694589877835,
+ "grad_norm": 0.05212831869721413,
+ "learning_rate": 8.486098039731384e-05,
+ "loss": 3.517983913421631,
+ "step": 2770
+ },
+ {
+ "epoch": 38.488656195462475,
+ "grad_norm": 0.05294293537735939,
+ "learning_rate": 8.46801530631075e-05,
+ "loss": 3.5454447269439697,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50261780104712,
+ "grad_norm": 0.05115564912557602,
+ "learning_rate": 8.449950162525859e-05,
+ "loss": 3.562196731567383,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51657940663176,
+ "grad_norm": 0.05171671882271767,
+ "learning_rate": 8.431902622931443e-05,
+ "loss": 3.526003360748291,
+ "step": 2773
+ },
+ {
+ "epoch": 38.5305410122164,
+ "grad_norm": 0.052197013050317764,
+ "learning_rate": 8.413872702068119e-05,
+ "loss": 3.5459299087524414,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544502617801044,
+ "grad_norm": 0.05189663916826248,
+ "learning_rate": 8.395860414462238e-05,
+ "loss": 3.572622299194336,
+ "step": 2775
+ },
+ {
+ "epoch": 38.55846422338569,
+ "grad_norm": 0.05171465501189232,
+ "learning_rate": 8.377865774625985e-05,
+ "loss": 3.535243034362793,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57242582897033,
+ "grad_norm": 0.052626628428697586,
+ "learning_rate": 8.359888797057316e-05,
+ "loss": 3.555469036102295,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58638743455497,
+ "grad_norm": 0.05242423340678215,
+ "learning_rate": 8.341929496239958e-05,
+ "loss": 3.5629148483276367,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60034904013962,
+ "grad_norm": 0.051777057349681854,
+ "learning_rate": 8.323987886643402e-05,
+ "loss": 3.5607125759124756,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61431064572426,
+ "grad_norm": 0.05259648710489273,
+ "learning_rate": 8.306063982722855e-05,
+ "loss": 3.5630576610565186,
+ "step": 2780
+ },
+ {
+ "epoch": 38.6282722513089,
+ "grad_norm": 0.05220061168074608,
+ "learning_rate": 8.288157798919318e-05,
+ "loss": 3.545198917388916,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64223385689354,
+ "grad_norm": 0.05149109661579132,
+ "learning_rate": 8.27026934965945e-05,
+ "loss": 3.5277156829833984,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65619546247819,
+ "grad_norm": 0.05252330005168915,
+ "learning_rate": 8.25239864935566e-05,
+ "loss": 3.541299343109131,
+ "step": 2783
+ },
+ {
+ "epoch": 38.67015706806283,
+ "grad_norm": 0.05302104353904724,
+ "learning_rate": 8.23454571240605e-05,
+ "loss": 3.5433006286621094,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68411867364747,
+ "grad_norm": 0.053501565009355545,
+ "learning_rate": 8.216710553194416e-05,
+ "loss": 3.5382585525512695,
+ "step": 2785
+ },
+ {
+ "epoch": 38.69808027923211,
+ "grad_norm": 0.053820304572582245,
+ "learning_rate": 8.198893186090222e-05,
+ "loss": 3.5610904693603516,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712041884816756,
+ "grad_norm": 0.05330271273851395,
+ "learning_rate": 8.181093625448585e-05,
+ "loss": 3.5504231452941895,
+ "step": 2787
+ },
+ {
+ "epoch": 38.726003490401396,
+ "grad_norm": 0.05421965569257736,
+ "learning_rate": 8.163311885610299e-05,
+ "loss": 3.550283432006836,
+ "step": 2788
+ },
+ {
+ "epoch": 38.73996509598604,
+ "grad_norm": 0.053781405091285706,
+ "learning_rate": 8.145547980901791e-05,
+ "loss": 3.5624003410339355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.753926701570684,
+ "grad_norm": 0.05342188850045204,
+ "learning_rate": 8.127801925635126e-05,
+ "loss": 3.524596691131592,
+ "step": 2790
+ },
+ {
+ "epoch": 38.767888307155324,
+ "grad_norm": 0.05213013291358948,
+ "learning_rate": 8.110073734107954e-05,
+ "loss": 3.5316734313964844,
+ "step": 2791
+ },
+ {
+ "epoch": 38.781849912739965,
+ "grad_norm": 0.052924077957868576,
+ "learning_rate": 8.092363420603584e-05,
+ "loss": 3.5611703395843506,
+ "step": 2792
+ },
+ {
+ "epoch": 38.795811518324605,
+ "grad_norm": 0.05358070507645607,
+ "learning_rate": 8.074670999390896e-05,
+ "loss": 3.545983076095581,
+ "step": 2793
+ },
+ {
+ "epoch": 38.80977312390925,
+ "grad_norm": 0.053544074296951294,
+ "learning_rate": 8.056996484724342e-05,
+ "loss": 3.5701904296875,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82373472949389,
+ "grad_norm": 0.054532721638679504,
+ "learning_rate": 8.039339890843958e-05,
+ "loss": 3.543095111846924,
+ "step": 2795
+ },
+ {
+ "epoch": 38.83769633507853,
+ "grad_norm": 0.0532798208296299,
+ "learning_rate": 8.02170123197535e-05,
+ "loss": 3.5167088508605957,
+ "step": 2796
+ },
+ {
+ "epoch": 38.85165794066317,
+ "grad_norm": 0.05314210429787636,
+ "learning_rate": 8.004080522329674e-05,
+ "loss": 3.5611109733581543,
+ "step": 2797
+ },
+ {
+ "epoch": 38.86561954624782,
+ "grad_norm": 0.053358301520347595,
+ "learning_rate": 7.986477776103601e-05,
+ "loss": 3.550701141357422,
+ "step": 2798
+ },
+ {
+ "epoch": 38.87958115183246,
+ "grad_norm": 0.0538727343082428,
+ "learning_rate": 7.968893007479343e-05,
+ "loss": 3.562330722808838,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8935427574171,
+ "grad_norm": 0.05346875637769699,
+ "learning_rate": 7.951326230624658e-05,
+ "loss": 3.6011950969696045,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90750436300174,
+ "grad_norm": 0.05399498715996742,
+ "learning_rate": 7.933777459692756e-05,
+ "loss": 3.552870750427246,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92146596858639,
+ "grad_norm": 0.053817737847566605,
+ "learning_rate": 7.916246708822373e-05,
+ "loss": 3.5696794986724854,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93542757417103,
+ "grad_norm": 0.05272316932678223,
+ "learning_rate": 7.898733992137715e-05,
+ "loss": 3.55513334274292,
+ "step": 2803
+ },
+ {
+ "epoch": 38.94938917975567,
+ "grad_norm": 0.053616899996995926,
+ "learning_rate": 7.881239323748475e-05,
+ "loss": 3.5549991130828857,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96335078534032,
+ "grad_norm": 0.053872063755989075,
+ "learning_rate": 7.863762717749771e-05,
+ "loss": 3.537346839904785,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97731239092496,
+ "grad_norm": 0.054122194647789,
+ "learning_rate": 7.846304188222189e-05,
+ "loss": 3.552178144454956,
+ "step": 2806
+ },
+ {
+ "epoch": 38.9912739965096,
+ "grad_norm": 0.053457897156476974,
+ "learning_rate": 7.828863749231777e-05,
+ "loss": 3.574368715286255,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.04022035375237465,
+ "learning_rate": 7.811441414829958e-05,
+ "loss": 2.2162294387817383,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 0.6184564828872681,
+ "eval_runtime": 60.5596,
+ "eval_samples_per_second": 40.324,
+ "eval_steps_per_second": 0.644,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01396160558464,
+ "grad_norm": 0.04818584397435188,
+ "learning_rate": 7.794037199053598e-05,
+ "loss": 3.5161890983581543,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02792321116929,
+ "grad_norm": 0.04768325760960579,
+ "learning_rate": 7.776651115924959e-05,
+ "loss": 3.5117123126983643,
+ "step": 2810
+ },
+ {
+ "epoch": 39.04188481675393,
+ "grad_norm": 0.049668096005916595,
+ "learning_rate": 7.759283179451704e-05,
+ "loss": 3.5076825618743896,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05584642233857,
+ "grad_norm": 0.04928773269057274,
+ "learning_rate": 7.741933403626848e-05,
+ "loss": 3.522817611694336,
+ "step": 2812
+ },
+ {
+ "epoch": 39.06980802792321,
+ "grad_norm": 0.049023669213056564,
+ "learning_rate": 7.724601802428805e-05,
+ "loss": 3.4985058307647705,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083769633507856,
+ "grad_norm": 0.04910000413656235,
+ "learning_rate": 7.707288389821329e-05,
+ "loss": 3.5706303119659424,
+ "step": 2814
+ },
+ {
+ "epoch": 39.0977312390925,
+ "grad_norm": 0.04971414431929588,
+ "learning_rate": 7.689993179753519e-05,
+ "loss": 3.5187277793884277,
+ "step": 2815
+ },
+ {
+ "epoch": 39.11169284467714,
+ "grad_norm": 0.049427829682826996,
+ "learning_rate": 7.672716186159829e-05,
+ "loss": 3.5230023860931396,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12565445026178,
+ "grad_norm": 0.04888913407921791,
+ "learning_rate": 7.655457422959993e-05,
+ "loss": 3.525777816772461,
+ "step": 2817
+ },
+ {
+ "epoch": 39.139616055846425,
+ "grad_norm": 0.04879172518849373,
+ "learning_rate": 7.638216904059122e-05,
+ "loss": 3.5179781913757324,
+ "step": 2818
+ },
+ {
+ "epoch": 39.153577661431065,
+ "grad_norm": 0.04999098926782608,
+ "learning_rate": 7.620994643347559e-05,
+ "loss": 3.53208327293396,
+ "step": 2819
+ },
+ {
+ "epoch": 39.167539267015705,
+ "grad_norm": 0.051719557493925095,
+ "learning_rate": 7.60379065470098e-05,
+ "loss": 3.500316619873047,
+ "step": 2820
+ },
+ {
+ "epoch": 39.181500872600346,
+ "grad_norm": 0.05016033351421356,
+ "learning_rate": 7.586604951980326e-05,
+ "loss": 3.4964962005615234,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19546247818499,
+ "grad_norm": 0.048761315643787384,
+ "learning_rate": 7.569437549031813e-05,
+ "loss": 3.5171115398406982,
+ "step": 2822
+ },
+ {
+ "epoch": 39.20942408376963,
+ "grad_norm": 0.05029286816716194,
+ "learning_rate": 7.55228845968691e-05,
+ "loss": 3.5414791107177734,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223385689354274,
+ "grad_norm": 0.0502147302031517,
+ "learning_rate": 7.535157697762313e-05,
+ "loss": 3.519467830657959,
+ "step": 2824
+ },
+ {
+ "epoch": 39.23734729493892,
+ "grad_norm": 0.05098641291260719,
+ "learning_rate": 7.518045277059979e-05,
+ "loss": 3.539252519607544,
+ "step": 2825
+ },
+ {
+ "epoch": 39.25130890052356,
+ "grad_norm": 0.04990671947598457,
+ "learning_rate": 7.500951211367068e-05,
+ "loss": 3.5308918952941895,
+ "step": 2826
+ },
+ {
+ "epoch": 39.2652705061082,
+ "grad_norm": 0.04992193728685379,
+ "learning_rate": 7.483875514455979e-05,
+ "loss": 3.521590232849121,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27923211169284,
+ "grad_norm": 0.05016932636499405,
+ "learning_rate": 7.466818200084264e-05,
+ "loss": 3.5208449363708496,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29319371727749,
+ "grad_norm": 0.050798505544662476,
+ "learning_rate": 7.449779281994712e-05,
+ "loss": 3.5616073608398438,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30715532286213,
+ "grad_norm": 0.05047072842717171,
+ "learning_rate": 7.43275877391528e-05,
+ "loss": 3.540660858154297,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32111692844677,
+ "grad_norm": 0.051101770251989365,
+ "learning_rate": 7.415756689559061e-05,
+ "loss": 3.532931327819824,
+ "step": 2831
+ },
+ {
+ "epoch": 39.33507853403141,
+ "grad_norm": 0.049927644431591034,
+ "learning_rate": 7.398773042624341e-05,
+ "loss": 3.5045418739318848,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34904013961606,
+ "grad_norm": 0.05085239186882973,
+ "learning_rate": 7.381807846794532e-05,
+ "loss": 3.4864392280578613,
+ "step": 2833
+ },
+ {
+ "epoch": 39.3630017452007,
+ "grad_norm": 0.05125649273395538,
+ "learning_rate": 7.3648611157382e-05,
+ "loss": 3.51094913482666,
+ "step": 2834
+ },
+ {
+ "epoch": 39.37696335078534,
+ "grad_norm": 0.05072709918022156,
+ "learning_rate": 7.34793286310899e-05,
+ "loss": 3.531437873840332,
+ "step": 2835
+ },
+ {
+ "epoch": 39.390924956369986,
+ "grad_norm": 0.050076793879270554,
+ "learning_rate": 7.331023102545716e-05,
+ "loss": 3.5222954750061035,
+ "step": 2836
+ },
+ {
+ "epoch": 39.404886561954626,
+ "grad_norm": 0.050684161484241486,
+ "learning_rate": 7.314131847672268e-05,
+ "loss": 3.525660991668701,
+ "step": 2837
+ },
+ {
+ "epoch": 39.41884816753927,
+ "grad_norm": 0.05221536383032799,
+ "learning_rate": 7.297259112097608e-05,
+ "loss": 3.547537326812744,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43280977312391,
+ "grad_norm": 0.050070326775312424,
+ "learning_rate": 7.280404909415801e-05,
+ "loss": 3.524541139602661,
+ "step": 2839
+ },
+ {
+ "epoch": 39.446771378708554,
+ "grad_norm": 0.05107896402478218,
+ "learning_rate": 7.263569253205973e-05,
+ "loss": 3.4862544536590576,
+ "step": 2840
+ },
+ {
+ "epoch": 39.460732984293195,
+ "grad_norm": 0.05145040526986122,
+ "learning_rate": 7.24675215703232e-05,
+ "loss": 3.5242197513580322,
+ "step": 2841
+ },
+ {
+ "epoch": 39.474694589877835,
+ "grad_norm": 0.05019835755228996,
+ "learning_rate": 7.229953634444055e-05,
+ "loss": 3.5409276485443115,
+ "step": 2842
+ },
+ {
+ "epoch": 39.488656195462475,
+ "grad_norm": 0.051458705216646194,
+ "learning_rate": 7.213173698975444e-05,
+ "loss": 3.5637054443359375,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50261780104712,
+ "grad_norm": 0.05186862125992775,
+ "learning_rate": 7.1964123641458e-05,
+ "loss": 3.5018935203552246,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51657940663176,
+ "grad_norm": 0.05114978924393654,
+ "learning_rate": 7.179669643459405e-05,
+ "loss": 3.5413007736206055,
+ "step": 2845
+ },
+ {
+ "epoch": 39.5305410122164,
+ "grad_norm": 0.051313482224941254,
+ "learning_rate": 7.162945550405575e-05,
+ "loss": 3.5065975189208984,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544502617801044,
+ "grad_norm": 0.0509142242372036,
+ "learning_rate": 7.146240098458613e-05,
+ "loss": 3.521883964538574,
+ "step": 2847
+ },
+ {
+ "epoch": 39.55846422338569,
+ "grad_norm": 0.05225839465856552,
+ "learning_rate": 7.129553301077801e-05,
+ "loss": 3.534417152404785,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57242582897033,
+ "grad_norm": 0.05145575851202011,
+ "learning_rate": 7.112885171707378e-05,
+ "loss": 3.5341527462005615,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58638743455497,
+ "grad_norm": 0.0504966638982296,
+ "learning_rate": 7.096235723776562e-05,
+ "loss": 3.529374599456787,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60034904013962,
+ "grad_norm": 0.05249404534697533,
+ "learning_rate": 7.079604970699514e-05,
+ "loss": 3.537140130996704,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61431064572426,
+ "grad_norm": 0.052380647510290146,
+ "learning_rate": 7.062992925875332e-05,
+ "loss": 3.529085159301758,
+ "step": 2852
+ },
+ {
+ "epoch": 39.6282722513089,
+ "grad_norm": 0.05162700638175011,
+ "learning_rate": 7.046399602688045e-05,
+ "loss": 3.549781084060669,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64223385689354,
+ "grad_norm": 0.05245045945048332,
+ "learning_rate": 7.029825014506572e-05,
+ "loss": 3.5424249172210693,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65619546247819,
+ "grad_norm": 0.05116629973053932,
+ "learning_rate": 7.013269174684795e-05,
+ "loss": 3.5407791137695312,
+ "step": 2855
+ },
+ {
+ "epoch": 39.67015706806283,
+ "grad_norm": 0.0507299080491066,
+ "learning_rate": 6.996732096561422e-05,
+ "loss": 3.550326108932495,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68411867364747,
+ "grad_norm": 0.05050551891326904,
+ "learning_rate": 6.980213793460092e-05,
+ "loss": 3.50667667388916,
+ "step": 2857
+ },
+ {
+ "epoch": 39.69808027923211,
+ "grad_norm": 0.052720312029123306,
+ "learning_rate": 6.963714278689304e-05,
+ "loss": 3.5426530838012695,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712041884816756,
+ "grad_norm": 0.05040572211146355,
+ "learning_rate": 6.947233565542417e-05,
+ "loss": 3.50482177734375,
+ "step": 2859
+ },
+ {
+ "epoch": 39.726003490401396,
+ "grad_norm": 0.05204465985298157,
+ "learning_rate": 6.930771667297653e-05,
+ "loss": 3.5070583820343018,
+ "step": 2860
+ },
+ {
+ "epoch": 39.73996509598604,
+ "grad_norm": 0.05265089496970177,
+ "learning_rate": 6.914328597218042e-05,
+ "loss": 3.53663969039917,
+ "step": 2861
+ },
+ {
+ "epoch": 39.753926701570684,
+ "grad_norm": 0.05254863202571869,
+ "learning_rate": 6.8979043685515e-05,
+ "loss": 3.5625903606414795,
+ "step": 2862
+ },
+ {
+ "epoch": 39.767888307155324,
+ "grad_norm": 0.05163634940981865,
+ "learning_rate": 6.881498994530708e-05,
+ "loss": 3.541079044342041,
+ "step": 2863
+ },
+ {
+ "epoch": 39.781849912739965,
+ "grad_norm": 0.05227374657988548,
+ "learning_rate": 6.865112488373186e-05,
+ "loss": 3.52247953414917,
+ "step": 2864
+ },
+ {
+ "epoch": 39.795811518324605,
+ "grad_norm": 0.051065072417259216,
+ "learning_rate": 6.84874486328125e-05,
+ "loss": 3.5435099601745605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.80977312390925,
+ "grad_norm": 0.05301819369196892,
+ "learning_rate": 6.832396132441993e-05,
+ "loss": 3.551717758178711,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82373472949389,
+ "grad_norm": 0.052438024431467056,
+ "learning_rate": 6.816066309027309e-05,
+ "loss": 3.545943260192871,
+ "step": 2867
+ },
+ {
+ "epoch": 39.83769633507853,
+ "grad_norm": 0.05212542042136192,
+ "learning_rate": 6.799755406193815e-05,
+ "loss": 3.5160136222839355,
+ "step": 2868
+ },
+ {
+ "epoch": 39.85165794066317,
+ "grad_norm": 0.05442605912685394,
+ "learning_rate": 6.783463437082929e-05,
+ "loss": 3.5255932807922363,
+ "step": 2869
+ },
+ {
+ "epoch": 39.86561954624782,
+ "grad_norm": 0.05317089334130287,
+ "learning_rate": 6.767190414820792e-05,
+ "loss": 3.5328879356384277,
+ "step": 2870
+ },
+ {
+ "epoch": 39.87958115183246,
+ "grad_norm": 0.05377183482050896,
+ "learning_rate": 6.750936352518284e-05,
+ "loss": 3.512451171875,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8935427574171,
+ "grad_norm": 0.05238938704133034,
+ "learning_rate": 6.734701263271011e-05,
+ "loss": 3.5253477096557617,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90750436300174,
+ "grad_norm": 0.0536922849714756,
+ "learning_rate": 6.718485160159289e-05,
+ "loss": 3.532573938369751,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92146596858639,
+ "grad_norm": 0.05387766659259796,
+ "learning_rate": 6.702288056248145e-05,
+ "loss": 3.535872459411621,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93542757417103,
+ "grad_norm": 0.05426036939024925,
+ "learning_rate": 6.686109964587285e-05,
+ "loss": 3.534602165222168,
+ "step": 2875
+ },
+ {
+ "epoch": 39.94938917975567,
+ "grad_norm": 0.05351133272051811,
+ "learning_rate": 6.669950898211106e-05,
+ "loss": 3.507835865020752,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96335078534032,
+ "grad_norm": 0.05391782522201538,
+ "learning_rate": 6.653810870138676e-05,
+ "loss": 3.543053150177002,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97731239092496,
+ "grad_norm": 0.05348298326134682,
+ "learning_rate": 6.637689893373729e-05,
+ "loss": 3.528005361557007,
+ "step": 2878
+ },
+ {
+ "epoch": 39.9912739965096,
+ "grad_norm": 0.05260506272315979,
+ "learning_rate": 6.621587980904651e-05,
+ "loss": 3.542234420776367,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.040236134082078934,
+ "learning_rate": 6.605505145704437e-05,
+ "loss": 2.2050814628601074,
+ "step": 2880
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.228917698545582e+18,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-2880/training_args.bin b/runs/l2r90-wd0033-ssl/checkpoint-2880/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..313c41733ff55284c06034a59bb885b30e40ffc8
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-2880/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7ddd0bdddba68d3e19ab33477000f3fc684fe4ee8e3580eeb8c02646e5fc27ac
+size 4856
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3240/chat_template.jinja b/runs/l2r90-wd0033-ssl/checkpoint-3240/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3240/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3240/config.json b/runs/l2r90-wd0033-ssl/checkpoint-3240/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3240/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3240/generation_config.json b/runs/l2r90-wd0033-ssl/checkpoint-3240/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3240/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3240/model.safetensors b/runs/l2r90-wd0033-ssl/checkpoint-3240/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..3abfbb03f0ed055d2e78c1efe0c4557782bc1c25
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3240/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bb00dbf31d123f6c1dae6bfcf4fbb28c8b1c6b9e82e33940cc225f897f869a92
+size 583356232
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3240/optimizer.pt b/runs/l2r90-wd0033-ssl/checkpoint-3240/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..9afc77f81475a12a1b1cd38861c084100d7327d6
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3240/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e078e38b3e17d6baa17852253117ab529101c16d4dddea12a63f722f11fb6c2c
+size 1166827898
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3240/rng_state_0.pth b/runs/l2r90-wd0033-ssl/checkpoint-3240/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..6fb901f11f49c59b39a622f7767ba7f3f47a8e68
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3240/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c07d15af951fd54259f185b472ffc81cf9402e4a8ec70b56c9edd912c69850c4
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3240/rng_state_1.pth b/runs/l2r90-wd0033-ssl/checkpoint-3240/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..7c6be540f2709b4585034dcadb7428c21ac6b6e5
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3240/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9b6d8e0cbda51cdd8640c4f6ac246f9e2d927fedbdb6bbb9ef1e60d78520899d
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3240/scheduler.pt b/runs/l2r90-wd0033-ssl/checkpoint-3240/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..84479f3138d6041f59e136173e080a0ddbdf81b1
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3240/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1045dafc67457b916f466f6a7399234fef3fce01fc96958c0e146dea7784e901
+size 1064
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3240/tokenizer.json b/runs/l2r90-wd0033-ssl/checkpoint-3240/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3240/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3240/tokenizer_config.json b/runs/l2r90-wd0033-ssl/checkpoint-3240/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3240/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3240/trainer_state.json b/runs/l2r90-wd0033-ssl/checkpoint-3240/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..6a8e80101381b5978553f037b3bb2319db1d4250
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3240/trainer_state.json
@@ -0,0 +1,23058 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 45.0,
+ "eval_steps": 500,
+ "global_step": 3240,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013961605584642234,
+ "grad_norm": 0.26733073592185974,
+ "learning_rate": 0.0,
+ "loss": 7.990724086761475,
+ "step": 1
+ },
+ {
+ "epoch": 0.027923211169284468,
+ "grad_norm": 0.26712340116500854,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 7.990110874176025,
+ "step": 2
+ },
+ {
+ "epoch": 0.041884816753926704,
+ "grad_norm": 0.2646932601928711,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 7.956935405731201,
+ "step": 3
+ },
+ {
+ "epoch": 0.055846422338568937,
+ "grad_norm": 0.24684545397758484,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 7.890539646148682,
+ "step": 4
+ },
+ {
+ "epoch": 0.06980802792321117,
+ "grad_norm": 0.21016643941402435,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 7.8146138191223145,
+ "step": 5
+ },
+ {
+ "epoch": 0.08376963350785341,
+ "grad_norm": 0.17992742359638214,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 7.724307060241699,
+ "step": 6
+ },
+ {
+ "epoch": 0.09773123909249563,
+ "grad_norm": 0.14356893301010132,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 7.645662307739258,
+ "step": 7
+ },
+ {
+ "epoch": 0.11169284467713787,
+ "grad_norm": 0.10830964893102646,
+ "learning_rate": 4.2e-05,
+ "loss": 7.582019329071045,
+ "step": 8
+ },
+ {
+ "epoch": 0.1256544502617801,
+ "grad_norm": 0.08265161514282227,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 7.5230793952941895,
+ "step": 9
+ },
+ {
+ "epoch": 0.13961605584642234,
+ "grad_norm": 0.06423710286617279,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 7.473812103271484,
+ "step": 10
+ },
+ {
+ "epoch": 0.15357766143106458,
+ "grad_norm": 0.05132246017456055,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 7.440786838531494,
+ "step": 11
+ },
+ {
+ "epoch": 0.16753926701570682,
+ "grad_norm": 0.040572404861450195,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 7.41290283203125,
+ "step": 12
+ },
+ {
+ "epoch": 0.18150087260034903,
+ "grad_norm": 0.03138705715537071,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 7.39749002456665,
+ "step": 13
+ },
+ {
+ "epoch": 0.19546247818499127,
+ "grad_norm": 0.02553008496761322,
+ "learning_rate": 7.8e-05,
+ "loss": 7.387092590332031,
+ "step": 14
+ },
+ {
+ "epoch": 0.2094240837696335,
+ "grad_norm": 0.022178977727890015,
+ "learning_rate": 8.4e-05,
+ "loss": 7.378993511199951,
+ "step": 15
+ },
+ {
+ "epoch": 0.22338568935427575,
+ "grad_norm": 0.019915293902158737,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 7.37220573425293,
+ "step": 16
+ },
+ {
+ "epoch": 0.23734729493891799,
+ "grad_norm": 0.017390629276633263,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 7.366759300231934,
+ "step": 17
+ },
+ {
+ "epoch": 0.2513089005235602,
+ "grad_norm": 0.015472771599888802,
+ "learning_rate": 0.000102,
+ "loss": 7.3652663230896,
+ "step": 18
+ },
+ {
+ "epoch": 0.26527050610820246,
+ "grad_norm": 0.013599214144051075,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 7.359714984893799,
+ "step": 19
+ },
+ {
+ "epoch": 0.2792321116928447,
+ "grad_norm": 0.012442861683666706,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 7.353302001953125,
+ "step": 20
+ },
+ {
+ "epoch": 0.2931937172774869,
+ "grad_norm": 0.010471965186297894,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 7.357417583465576,
+ "step": 21
+ },
+ {
+ "epoch": 0.30715532286212915,
+ "grad_norm": 0.009647224098443985,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 7.351763725280762,
+ "step": 22
+ },
+ {
+ "epoch": 0.32111692844677137,
+ "grad_norm": 0.008228210732340813,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 7.345434188842773,
+ "step": 23
+ },
+ {
+ "epoch": 0.33507853403141363,
+ "grad_norm": 0.007454514969140291,
+ "learning_rate": 0.000138,
+ "loss": 7.342520236968994,
+ "step": 24
+ },
+ {
+ "epoch": 0.34904013961605584,
+ "grad_norm": 0.0069856527261435986,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 7.338870048522949,
+ "step": 25
+ },
+ {
+ "epoch": 0.36300174520069806,
+ "grad_norm": 0.006091665010899305,
+ "learning_rate": 0.00015,
+ "loss": 7.328121662139893,
+ "step": 26
+ },
+ {
+ "epoch": 0.3769633507853403,
+ "grad_norm": 0.005525320768356323,
+ "learning_rate": 0.000156,
+ "loss": 7.335029602050781,
+ "step": 27
+ },
+ {
+ "epoch": 0.39092495636998253,
+ "grad_norm": 0.0061133443377912045,
+ "learning_rate": 0.000162,
+ "loss": 7.319417953491211,
+ "step": 28
+ },
+ {
+ "epoch": 0.4048865619546248,
+ "grad_norm": 0.006040909793227911,
+ "learning_rate": 0.000168,
+ "loss": 7.326376438140869,
+ "step": 29
+ },
+ {
+ "epoch": 0.418848167539267,
+ "grad_norm": 0.006003216840326786,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 7.319552421569824,
+ "step": 30
+ },
+ {
+ "epoch": 0.4328097731239092,
+ "grad_norm": 0.007096003741025925,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 7.323486328125,
+ "step": 31
+ },
+ {
+ "epoch": 0.4467713787085515,
+ "grad_norm": 0.007725459989160299,
+ "learning_rate": 0.000186,
+ "loss": 7.3190717697143555,
+ "step": 32
+ },
+ {
+ "epoch": 0.4607329842931937,
+ "grad_norm": 0.007338670082390308,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 7.317154884338379,
+ "step": 33
+ },
+ {
+ "epoch": 0.47469458987783597,
+ "grad_norm": 0.007709544617682695,
+ "learning_rate": 0.000198,
+ "loss": 7.311826705932617,
+ "step": 34
+ },
+ {
+ "epoch": 0.4886561954624782,
+ "grad_norm": 0.009249784983694553,
+ "learning_rate": 0.000204,
+ "loss": 7.322498798370361,
+ "step": 35
+ },
+ {
+ "epoch": 0.5026178010471204,
+ "grad_norm": 0.006950767710804939,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 7.314571380615234,
+ "step": 36
+ },
+ {
+ "epoch": 0.5165794066317626,
+ "grad_norm": 0.007583985570818186,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 7.309540748596191,
+ "step": 37
+ },
+ {
+ "epoch": 0.5305410122164049,
+ "grad_norm": 0.014731820672750473,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 7.307004928588867,
+ "step": 38
+ },
+ {
+ "epoch": 0.5445026178010471,
+ "grad_norm": 0.016276473179459572,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 7.299642562866211,
+ "step": 39
+ },
+ {
+ "epoch": 0.5584642233856894,
+ "grad_norm": 0.020018689334392548,
+ "learning_rate": 0.000234,
+ "loss": 7.298466682434082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5724258289703316,
+ "grad_norm": 0.014941536821424961,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 7.289968967437744,
+ "step": 41
+ },
+ {
+ "epoch": 0.5863874345549738,
+ "grad_norm": 0.018299812451004982,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 7.276564598083496,
+ "step": 42
+ },
+ {
+ "epoch": 0.6003490401396161,
+ "grad_norm": 0.01615075021982193,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.2575554847717285,
+ "step": 43
+ },
+ {
+ "epoch": 0.6143106457242583,
+ "grad_norm": 0.013063831254839897,
+ "learning_rate": 0.000258,
+ "loss": 7.253885269165039,
+ "step": 44
+ },
+ {
+ "epoch": 0.6282722513089005,
+ "grad_norm": 0.015211676247417927,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.245779037475586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6422338568935427,
+ "grad_norm": 0.02093500830233097,
+ "learning_rate": 0.00027,
+ "loss": 7.225130081176758,
+ "step": 46
+ },
+ {
+ "epoch": 0.6561954624781849,
+ "grad_norm": 0.01808053068816662,
+ "learning_rate": 0.000276,
+ "loss": 7.209654331207275,
+ "step": 47
+ },
+ {
+ "epoch": 0.6701570680628273,
+ "grad_norm": 0.02209644205868244,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.197124481201172,
+ "step": 48
+ },
+ {
+ "epoch": 0.6841186736474695,
+ "grad_norm": 0.023524878546595573,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.1841936111450195,
+ "step": 49
+ },
+ {
+ "epoch": 0.6980802792321117,
+ "grad_norm": 0.031148087233304977,
+ "learning_rate": 0.000294,
+ "loss": 7.188040733337402,
+ "step": 50
+ },
+ {
+ "epoch": 0.7120418848167539,
+ "grad_norm": 0.02931929938495159,
+ "learning_rate": 0.0003,
+ "loss": 7.17371940612793,
+ "step": 51
+ },
+ {
+ "epoch": 0.7260034904013961,
+ "grad_norm": 0.01611483097076416,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.1714324951171875,
+ "step": 52
+ },
+ {
+ "epoch": 0.7399650959860384,
+ "grad_norm": 0.02619907818734646,
+ "learning_rate": 0.000312,
+ "loss": 7.158261299133301,
+ "step": 53
+ },
+ {
+ "epoch": 0.7539267015706806,
+ "grad_norm": 0.049462560564279556,
+ "learning_rate": 0.000318,
+ "loss": 7.167076110839844,
+ "step": 54
+ },
+ {
+ "epoch": 0.7678883071553229,
+ "grad_norm": 0.09835070371627808,
+ "learning_rate": 0.000324,
+ "loss": 7.154292106628418,
+ "step": 55
+ },
+ {
+ "epoch": 0.7818499127399651,
+ "grad_norm": 0.11406510323286057,
+ "learning_rate": 0.00033,
+ "loss": 7.162016868591309,
+ "step": 56
+ },
+ {
+ "epoch": 0.7958115183246073,
+ "grad_norm": 0.027026303112506866,
+ "learning_rate": 0.000336,
+ "loss": 7.141416549682617,
+ "step": 57
+ },
+ {
+ "epoch": 0.8097731239092496,
+ "grad_norm": 0.06690579652786255,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.13959264755249,
+ "step": 58
+ },
+ {
+ "epoch": 0.8237347294938918,
+ "grad_norm": 0.020919431000947952,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.126513481140137,
+ "step": 59
+ },
+ {
+ "epoch": 0.837696335078534,
+ "grad_norm": 0.05618130415678024,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.124345779418945,
+ "step": 60
+ },
+ {
+ "epoch": 0.8516579406631762,
+ "grad_norm": 0.03402302414178848,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.100615501403809,
+ "step": 61
+ },
+ {
+ "epoch": 0.8656195462478184,
+ "grad_norm": 0.03646906837821007,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.116507530212402,
+ "step": 62
+ },
+ {
+ "epoch": 0.8795811518324608,
+ "grad_norm": 0.03076397068798542,
+ "learning_rate": 0.000372,
+ "loss": 7.097659111022949,
+ "step": 63
+ },
+ {
+ "epoch": 0.893542757417103,
+ "grad_norm": 0.03329646959900856,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.097290992736816,
+ "step": 64
+ },
+ {
+ "epoch": 0.9075043630017452,
+ "grad_norm": 0.01998630352318287,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.092268943786621,
+ "step": 65
+ },
+ {
+ "epoch": 0.9214659685863874,
+ "grad_norm": 0.03278940171003342,
+ "learning_rate": 0.00039,
+ "loss": 7.088292121887207,
+ "step": 66
+ },
+ {
+ "epoch": 0.9354275741710296,
+ "grad_norm": 0.024620421230793,
+ "learning_rate": 0.000396,
+ "loss": 7.078650951385498,
+ "step": 67
+ },
+ {
+ "epoch": 0.9493891797556719,
+ "grad_norm": 0.03750383481383324,
+ "learning_rate": 0.000402,
+ "loss": 7.052778244018555,
+ "step": 68
+ },
+ {
+ "epoch": 0.9633507853403142,
+ "grad_norm": 0.07558075338602066,
+ "learning_rate": 0.000408,
+ "loss": 7.045991897583008,
+ "step": 69
+ },
+ {
+ "epoch": 0.9773123909249564,
+ "grad_norm": 0.16021482646465302,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.023152828216553,
+ "step": 70
+ },
+ {
+ "epoch": 0.9912739965095986,
+ "grad_norm": 0.24981893599033356,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.067540645599365,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.10255520790815353,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 4.425640106201172,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8861758708953857,
+ "eval_runtime": 58.2699,
+ "eval_samples_per_second": 41.908,
+ "eval_steps_per_second": 0.669,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139616055846423,
+ "grad_norm": 0.13374797999858856,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.073373317718506,
+ "step": 73
+ },
+ {
+ "epoch": 1.0279232111692844,
+ "grad_norm": 0.05079387128353119,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.0243306159973145,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418848167539267,
+ "grad_norm": 0.10056068748235703,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.014476299285889,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558464223385688,
+ "grad_norm": 0.06995758414268494,
+ "learning_rate": 0.00045,
+ "loss": 7.011931419372559,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698080279232112,
+ "grad_norm": 0.0483892522752285,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 6.9822797775268555,
+ "step": 77
+ },
+ {
+ "epoch": 1.0837696335078535,
+ "grad_norm": 0.08408170193433762,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 6.964348793029785,
+ "step": 78
+ },
+ {
+ "epoch": 1.0977312390924956,
+ "grad_norm": 0.08027710020542145,
+ "learning_rate": 0.000468,
+ "loss": 6.96852445602417,
+ "step": 79
+ },
+ {
+ "epoch": 1.111692844677138,
+ "grad_norm": 0.05647159367799759,
+ "learning_rate": 0.000474,
+ "loss": 6.944484710693359,
+ "step": 80
+ },
+ {
+ "epoch": 1.12565445026178,
+ "grad_norm": 0.07464330643415451,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 6.9246745109558105,
+ "step": 81
+ },
+ {
+ "epoch": 1.1396160558464223,
+ "grad_norm": 0.04451674595475197,
+ "learning_rate": 0.000486,
+ "loss": 6.895403861999512,
+ "step": 82
+ },
+ {
+ "epoch": 1.1535776614310647,
+ "grad_norm": 0.05820256471633911,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 6.887338638305664,
+ "step": 83
+ },
+ {
+ "epoch": 1.1675392670157068,
+ "grad_norm": 0.02925296127796173,
+ "learning_rate": 0.000498,
+ "loss": 6.857797622680664,
+ "step": 84
+ },
+ {
+ "epoch": 1.181500872600349,
+ "grad_norm": 0.05096987634897232,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 6.854822158813477,
+ "step": 85
+ },
+ {
+ "epoch": 1.1954624781849912,
+ "grad_norm": 0.03483197093009949,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 6.840914726257324,
+ "step": 86
+ },
+ {
+ "epoch": 1.2094240837696335,
+ "grad_norm": 0.03323264792561531,
+ "learning_rate": 0.000516,
+ "loss": 6.834497451782227,
+ "step": 87
+ },
+ {
+ "epoch": 1.2233856893542758,
+ "grad_norm": 0.03290868178009987,
+ "learning_rate": 0.000522,
+ "loss": 6.820746898651123,
+ "step": 88
+ },
+ {
+ "epoch": 1.237347294938918,
+ "grad_norm": 0.025398844853043556,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 6.795746803283691,
+ "step": 89
+ },
+ {
+ "epoch": 1.2513089005235603,
+ "grad_norm": 0.030700596049427986,
+ "learning_rate": 0.000534,
+ "loss": 6.7867889404296875,
+ "step": 90
+ },
+ {
+ "epoch": 1.2652705061082026,
+ "grad_norm": 0.019589420408010483,
+ "learning_rate": 0.00054,
+ "loss": 6.77443790435791,
+ "step": 91
+ },
+ {
+ "epoch": 1.2792321116928447,
+ "grad_norm": 0.03313920646905899,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 6.762271881103516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2931937172774868,
+ "grad_norm": 0.025389468297362328,
+ "learning_rate": 0.000552,
+ "loss": 6.750258445739746,
+ "step": 93
+ },
+ {
+ "epoch": 1.307155322862129,
+ "grad_norm": 0.040296655148267746,
+ "learning_rate": 0.000558,
+ "loss": 6.729808807373047,
+ "step": 94
+ },
+ {
+ "epoch": 1.3211169284467714,
+ "grad_norm": 0.04735933244228363,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 6.737982749938965,
+ "step": 95
+ },
+ {
+ "epoch": 1.3350785340314135,
+ "grad_norm": 0.05529680848121643,
+ "learning_rate": 0.00057,
+ "loss": 6.735714912414551,
+ "step": 96
+ },
+ {
+ "epoch": 1.3490401396160558,
+ "grad_norm": 0.07148554176092148,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 6.708859920501709,
+ "step": 97
+ },
+ {
+ "epoch": 1.3630017452006982,
+ "grad_norm": 0.059708479791879654,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 6.708860397338867,
+ "step": 98
+ },
+ {
+ "epoch": 1.3769633507853403,
+ "grad_norm": 0.027111195027828217,
+ "learning_rate": 0.000588,
+ "loss": 6.686685085296631,
+ "step": 99
+ },
+ {
+ "epoch": 1.3909249563699826,
+ "grad_norm": 0.0500967763364315,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 6.676325798034668,
+ "step": 100
+ },
+ {
+ "epoch": 1.404886561954625,
+ "grad_norm": 0.044565796852111816,
+ "learning_rate": 0.0006,
+ "loss": 6.674530029296875,
+ "step": 101
+ },
+ {
+ "epoch": 1.418848167539267,
+ "grad_norm": 0.03608180209994316,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 6.653741359710693,
+ "step": 102
+ },
+ {
+ "epoch": 1.4328097731239091,
+ "grad_norm": 0.030394982546567917,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 6.643152236938477,
+ "step": 103
+ },
+ {
+ "epoch": 1.4467713787085514,
+ "grad_norm": 0.040185656398534775,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 6.628686904907227,
+ "step": 104
+ },
+ {
+ "epoch": 1.4607329842931938,
+ "grad_norm": 0.021049322560429573,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 6.610864639282227,
+ "step": 105
+ },
+ {
+ "epoch": 1.4746945898778359,
+ "grad_norm": 0.04164277762174606,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 6.602880954742432,
+ "step": 106
+ },
+ {
+ "epoch": 1.4886561954624782,
+ "grad_norm": 0.0389864556491375,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 6.599456787109375,
+ "step": 107
+ },
+ {
+ "epoch": 1.5026178010471205,
+ "grad_norm": 0.05805973336100578,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 6.588700294494629,
+ "step": 108
+ },
+ {
+ "epoch": 1.5165794066317626,
+ "grad_norm": 0.08825036883354187,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 6.587496757507324,
+ "step": 109
+ },
+ {
+ "epoch": 1.530541012216405,
+ "grad_norm": 0.12521523237228394,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 6.626501083374023,
+ "step": 110
+ },
+ {
+ "epoch": 1.5445026178010473,
+ "grad_norm": 0.07429645210504532,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 6.590241432189941,
+ "step": 111
+ },
+ {
+ "epoch": 1.5584642233856894,
+ "grad_norm": 0.06463009119033813,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 6.578714370727539,
+ "step": 112
+ },
+ {
+ "epoch": 1.5724258289703315,
+ "grad_norm": 0.07508203387260437,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 6.561270713806152,
+ "step": 113
+ },
+ {
+ "epoch": 1.5863874345549738,
+ "grad_norm": 0.04125197231769562,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 6.558821678161621,
+ "step": 114
+ },
+ {
+ "epoch": 1.600349040139616,
+ "grad_norm": 0.05457386001944542,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.541611671447754,
+ "step": 115
+ },
+ {
+ "epoch": 1.6143106457242582,
+ "grad_norm": 0.03684385493397713,
+ "learning_rate": 0.000599973080557108,
+ "loss": 6.547738075256348,
+ "step": 116
+ },
+ {
+ "epoch": 1.6282722513089005,
+ "grad_norm": 0.05888746306300163,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.529491424560547,
+ "step": 117
+ },
+ {
+ "epoch": 1.6422338568935428,
+ "grad_norm": 0.056668370962142944,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.519444942474365,
+ "step": 118
+ },
+ {
+ "epoch": 1.656195462478185,
+ "grad_norm": 0.09123896062374115,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.522984981536865,
+ "step": 119
+ },
+ {
+ "epoch": 1.6701570680628273,
+ "grad_norm": 0.12380023300647736,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.515375137329102,
+ "step": 120
+ },
+ {
+ "epoch": 1.6841186736474696,
+ "grad_norm": 0.06953322887420654,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.500491142272949,
+ "step": 121
+ },
+ {
+ "epoch": 1.6980802792321117,
+ "grad_norm": 0.08613869547843933,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.5097198486328125,
+ "step": 122
+ },
+ {
+ "epoch": 1.7120418848167538,
+ "grad_norm": 0.08296383172273636,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.487521171569824,
+ "step": 123
+ },
+ {
+ "epoch": 1.7260034904013961,
+ "grad_norm": 0.05744783207774162,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.485134124755859,
+ "step": 124
+ },
+ {
+ "epoch": 1.7399650959860384,
+ "grad_norm": 0.04108566418290138,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.466182708740234,
+ "step": 125
+ },
+ {
+ "epoch": 1.7539267015706805,
+ "grad_norm": 0.05374931916594505,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.479011535644531,
+ "step": 126
+ },
+ {
+ "epoch": 1.7678883071553229,
+ "grad_norm": 0.03867914527654648,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.465902328491211,
+ "step": 127
+ },
+ {
+ "epoch": 1.7818499127399652,
+ "grad_norm": 0.05477374792098999,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.453082084655762,
+ "step": 128
+ },
+ {
+ "epoch": 1.7958115183246073,
+ "grad_norm": 0.06508298963308334,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.446089744567871,
+ "step": 129
+ },
+ {
+ "epoch": 1.8097731239092496,
+ "grad_norm": 0.06388624757528305,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.451748847961426,
+ "step": 130
+ },
+ {
+ "epoch": 1.823734729493892,
+ "grad_norm": 0.04223188757896423,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.4492387771606445,
+ "step": 131
+ },
+ {
+ "epoch": 1.837696335078534,
+ "grad_norm": 0.04528658092021942,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.409390926361084,
+ "step": 132
+ },
+ {
+ "epoch": 1.8516579406631761,
+ "grad_norm": 0.04962967336177826,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.415774345397949,
+ "step": 133
+ },
+ {
+ "epoch": 1.8656195462478184,
+ "grad_norm": 0.031944390386343,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.411528587341309,
+ "step": 134
+ },
+ {
+ "epoch": 1.8795811518324608,
+ "grad_norm": 0.028411149978637695,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.392329216003418,
+ "step": 135
+ },
+ {
+ "epoch": 1.8935427574171029,
+ "grad_norm": 0.03516148403286934,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.392238616943359,
+ "step": 136
+ },
+ {
+ "epoch": 1.9075043630017452,
+ "grad_norm": 0.027800120413303375,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.39179801940918,
+ "step": 137
+ },
+ {
+ "epoch": 1.9214659685863875,
+ "grad_norm": 0.02362987957894802,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.378960609436035,
+ "step": 138
+ },
+ {
+ "epoch": 1.9354275741710296,
+ "grad_norm": 0.033999308943748474,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.36802339553833,
+ "step": 139
+ },
+ {
+ "epoch": 1.949389179755672,
+ "grad_norm": 0.03290879726409912,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.366308689117432,
+ "step": 140
+ },
+ {
+ "epoch": 1.9633507853403143,
+ "grad_norm": 0.0286843404173851,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.350399971008301,
+ "step": 141
+ },
+ {
+ "epoch": 1.9773123909249564,
+ "grad_norm": 0.027145877480506897,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.33477783203125,
+ "step": 142
+ },
+ {
+ "epoch": 1.9912739965095985,
+ "grad_norm": 0.03530878946185112,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.322859764099121,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.028133150190114975,
+ "learning_rate": 0.000599778806120077,
+ "loss": 3.955288887023926,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7927101254463196,
+ "eval_runtime": 60.8372,
+ "eval_samples_per_second": 40.14,
+ "eval_steps_per_second": 0.641,
+ "step": 144
+ },
+ {
+ "epoch": 2.013961605584642,
+ "grad_norm": 0.03054557554423809,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.311975479125977,
+ "step": 145
+ },
+ {
+ "epoch": 2.0279232111692846,
+ "grad_norm": 0.05406459793448448,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.3030853271484375,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418848167539267,
+ "grad_norm": 0.10911019891500473,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.329056739807129,
+ "step": 147
+ },
+ {
+ "epoch": 2.055846422338569,
+ "grad_norm": 0.13948315382003784,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.362421035766602,
+ "step": 148
+ },
+ {
+ "epoch": 2.0698080279232114,
+ "grad_norm": 0.13176991045475006,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.3888139724731445,
+ "step": 149
+ },
+ {
+ "epoch": 2.0837696335078535,
+ "grad_norm": 0.09615885466337204,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.35154914855957,
+ "step": 150
+ },
+ {
+ "epoch": 2.0977312390924956,
+ "grad_norm": 0.06462561339139938,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.335790634155273,
+ "step": 151
+ },
+ {
+ "epoch": 2.1116928446771377,
+ "grad_norm": 0.05186621472239494,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.324396133422852,
+ "step": 152
+ },
+ {
+ "epoch": 2.1256544502617802,
+ "grad_norm": 0.06095981225371361,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.318549156188965,
+ "step": 153
+ },
+ {
+ "epoch": 2.1396160558464223,
+ "grad_norm": 0.036121442914009094,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.300607681274414,
+ "step": 154
+ },
+ {
+ "epoch": 2.1535776614310644,
+ "grad_norm": 0.055716633796691895,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.307400703430176,
+ "step": 155
+ },
+ {
+ "epoch": 2.167539267015707,
+ "grad_norm": 0.037626396864652634,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.287364959716797,
+ "step": 156
+ },
+ {
+ "epoch": 2.181500872600349,
+ "grad_norm": 0.04329098388552666,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.28244686126709,
+ "step": 157
+ },
+ {
+ "epoch": 2.195462478184991,
+ "grad_norm": 0.031868450343608856,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.2602410316467285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2094240837696333,
+ "grad_norm": 0.03647344186902046,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.263988971710205,
+ "step": 159
+ },
+ {
+ "epoch": 2.223385689354276,
+ "grad_norm": 0.03141765296459198,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.263205051422119,
+ "step": 160
+ },
+ {
+ "epoch": 2.237347294938918,
+ "grad_norm": 0.028253229334950447,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.253107070922852,
+ "step": 161
+ },
+ {
+ "epoch": 2.25130890052356,
+ "grad_norm": 0.03608971834182739,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.2571702003479,
+ "step": 162
+ },
+ {
+ "epoch": 2.2652705061082026,
+ "grad_norm": 0.02644292451441288,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.241732597351074,
+ "step": 163
+ },
+ {
+ "epoch": 2.2792321116928447,
+ "grad_norm": 0.028604455292224884,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.230289459228516,
+ "step": 164
+ },
+ {
+ "epoch": 2.2931937172774868,
+ "grad_norm": 0.023704614490270615,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.226101398468018,
+ "step": 165
+ },
+ {
+ "epoch": 2.3071553228621293,
+ "grad_norm": 0.020691027864813805,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.2096405029296875,
+ "step": 166
+ },
+ {
+ "epoch": 2.3211169284467714,
+ "grad_norm": 0.025598427280783653,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.22075080871582,
+ "step": 167
+ },
+ {
+ "epoch": 2.3350785340314135,
+ "grad_norm": 0.030546002089977264,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.216886520385742,
+ "step": 168
+ },
+ {
+ "epoch": 2.349040139616056,
+ "grad_norm": 0.026066245511174202,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.2022504806518555,
+ "step": 169
+ },
+ {
+ "epoch": 2.363001745200698,
+ "grad_norm": 0.023802239447832108,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.189509391784668,
+ "step": 170
+ },
+ {
+ "epoch": 2.3769633507853403,
+ "grad_norm": 0.029805002734065056,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.186277389526367,
+ "step": 171
+ },
+ {
+ "epoch": 2.3909249563699824,
+ "grad_norm": 0.043327830731868744,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.164876937866211,
+ "step": 172
+ },
+ {
+ "epoch": 2.404886561954625,
+ "grad_norm": 0.058787211775779724,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.179936408996582,
+ "step": 173
+ },
+ {
+ "epoch": 2.418848167539267,
+ "grad_norm": 0.0787397176027298,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.1830244064331055,
+ "step": 174
+ },
+ {
+ "epoch": 2.432809773123909,
+ "grad_norm": 0.0899854376912117,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.190343856811523,
+ "step": 175
+ },
+ {
+ "epoch": 2.4467713787085517,
+ "grad_norm": 0.06520707905292511,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.185057163238525,
+ "step": 176
+ },
+ {
+ "epoch": 2.4607329842931938,
+ "grad_norm": 0.1002504974603653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.171279430389404,
+ "step": 177
+ },
+ {
+ "epoch": 2.474694589877836,
+ "grad_norm": 0.09591004997491837,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.174382209777832,
+ "step": 178
+ },
+ {
+ "epoch": 2.488656195462478,
+ "grad_norm": 0.12127121537923813,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.185717582702637,
+ "step": 179
+ },
+ {
+ "epoch": 2.5026178010471205,
+ "grad_norm": 0.09640590846538544,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.196681976318359,
+ "step": 180
+ },
+ {
+ "epoch": 2.5165794066317626,
+ "grad_norm": 0.048341453075408936,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.158774375915527,
+ "step": 181
+ },
+ {
+ "epoch": 2.530541012216405,
+ "grad_norm": 0.08206334710121155,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.17445707321167,
+ "step": 182
+ },
+ {
+ "epoch": 2.5445026178010473,
+ "grad_norm": 0.038934968411922455,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.170527458190918,
+ "step": 183
+ },
+ {
+ "epoch": 2.5584642233856894,
+ "grad_norm": 0.060987457633018494,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.1501994132995605,
+ "step": 184
+ },
+ {
+ "epoch": 2.5724258289703315,
+ "grad_norm": 0.033269546926021576,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.144186019897461,
+ "step": 185
+ },
+ {
+ "epoch": 2.5863874345549736,
+ "grad_norm": 0.046137236058712006,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.1236677169799805,
+ "step": 186
+ },
+ {
+ "epoch": 2.600349040139616,
+ "grad_norm": 0.038265109062194824,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.121979236602783,
+ "step": 187
+ },
+ {
+ "epoch": 2.614310645724258,
+ "grad_norm": 0.039128001779317856,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.124156951904297,
+ "step": 188
+ },
+ {
+ "epoch": 2.6282722513089007,
+ "grad_norm": 0.025303540751338005,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.108036994934082,
+ "step": 189
+ },
+ {
+ "epoch": 2.642233856893543,
+ "grad_norm": 0.04441308230161667,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.1211161613464355,
+ "step": 190
+ },
+ {
+ "epoch": 2.656195462478185,
+ "grad_norm": 0.03083532303571701,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.094561576843262,
+ "step": 191
+ },
+ {
+ "epoch": 2.670157068062827,
+ "grad_norm": 0.02951296605169773,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.091608047485352,
+ "step": 192
+ },
+ {
+ "epoch": 2.6841186736474696,
+ "grad_norm": 0.02385987527668476,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.087080001831055,
+ "step": 193
+ },
+ {
+ "epoch": 2.6980802792321117,
+ "grad_norm": 0.03315780311822891,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.088132381439209,
+ "step": 194
+ },
+ {
+ "epoch": 2.712041884816754,
+ "grad_norm": 0.02410835586488247,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.081460952758789,
+ "step": 195
+ },
+ {
+ "epoch": 2.7260034904013963,
+ "grad_norm": 0.033069487661123276,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.0696539878845215,
+ "step": 196
+ },
+ {
+ "epoch": 2.7399650959860384,
+ "grad_norm": 0.031288471072912216,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.047612190246582,
+ "step": 197
+ },
+ {
+ "epoch": 2.7539267015706805,
+ "grad_norm": 0.03399677947163582,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.019026756286621,
+ "step": 198
+ },
+ {
+ "epoch": 2.7678883071553226,
+ "grad_norm": 0.031948331743478775,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.050133228302002,
+ "step": 199
+ },
+ {
+ "epoch": 2.781849912739965,
+ "grad_norm": 0.049206435680389404,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.045332431793213,
+ "step": 200
+ },
+ {
+ "epoch": 2.7958115183246073,
+ "grad_norm": 0.07528849691152573,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.059202194213867,
+ "step": 201
+ },
+ {
+ "epoch": 2.80977312390925,
+ "grad_norm": 0.11815470457077026,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.077486038208008,
+ "step": 202
+ },
+ {
+ "epoch": 2.823734729493892,
+ "grad_norm": 0.0867348462343216,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.045670032501221,
+ "step": 203
+ },
+ {
+ "epoch": 2.837696335078534,
+ "grad_norm": 0.05716705322265625,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.034373760223389,
+ "step": 204
+ },
+ {
+ "epoch": 2.851657940663176,
+ "grad_norm": 0.05944284051656723,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.049911975860596,
+ "step": 205
+ },
+ {
+ "epoch": 2.8656195462478182,
+ "grad_norm": 0.03220470994710922,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.039861679077148,
+ "step": 206
+ },
+ {
+ "epoch": 2.8795811518324608,
+ "grad_norm": 0.057081177830696106,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.010714530944824,
+ "step": 207
+ },
+ {
+ "epoch": 2.893542757417103,
+ "grad_norm": 0.04037516191601753,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.018877983093262,
+ "step": 208
+ },
+ {
+ "epoch": 2.9075043630017454,
+ "grad_norm": 0.036492880433797836,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.009493827819824,
+ "step": 209
+ },
+ {
+ "epoch": 2.9214659685863875,
+ "grad_norm": 0.04785023257136345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.0140862464904785,
+ "step": 210
+ },
+ {
+ "epoch": 2.9354275741710296,
+ "grad_norm": 0.07721036672592163,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.030359268188477,
+ "step": 211
+ },
+ {
+ "epoch": 2.9493891797556717,
+ "grad_norm": 0.14220190048217773,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.045989990234375,
+ "step": 212
+ },
+ {
+ "epoch": 2.9633507853403143,
+ "grad_norm": 0.10265158116817474,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.043968200683594,
+ "step": 213
+ },
+ {
+ "epoch": 2.9773123909249564,
+ "grad_norm": 0.051398951560258865,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.022278785705566,
+ "step": 214
+ },
+ {
+ "epoch": 2.9912739965095985,
+ "grad_norm": 0.06044618785381317,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.017296314239502,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.03284750506281853,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 3.7544798851013184,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7531244158744812,
+ "eval_runtime": 60.3986,
+ "eval_samples_per_second": 40.431,
+ "eval_steps_per_second": 0.646,
+ "step": 216
+ },
+ {
+ "epoch": 3.013961605584642,
+ "grad_norm": 0.03895840048789978,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 5.9985857009887695,
+ "step": 217
+ },
+ {
+ "epoch": 3.0279232111692846,
+ "grad_norm": 0.04554992541670799,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 5.984710693359375,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418848167539267,
+ "grad_norm": 0.03359116613864899,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 5.984078407287598,
+ "step": 219
+ },
+ {
+ "epoch": 3.055846422338569,
+ "grad_norm": 0.035171881318092346,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 5.979424476623535,
+ "step": 220
+ },
+ {
+ "epoch": 3.0698080279232114,
+ "grad_norm": 0.032774072140455246,
+ "learning_rate": 0.000598278794687186,
+ "loss": 5.94467830657959,
+ "step": 221
+ },
+ {
+ "epoch": 3.0837696335078535,
+ "grad_norm": 0.032842908054590225,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 5.95986270904541,
+ "step": 222
+ },
+ {
+ "epoch": 3.0977312390924956,
+ "grad_norm": 0.02870488539338112,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 5.951186656951904,
+ "step": 223
+ },
+ {
+ "epoch": 3.1116928446771377,
+ "grad_norm": 0.030777374282479286,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 5.942493438720703,
+ "step": 224
+ },
+ {
+ "epoch": 3.1256544502617802,
+ "grad_norm": 0.026439018547534943,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 5.938275337219238,
+ "step": 225
+ },
+ {
+ "epoch": 3.1396160558464223,
+ "grad_norm": 0.028689472004771233,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 5.931982040405273,
+ "step": 226
+ },
+ {
+ "epoch": 3.1535776614310644,
+ "grad_norm": 0.028100362047553062,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 5.919692039489746,
+ "step": 227
+ },
+ {
+ "epoch": 3.167539267015707,
+ "grad_norm": 0.02407177910208702,
+ "learning_rate": 0.00059807235435179,
+ "loss": 5.909029006958008,
+ "step": 228
+ },
+ {
+ "epoch": 3.181500872600349,
+ "grad_norm": 0.0237748846411705,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 5.9131598472595215,
+ "step": 229
+ },
+ {
+ "epoch": 3.195462478184991,
+ "grad_norm": 0.022418556734919548,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 5.909396171569824,
+ "step": 230
+ },
+ {
+ "epoch": 3.2094240837696333,
+ "grad_norm": 0.024656571447849274,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 5.885243892669678,
+ "step": 231
+ },
+ {
+ "epoch": 3.223385689354276,
+ "grad_norm": 0.02136118710041046,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 5.898958206176758,
+ "step": 232
+ },
+ {
+ "epoch": 3.237347294938918,
+ "grad_norm": 0.025290116667747498,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 5.886986255645752,
+ "step": 233
+ },
+ {
+ "epoch": 3.25130890052356,
+ "grad_norm": 0.02656404674053192,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 5.897647857666016,
+ "step": 234
+ },
+ {
+ "epoch": 3.2652705061082026,
+ "grad_norm": 0.019188793376088142,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 5.875659942626953,
+ "step": 235
+ },
+ {
+ "epoch": 3.2792321116928447,
+ "grad_norm": 0.02308926172554493,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 5.859380722045898,
+ "step": 236
+ },
+ {
+ "epoch": 3.2931937172774868,
+ "grad_norm": 0.029775725677609444,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 5.873961448669434,
+ "step": 237
+ },
+ {
+ "epoch": 3.3071553228621293,
+ "grad_norm": 0.03875259310007095,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 5.863171100616455,
+ "step": 238
+ },
+ {
+ "epoch": 3.3211169284467714,
+ "grad_norm": 0.052446842193603516,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 5.886547088623047,
+ "step": 239
+ },
+ {
+ "epoch": 3.3350785340314135,
+ "grad_norm": 0.05583837628364563,
+ "learning_rate": 0.000597691359743243,
+ "loss": 5.8849287033081055,
+ "step": 240
+ },
+ {
+ "epoch": 3.349040139616056,
+ "grad_norm": 0.034729067236185074,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 5.863419532775879,
+ "step": 241
+ },
+ {
+ "epoch": 3.363001745200698,
+ "grad_norm": 0.03590013459324837,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 5.852169990539551,
+ "step": 242
+ },
+ {
+ "epoch": 3.3769633507853403,
+ "grad_norm": 0.035041507333517075,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 5.836690902709961,
+ "step": 243
+ },
+ {
+ "epoch": 3.3909249563699824,
+ "grad_norm": 0.02857939340174198,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 5.836485862731934,
+ "step": 244
+ },
+ {
+ "epoch": 3.404886561954625,
+ "grad_norm": 0.03176642581820488,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 5.860719203948975,
+ "step": 245
+ },
+ {
+ "epoch": 3.418848167539267,
+ "grad_norm": 0.026105748489499092,
+ "learning_rate": 0.000597488038559129,
+ "loss": 5.8354949951171875,
+ "step": 246
+ },
+ {
+ "epoch": 3.432809773123909,
+ "grad_norm": 0.029094981029629707,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 5.846247673034668,
+ "step": 247
+ },
+ {
+ "epoch": 3.4467713787085517,
+ "grad_norm": 0.02388785220682621,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 5.837607383728027,
+ "step": 248
+ },
+ {
+ "epoch": 3.4607329842931938,
+ "grad_norm": 0.03094472736120224,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 5.832090377807617,
+ "step": 249
+ },
+ {
+ "epoch": 3.474694589877836,
+ "grad_norm": 0.03227658197283745,
+ "learning_rate": 0.000597347745466109,
+ "loss": 5.83722448348999,
+ "step": 250
+ },
+ {
+ "epoch": 3.488656195462478,
+ "grad_norm": 0.042813923209905624,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 5.851722717285156,
+ "step": 251
+ },
+ {
+ "epoch": 3.5026178010471205,
+ "grad_norm": 0.05825525149703026,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.795719146728516,
+ "step": 252
+ },
+ {
+ "epoch": 3.5165794066317626,
+ "grad_norm": 0.0867447778582573,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.813750267028809,
+ "step": 253
+ },
+ {
+ "epoch": 3.530541012216405,
+ "grad_norm": 0.09192710369825363,
+ "learning_rate": 0.000597203657966319,
+ "loss": 5.823488235473633,
+ "step": 254
+ },
+ {
+ "epoch": 3.5445026178010473,
+ "grad_norm": 0.06069248169660568,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 5.814342021942139,
+ "step": 255
+ },
+ {
+ "epoch": 3.5584642233856894,
+ "grad_norm": 0.06259952485561371,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 5.834272384643555,
+ "step": 256
+ },
+ {
+ "epoch": 3.5724258289703315,
+ "grad_norm": 0.0513639897108078,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.807951927185059,
+ "step": 257
+ },
+ {
+ "epoch": 3.5863874345549736,
+ "grad_norm": 0.03327798470854759,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.790238857269287,
+ "step": 258
+ },
+ {
+ "epoch": 3.600349040139616,
+ "grad_norm": 0.03963128477334976,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.7956695556640625,
+ "step": 259
+ },
+ {
+ "epoch": 3.614310645724258,
+ "grad_norm": 0.031153453513979912,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 5.7874531745910645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6282722513089007,
+ "grad_norm": 0.031419433653354645,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 5.779650688171387,
+ "step": 261
+ },
+ {
+ "epoch": 3.642233856893543,
+ "grad_norm": 0.0288684144616127,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.776577949523926,
+ "step": 262
+ },
+ {
+ "epoch": 3.656195462478185,
+ "grad_norm": 0.028315918520092964,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.778571128845215,
+ "step": 263
+ },
+ {
+ "epoch": 3.670157068062827,
+ "grad_norm": 0.024977795779705048,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 5.776470184326172,
+ "step": 264
+ },
+ {
+ "epoch": 3.6841186736474696,
+ "grad_norm": 0.026822632178664207,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.755108833312988,
+ "step": 265
+ },
+ {
+ "epoch": 3.6980802792321117,
+ "grad_norm": 0.02738630585372448,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.74041748046875,
+ "step": 266
+ },
+ {
+ "epoch": 3.712041884816754,
+ "grad_norm": 0.02361280657351017,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.764027118682861,
+ "step": 267
+ },
+ {
+ "epoch": 3.7260034904013963,
+ "grad_norm": 0.029926009476184845,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.780390739440918,
+ "step": 268
+ },
+ {
+ "epoch": 3.7399650959860384,
+ "grad_norm": 0.03519507870078087,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.758105278015137,
+ "step": 269
+ },
+ {
+ "epoch": 3.7539267015706805,
+ "grad_norm": 0.030114633962512016,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.757230758666992,
+ "step": 270
+ },
+ {
+ "epoch": 3.7678883071553226,
+ "grad_norm": 0.02467748522758484,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.7615461349487305,
+ "step": 271
+ },
+ {
+ "epoch": 3.781849912739965,
+ "grad_norm": 0.03425183147192001,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.749448776245117,
+ "step": 272
+ },
+ {
+ "epoch": 3.7958115183246073,
+ "grad_norm": 0.046307582408189774,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.7356977462768555,
+ "step": 273
+ },
+ {
+ "epoch": 3.80977312390925,
+ "grad_norm": 0.054102856665849686,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.741869926452637,
+ "step": 274
+ },
+ {
+ "epoch": 3.823734729493892,
+ "grad_norm": 0.07026689499616623,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.721103668212891,
+ "step": 275
+ },
+ {
+ "epoch": 3.837696335078534,
+ "grad_norm": 0.0733039602637291,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.731414794921875,
+ "step": 276
+ },
+ {
+ "epoch": 3.851657940663176,
+ "grad_norm": 0.05342920497059822,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.734922885894775,
+ "step": 277
+ },
+ {
+ "epoch": 3.8656195462478182,
+ "grad_norm": 0.0504680834710598,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.732746124267578,
+ "step": 278
+ },
+ {
+ "epoch": 3.8795811518324608,
+ "grad_norm": 0.038991592824459076,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.719473838806152,
+ "step": 279
+ },
+ {
+ "epoch": 3.893542757417103,
+ "grad_norm": 0.04434404894709587,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.709481239318848,
+ "step": 280
+ },
+ {
+ "epoch": 3.9075043630017454,
+ "grad_norm": 0.047285184264183044,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.693531036376953,
+ "step": 281
+ },
+ {
+ "epoch": 3.9214659685863875,
+ "grad_norm": 0.045081138610839844,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.728959083557129,
+ "step": 282
+ },
+ {
+ "epoch": 3.9354275741710296,
+ "grad_norm": 0.025133663788437843,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.719696998596191,
+ "step": 283
+ },
+ {
+ "epoch": 3.9493891797556717,
+ "grad_norm": 0.03353596478700638,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.700235366821289,
+ "step": 284
+ },
+ {
+ "epoch": 3.9633507853403143,
+ "grad_norm": 0.035162411630153656,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.684836387634277,
+ "step": 285
+ },
+ {
+ "epoch": 3.9773123909249564,
+ "grad_norm": 0.04025450348854065,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.692097187042236,
+ "step": 286
+ },
+ {
+ "epoch": 3.9912739965095985,
+ "grad_norm": 0.04995397478342056,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.6766862869262695,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.031619708985090256,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 3.5519747734069824,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.714687705039978,
+ "eval_runtime": 59.637,
+ "eval_samples_per_second": 40.948,
+ "eval_steps_per_second": 0.654,
+ "step": 288
+ },
+ {
+ "epoch": 4.013961605584642,
+ "grad_norm": 0.036328598856925964,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.666294097900391,
+ "step": 289
+ },
+ {
+ "epoch": 4.027923211169284,
+ "grad_norm": 0.07326336950063705,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.6879167556762695,
+ "step": 290
+ },
+ {
+ "epoch": 4.041884816753926,
+ "grad_norm": 0.07947526127099991,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.674883842468262,
+ "step": 291
+ },
+ {
+ "epoch": 4.055846422338569,
+ "grad_norm": 0.04599885269999504,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.688086986541748,
+ "step": 292
+ },
+ {
+ "epoch": 4.069808027923211,
+ "grad_norm": 0.0465150885283947,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.649483680725098,
+ "step": 293
+ },
+ {
+ "epoch": 4.0837696335078535,
+ "grad_norm": 0.038542453199625015,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.653055667877197,
+ "step": 294
+ },
+ {
+ "epoch": 4.097731239092496,
+ "grad_norm": 0.03659164905548096,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.65955924987793,
+ "step": 295
+ },
+ {
+ "epoch": 4.111692844677138,
+ "grad_norm": 0.03824038803577423,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.650853157043457,
+ "step": 296
+ },
+ {
+ "epoch": 4.12565445026178,
+ "grad_norm": 0.03591204807162285,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.643499374389648,
+ "step": 297
+ },
+ {
+ "epoch": 4.139616055846423,
+ "grad_norm": 0.03269682824611664,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.657600402832031,
+ "step": 298
+ },
+ {
+ "epoch": 4.153577661431065,
+ "grad_norm": 0.03343537449836731,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.6352081298828125,
+ "step": 299
+ },
+ {
+ "epoch": 4.167539267015707,
+ "grad_norm": 0.034283801913261414,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.626757621765137,
+ "step": 300
+ },
+ {
+ "epoch": 4.181500872600349,
+ "grad_norm": 0.034614723175764084,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.6316680908203125,
+ "step": 301
+ },
+ {
+ "epoch": 4.195462478184991,
+ "grad_norm": 0.0340128131210804,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.648160934448242,
+ "step": 302
+ },
+ {
+ "epoch": 4.209424083769633,
+ "grad_norm": 0.03068280965089798,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.619738578796387,
+ "step": 303
+ },
+ {
+ "epoch": 4.223385689354275,
+ "grad_norm": 0.030269058421254158,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.609238624572754,
+ "step": 304
+ },
+ {
+ "epoch": 4.237347294938918,
+ "grad_norm": 0.03612440824508667,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.602194786071777,
+ "step": 305
+ },
+ {
+ "epoch": 4.2513089005235605,
+ "grad_norm": 0.044292811304330826,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.61159610748291,
+ "step": 306
+ },
+ {
+ "epoch": 4.265270506108203,
+ "grad_norm": 0.04449291527271271,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.6274237632751465,
+ "step": 307
+ },
+ {
+ "epoch": 4.279232111692845,
+ "grad_norm": 0.05014493316411972,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.589034080505371,
+ "step": 308
+ },
+ {
+ "epoch": 4.293193717277487,
+ "grad_norm": 0.05286206305027008,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.600770950317383,
+ "step": 309
+ },
+ {
+ "epoch": 4.307155322862129,
+ "grad_norm": 0.048459846526384354,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.5796098709106445,
+ "step": 310
+ },
+ {
+ "epoch": 4.321116928446771,
+ "grad_norm": 0.05329663306474686,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.590407371520996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335078534031414,
+ "grad_norm": 0.052595119923353195,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.583693504333496,
+ "step": 312
+ },
+ {
+ "epoch": 4.349040139616056,
+ "grad_norm": 0.05164601653814316,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.603281021118164,
+ "step": 313
+ },
+ {
+ "epoch": 4.363001745200698,
+ "grad_norm": 0.04657137766480446,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.591344833374023,
+ "step": 314
+ },
+ {
+ "epoch": 4.37696335078534,
+ "grad_norm": 0.03804599866271019,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.580578804016113,
+ "step": 315
+ },
+ {
+ "epoch": 4.390924956369982,
+ "grad_norm": 0.03835320845246315,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.597836971282959,
+ "step": 316
+ },
+ {
+ "epoch": 4.4048865619546245,
+ "grad_norm": 0.034059133380651474,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.567502021789551,
+ "step": 317
+ },
+ {
+ "epoch": 4.418848167539267,
+ "grad_norm": 0.03350386768579483,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.552022933959961,
+ "step": 318
+ },
+ {
+ "epoch": 4.43280977312391,
+ "grad_norm": 0.032574303448200226,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.559177875518799,
+ "step": 319
+ },
+ {
+ "epoch": 4.446771378708552,
+ "grad_norm": 0.030893854796886444,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.549288272857666,
+ "step": 320
+ },
+ {
+ "epoch": 4.460732984293194,
+ "grad_norm": 0.03255036100745201,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.5400390625,
+ "step": 321
+ },
+ {
+ "epoch": 4.474694589877836,
+ "grad_norm": 0.03009258769452572,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.5328474044799805,
+ "step": 322
+ },
+ {
+ "epoch": 4.488656195462478,
+ "grad_norm": 0.03178320452570915,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.5594482421875,
+ "step": 323
+ },
+ {
+ "epoch": 4.50261780104712,
+ "grad_norm": 0.03360465541481972,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.5402727127075195,
+ "step": 324
+ },
+ {
+ "epoch": 4.516579406631763,
+ "grad_norm": 0.04275049269199371,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.534474849700928,
+ "step": 325
+ },
+ {
+ "epoch": 4.530541012216405,
+ "grad_norm": 0.04140641540288925,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.5324859619140625,
+ "step": 326
+ },
+ {
+ "epoch": 4.544502617801047,
+ "grad_norm": 0.04769204929471016,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.4919352531433105,
+ "step": 327
+ },
+ {
+ "epoch": 4.558464223385689,
+ "grad_norm": 0.04402285814285278,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.520858287811279,
+ "step": 328
+ },
+ {
+ "epoch": 4.5724258289703315,
+ "grad_norm": 0.038526229560375214,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.523778438568115,
+ "step": 329
+ },
+ {
+ "epoch": 4.5863874345549736,
+ "grad_norm": 0.05422377213835716,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.517973899841309,
+ "step": 330
+ },
+ {
+ "epoch": 4.600349040139616,
+ "grad_norm": 0.07236721366643906,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.548083782196045,
+ "step": 331
+ },
+ {
+ "epoch": 4.614310645724259,
+ "grad_norm": 0.06861061602830887,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.535577774047852,
+ "step": 332
+ },
+ {
+ "epoch": 4.628272251308901,
+ "grad_norm": 0.07148340344429016,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.54155158996582,
+ "step": 333
+ },
+ {
+ "epoch": 4.642233856893543,
+ "grad_norm": 0.07838302850723267,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.552852630615234,
+ "step": 334
+ },
+ {
+ "epoch": 4.656195462478185,
+ "grad_norm": 0.10946758091449738,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.555066108703613,
+ "step": 335
+ },
+ {
+ "epoch": 4.670157068062827,
+ "grad_norm": 0.10143530368804932,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.584932804107666,
+ "step": 336
+ },
+ {
+ "epoch": 4.684118673647469,
+ "grad_norm": 0.08926606178283691,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.570462226867676,
+ "step": 337
+ },
+ {
+ "epoch": 4.698080279232112,
+ "grad_norm": 0.07060826569795609,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.5716753005981445,
+ "step": 338
+ },
+ {
+ "epoch": 4.712041884816754,
+ "grad_norm": 0.0418456569314003,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.556496620178223,
+ "step": 339
+ },
+ {
+ "epoch": 4.726003490401396,
+ "grad_norm": 0.05952411890029907,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.567801475524902,
+ "step": 340
+ },
+ {
+ "epoch": 4.739965095986038,
+ "grad_norm": 0.042999885976314545,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.537403106689453,
+ "step": 341
+ },
+ {
+ "epoch": 4.7539267015706805,
+ "grad_norm": 0.04426449537277222,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.552387237548828,
+ "step": 342
+ },
+ {
+ "epoch": 4.767888307155323,
+ "grad_norm": 0.040835145860910416,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.53051233291626,
+ "step": 343
+ },
+ {
+ "epoch": 4.781849912739965,
+ "grad_norm": 0.03819291293621063,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.531174659729004,
+ "step": 344
+ },
+ {
+ "epoch": 4.795811518324607,
+ "grad_norm": 0.030088599771261215,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.499328136444092,
+ "step": 345
+ },
+ {
+ "epoch": 4.80977312390925,
+ "grad_norm": 0.029984362423419952,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.504540920257568,
+ "step": 346
+ },
+ {
+ "epoch": 4.823734729493892,
+ "grad_norm": 0.027402915060520172,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.50899076461792,
+ "step": 347
+ },
+ {
+ "epoch": 4.837696335078534,
+ "grad_norm": 0.025096213445067406,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.480479717254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.851657940663176,
+ "grad_norm": 0.026828132569789886,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.501729965209961,
+ "step": 349
+ },
+ {
+ "epoch": 4.865619546247818,
+ "grad_norm": 0.023715168237686157,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.466860771179199,
+ "step": 350
+ },
+ {
+ "epoch": 4.879581151832461,
+ "grad_norm": 0.024503640830516815,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.491827964782715,
+ "step": 351
+ },
+ {
+ "epoch": 4.893542757417103,
+ "grad_norm": 0.020891105756163597,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.478384971618652,
+ "step": 352
+ },
+ {
+ "epoch": 4.907504363001745,
+ "grad_norm": 0.021373765543103218,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.4612135887146,
+ "step": 353
+ },
+ {
+ "epoch": 4.9214659685863875,
+ "grad_norm": 0.021773340180516243,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.4547319412231445,
+ "step": 354
+ },
+ {
+ "epoch": 4.93542757417103,
+ "grad_norm": 0.019637586548924446,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.431741714477539,
+ "step": 355
+ },
+ {
+ "epoch": 4.949389179755672,
+ "grad_norm": 0.023108696565032005,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.448612213134766,
+ "step": 356
+ },
+ {
+ "epoch": 4.963350785340314,
+ "grad_norm": 0.021935103461146355,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.447112083435059,
+ "step": 357
+ },
+ {
+ "epoch": 4.977312390924956,
+ "grad_norm": 0.023260753601789474,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.436091423034668,
+ "step": 358
+ },
+ {
+ "epoch": 4.991273996509599,
+ "grad_norm": 0.021355057135224342,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.408421993255615,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.014992325566709042,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 3.4004223346710205,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6834986805915833,
+ "eval_runtime": 60.4148,
+ "eval_samples_per_second": 40.421,
+ "eval_steps_per_second": 0.646,
+ "step": 360
+ },
+ {
+ "epoch": 5.013961605584642,
+ "grad_norm": 0.02471218630671501,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.396238803863525,
+ "step": 361
+ },
+ {
+ "epoch": 5.027923211169284,
+ "grad_norm": 0.027672411873936653,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.4048919677734375,
+ "step": 362
+ },
+ {
+ "epoch": 5.041884816753926,
+ "grad_norm": 0.031560689210891724,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.399665832519531,
+ "step": 363
+ },
+ {
+ "epoch": 5.055846422338569,
+ "grad_norm": 0.03459061309695244,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.396360397338867,
+ "step": 364
+ },
+ {
+ "epoch": 5.069808027923211,
+ "grad_norm": 0.04040960222482681,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.405309677124023,
+ "step": 365
+ },
+ {
+ "epoch": 5.0837696335078535,
+ "grad_norm": 0.05444670096039772,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.408894062042236,
+ "step": 366
+ },
+ {
+ "epoch": 5.097731239092496,
+ "grad_norm": 0.05816745012998581,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.403566360473633,
+ "step": 367
+ },
+ {
+ "epoch": 5.111692844677138,
+ "grad_norm": 0.06330069154500961,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.395961761474609,
+ "step": 368
+ },
+ {
+ "epoch": 5.12565445026178,
+ "grad_norm": 0.07458917796611786,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.400968551635742,
+ "step": 369
+ },
+ {
+ "epoch": 5.139616055846423,
+ "grad_norm": 0.06231356039643288,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.412624359130859,
+ "step": 370
+ },
+ {
+ "epoch": 5.153577661431065,
+ "grad_norm": 0.05212346464395523,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.394794464111328,
+ "step": 371
+ },
+ {
+ "epoch": 5.167539267015707,
+ "grad_norm": 0.050003375858068466,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.380125045776367,
+ "step": 372
+ },
+ {
+ "epoch": 5.181500872600349,
+ "grad_norm": 0.048393815755844116,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.390806198120117,
+ "step": 373
+ },
+ {
+ "epoch": 5.195462478184991,
+ "grad_norm": 0.04200794920325279,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.387467384338379,
+ "step": 374
+ },
+ {
+ "epoch": 5.209424083769633,
+ "grad_norm": 0.04931448772549629,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.37552547454834,
+ "step": 375
+ },
+ {
+ "epoch": 5.223385689354275,
+ "grad_norm": 0.03851857781410217,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.373126983642578,
+ "step": 376
+ },
+ {
+ "epoch": 5.237347294938918,
+ "grad_norm": 0.03458717465400696,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.346826553344727,
+ "step": 377
+ },
+ {
+ "epoch": 5.2513089005235605,
+ "grad_norm": 0.031107677146792412,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.353341102600098,
+ "step": 378
+ },
+ {
+ "epoch": 5.265270506108203,
+ "grad_norm": 0.03077058307826519,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.358076095581055,
+ "step": 379
+ },
+ {
+ "epoch": 5.279232111692845,
+ "grad_norm": 0.028170401230454445,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.354019641876221,
+ "step": 380
+ },
+ {
+ "epoch": 5.293193717277487,
+ "grad_norm": 0.02871990203857422,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.368258476257324,
+ "step": 381
+ },
+ {
+ "epoch": 5.307155322862129,
+ "grad_norm": 0.03046584501862526,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.342505931854248,
+ "step": 382
+ },
+ {
+ "epoch": 5.321116928446771,
+ "grad_norm": 0.02781563811004162,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.339411735534668,
+ "step": 383
+ },
+ {
+ "epoch": 5.335078534031414,
+ "grad_norm": 0.028825288638472557,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.341948509216309,
+ "step": 384
+ },
+ {
+ "epoch": 5.349040139616056,
+ "grad_norm": 0.026684079319238663,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.340881824493408,
+ "step": 385
+ },
+ {
+ "epoch": 5.363001745200698,
+ "grad_norm": 0.02429167740046978,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.342733383178711,
+ "step": 386
+ },
+ {
+ "epoch": 5.37696335078534,
+ "grad_norm": 0.02855239436030388,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.313878536224365,
+ "step": 387
+ },
+ {
+ "epoch": 5.390924956369982,
+ "grad_norm": 0.029651548713445663,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.312665939331055,
+ "step": 388
+ },
+ {
+ "epoch": 5.4048865619546245,
+ "grad_norm": 0.027191132307052612,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3086838722229,
+ "step": 389
+ },
+ {
+ "epoch": 5.418848167539267,
+ "grad_norm": 0.026719525456428528,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.305359840393066,
+ "step": 390
+ },
+ {
+ "epoch": 5.43280977312391,
+ "grad_norm": 0.026763100177049637,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.3164381980896,
+ "step": 391
+ },
+ {
+ "epoch": 5.446771378708552,
+ "grad_norm": 0.03125109151005745,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.3083953857421875,
+ "step": 392
+ },
+ {
+ "epoch": 5.460732984293194,
+ "grad_norm": 0.0394827201962471,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.275996208190918,
+ "step": 393
+ },
+ {
+ "epoch": 5.474694589877836,
+ "grad_norm": 0.040579475462436676,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.292995929718018,
+ "step": 394
+ },
+ {
+ "epoch": 5.488656195462478,
+ "grad_norm": 0.037223003804683685,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.298526763916016,
+ "step": 395
+ },
+ {
+ "epoch": 5.50261780104712,
+ "grad_norm": 0.04207765311002731,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.287985801696777,
+ "step": 396
+ },
+ {
+ "epoch": 5.516579406631763,
+ "grad_norm": 0.048299964517354965,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.244608402252197,
+ "step": 397
+ },
+ {
+ "epoch": 5.530541012216405,
+ "grad_norm": 0.04650307446718216,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.285453796386719,
+ "step": 398
+ },
+ {
+ "epoch": 5.544502617801047,
+ "grad_norm": 0.045339711010456085,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.312187671661377,
+ "step": 399
+ },
+ {
+ "epoch": 5.558464223385689,
+ "grad_norm": 0.041067708283662796,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.292692184448242,
+ "step": 400
+ },
+ {
+ "epoch": 5.5724258289703315,
+ "grad_norm": 0.05455026403069496,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.253364562988281,
+ "step": 401
+ },
+ {
+ "epoch": 5.5863874345549736,
+ "grad_norm": 0.06660915911197662,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.262646675109863,
+ "step": 402
+ },
+ {
+ "epoch": 5.600349040139616,
+ "grad_norm": 0.06634793430566788,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.274322986602783,
+ "step": 403
+ },
+ {
+ "epoch": 5.614310645724259,
+ "grad_norm": 0.08996330946683884,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.287951469421387,
+ "step": 404
+ },
+ {
+ "epoch": 5.628272251308901,
+ "grad_norm": 0.08526470512151718,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.3007097244262695,
+ "step": 405
+ },
+ {
+ "epoch": 5.642233856893543,
+ "grad_norm": 0.08766836673021317,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.332538604736328,
+ "step": 406
+ },
+ {
+ "epoch": 5.656195462478185,
+ "grad_norm": 0.07055048644542694,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.296169281005859,
+ "step": 407
+ },
+ {
+ "epoch": 5.670157068062827,
+ "grad_norm": 0.07520575076341629,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.309683322906494,
+ "step": 408
+ },
+ {
+ "epoch": 5.684118673647469,
+ "grad_norm": 0.06115756183862686,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.311301231384277,
+ "step": 409
+ },
+ {
+ "epoch": 5.698080279232112,
+ "grad_norm": 0.05862593650817871,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.28066349029541,
+ "step": 410
+ },
+ {
+ "epoch": 5.712041884816754,
+ "grad_norm": 0.06169601157307625,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.286840438842773,
+ "step": 411
+ },
+ {
+ "epoch": 5.726003490401396,
+ "grad_norm": 0.05270719155669212,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.302594184875488,
+ "step": 412
+ },
+ {
+ "epoch": 5.739965095986038,
+ "grad_norm": 0.03912634402513504,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.262563705444336,
+ "step": 413
+ },
+ {
+ "epoch": 5.7539267015706805,
+ "grad_norm": 0.03215811401605606,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.269310474395752,
+ "step": 414
+ },
+ {
+ "epoch": 5.767888307155323,
+ "grad_norm": 0.03236516937613487,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.268766403198242,
+ "step": 415
+ },
+ {
+ "epoch": 5.781849912739965,
+ "grad_norm": 0.027429452165961266,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.283801555633545,
+ "step": 416
+ },
+ {
+ "epoch": 5.795811518324607,
+ "grad_norm": 0.02990538813173771,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.22291374206543,
+ "step": 417
+ },
+ {
+ "epoch": 5.80977312390925,
+ "grad_norm": 0.030410146340727806,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.2281599044799805,
+ "step": 418
+ },
+ {
+ "epoch": 5.823734729493892,
+ "grad_norm": 0.030068187043070793,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.250294208526611,
+ "step": 419
+ },
+ {
+ "epoch": 5.837696335078534,
+ "grad_norm": 0.02800513058900833,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.226508617401123,
+ "step": 420
+ },
+ {
+ "epoch": 5.851657940663176,
+ "grad_norm": 0.023823332041502,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.247838020324707,
+ "step": 421
+ },
+ {
+ "epoch": 5.865619546247818,
+ "grad_norm": 0.024974249303340912,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.2392473220825195,
+ "step": 422
+ },
+ {
+ "epoch": 5.879581151832461,
+ "grad_norm": 0.023320546373724937,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.209543704986572,
+ "step": 423
+ },
+ {
+ "epoch": 5.893542757417103,
+ "grad_norm": 0.024201491847634315,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.219725608825684,
+ "step": 424
+ },
+ {
+ "epoch": 5.907504363001745,
+ "grad_norm": 0.021649666130542755,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.203141212463379,
+ "step": 425
+ },
+ {
+ "epoch": 5.9214659685863875,
+ "grad_norm": 0.0248002577573061,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.179058074951172,
+ "step": 426
+ },
+ {
+ "epoch": 5.93542757417103,
+ "grad_norm": 0.028549587354063988,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.202126502990723,
+ "step": 427
+ },
+ {
+ "epoch": 5.949389179755672,
+ "grad_norm": 0.0331057533621788,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.208107948303223,
+ "step": 428
+ },
+ {
+ "epoch": 5.963350785340314,
+ "grad_norm": 0.031843364238739014,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.184683799743652,
+ "step": 429
+ },
+ {
+ "epoch": 5.977312390924956,
+ "grad_norm": 0.024927377700805664,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.180486679077148,
+ "step": 430
+ },
+ {
+ "epoch": 5.991273996509599,
+ "grad_norm": 0.026685267686843872,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.186746597290039,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.018545351922512054,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 3.217043876647949,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.653471827507019,
+ "eval_runtime": 61.1586,
+ "eval_samples_per_second": 39.929,
+ "eval_steps_per_second": 0.638,
+ "step": 432
+ },
+ {
+ "epoch": 6.013961605584642,
+ "grad_norm": 0.025291118770837784,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.156534194946289,
+ "step": 433
+ },
+ {
+ "epoch": 6.027923211169284,
+ "grad_norm": 0.03774752840399742,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.144041538238525,
+ "step": 434
+ },
+ {
+ "epoch": 6.041884816753926,
+ "grad_norm": 0.04244692996144295,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.158862113952637,
+ "step": 435
+ },
+ {
+ "epoch": 6.055846422338569,
+ "grad_norm": 0.039782389998435974,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.166165351867676,
+ "step": 436
+ },
+ {
+ "epoch": 6.069808027923211,
+ "grad_norm": 0.05489896610379219,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.155766487121582,
+ "step": 437
+ },
+ {
+ "epoch": 6.0837696335078535,
+ "grad_norm": 0.059844326227903366,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.153733253479004,
+ "step": 438
+ },
+ {
+ "epoch": 6.097731239092496,
+ "grad_norm": 0.05539938807487488,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.130112648010254,
+ "step": 439
+ },
+ {
+ "epoch": 6.111692844677138,
+ "grad_norm": 0.05283183231949806,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.136910915374756,
+ "step": 440
+ },
+ {
+ "epoch": 6.12565445026178,
+ "grad_norm": 0.054407261312007904,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.165216445922852,
+ "step": 441
+ },
+ {
+ "epoch": 6.139616055846423,
+ "grad_norm": 0.0509788915514946,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.140888214111328,
+ "step": 442
+ },
+ {
+ "epoch": 6.153577661431065,
+ "grad_norm": 0.05018748342990875,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.13115930557251,
+ "step": 443
+ },
+ {
+ "epoch": 6.167539267015707,
+ "grad_norm": 0.047012437134981155,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.146507263183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181500872600349,
+ "grad_norm": 0.04943335801362991,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.135608673095703,
+ "step": 445
+ },
+ {
+ "epoch": 6.195462478184991,
+ "grad_norm": 0.05304260924458504,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.136394023895264,
+ "step": 446
+ },
+ {
+ "epoch": 6.209424083769633,
+ "grad_norm": 0.04554047808051109,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.124661922454834,
+ "step": 447
+ },
+ {
+ "epoch": 6.223385689354275,
+ "grad_norm": 0.040672432631254196,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.129773139953613,
+ "step": 448
+ },
+ {
+ "epoch": 6.237347294938918,
+ "grad_norm": 0.03442879393696785,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.128028392791748,
+ "step": 449
+ },
+ {
+ "epoch": 6.2513089005235605,
+ "grad_norm": 0.03927934914827347,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.092485427856445,
+ "step": 450
+ },
+ {
+ "epoch": 6.265270506108203,
+ "grad_norm": 0.029380977153778076,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.077669143676758,
+ "step": 451
+ },
+ {
+ "epoch": 6.279232111692845,
+ "grad_norm": 0.03331821784377098,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.138857841491699,
+ "step": 452
+ },
+ {
+ "epoch": 6.293193717277487,
+ "grad_norm": 0.02939118817448616,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.082937240600586,
+ "step": 453
+ },
+ {
+ "epoch": 6.307155322862129,
+ "grad_norm": 0.02730412967503071,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.093183517456055,
+ "step": 454
+ },
+ {
+ "epoch": 6.321116928446771,
+ "grad_norm": 0.02964780665934086,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.1028151512146,
+ "step": 455
+ },
+ {
+ "epoch": 6.335078534031414,
+ "grad_norm": 0.026350129395723343,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.066362380981445,
+ "step": 456
+ },
+ {
+ "epoch": 6.349040139616056,
+ "grad_norm": 0.027010295540094376,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.070037841796875,
+ "step": 457
+ },
+ {
+ "epoch": 6.363001745200698,
+ "grad_norm": 0.02906128577888012,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.107708930969238,
+ "step": 458
+ },
+ {
+ "epoch": 6.37696335078534,
+ "grad_norm": 0.028619443997740746,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.073182582855225,
+ "step": 459
+ },
+ {
+ "epoch": 6.390924956369982,
+ "grad_norm": 0.029981834813952446,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.076602935791016,
+ "step": 460
+ },
+ {
+ "epoch": 6.4048865619546245,
+ "grad_norm": 0.03255423903465271,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.063301086425781,
+ "step": 461
+ },
+ {
+ "epoch": 6.418848167539267,
+ "grad_norm": 0.03863866627216339,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.041812896728516,
+ "step": 462
+ },
+ {
+ "epoch": 6.43280977312391,
+ "grad_norm": 0.03860306739807129,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.06298828125,
+ "step": 463
+ },
+ {
+ "epoch": 6.446771378708552,
+ "grad_norm": 0.03378435596823692,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.0434370040893555,
+ "step": 464
+ },
+ {
+ "epoch": 6.460732984293194,
+ "grad_norm": 0.03276759013533592,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.036187171936035,
+ "step": 465
+ },
+ {
+ "epoch": 6.474694589877836,
+ "grad_norm": 0.04130513221025467,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.039098739624023,
+ "step": 466
+ },
+ {
+ "epoch": 6.488656195462478,
+ "grad_norm": 0.05389940366148949,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.062798976898193,
+ "step": 467
+ },
+ {
+ "epoch": 6.50261780104712,
+ "grad_norm": 0.05881704390048981,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.067028522491455,
+ "step": 468
+ },
+ {
+ "epoch": 6.516579406631763,
+ "grad_norm": 0.061897799372673035,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.056947231292725,
+ "step": 469
+ },
+ {
+ "epoch": 6.530541012216405,
+ "grad_norm": 0.06549323350191116,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.089262008666992,
+ "step": 470
+ },
+ {
+ "epoch": 6.544502617801047,
+ "grad_norm": 0.048082590103149414,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.055979251861572,
+ "step": 471
+ },
+ {
+ "epoch": 6.558464223385689,
+ "grad_norm": 0.03768027946352959,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.0660600662231445,
+ "step": 472
+ },
+ {
+ "epoch": 6.5724258289703315,
+ "grad_norm": 0.043922603130340576,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 5.036745071411133,
+ "step": 473
+ },
+ {
+ "epoch": 6.5863874345549736,
+ "grad_norm": 0.044848836958408356,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.073650360107422,
+ "step": 474
+ },
+ {
+ "epoch": 6.600349040139616,
+ "grad_norm": 0.05841026455163956,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.053045272827148,
+ "step": 475
+ },
+ {
+ "epoch": 6.614310645724259,
+ "grad_norm": 0.0650017261505127,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 5.052206039428711,
+ "step": 476
+ },
+ {
+ "epoch": 6.628272251308901,
+ "grad_norm": 0.048979002982378006,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 5.054277420043945,
+ "step": 477
+ },
+ {
+ "epoch": 6.642233856893543,
+ "grad_norm": 0.05085799843072891,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 5.063844680786133,
+ "step": 478
+ },
+ {
+ "epoch": 6.656195462478185,
+ "grad_norm": 0.04336017370223999,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 5.045474052429199,
+ "step": 479
+ },
+ {
+ "epoch": 6.670157068062827,
+ "grad_norm": 0.04113380238413811,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 5.031338214874268,
+ "step": 480
+ },
+ {
+ "epoch": 6.684118673647469,
+ "grad_norm": 0.03199538215994835,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 5.030025959014893,
+ "step": 481
+ },
+ {
+ "epoch": 6.698080279232112,
+ "grad_norm": 0.03043113648891449,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 5.033163070678711,
+ "step": 482
+ },
+ {
+ "epoch": 6.712041884816754,
+ "grad_norm": 0.031694378703832626,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 5.004739761352539,
+ "step": 483
+ },
+ {
+ "epoch": 6.726003490401396,
+ "grad_norm": 0.030243348330259323,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 5.012566089630127,
+ "step": 484
+ },
+ {
+ "epoch": 6.739965095986038,
+ "grad_norm": 0.02937416359782219,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 5.039803504943848,
+ "step": 485
+ },
+ {
+ "epoch": 6.7539267015706805,
+ "grad_norm": 0.029616443440318108,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 5.026117324829102,
+ "step": 486
+ },
+ {
+ "epoch": 6.767888307155323,
+ "grad_norm": 0.029567131772637367,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 5.006569862365723,
+ "step": 487
+ },
+ {
+ "epoch": 6.781849912739965,
+ "grad_norm": 0.029443370178341866,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.9941487312316895,
+ "step": 488
+ },
+ {
+ "epoch": 6.795811518324607,
+ "grad_norm": 0.03434352949261665,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.976431846618652,
+ "step": 489
+ },
+ {
+ "epoch": 6.80977312390925,
+ "grad_norm": 0.03845234215259552,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.993168354034424,
+ "step": 490
+ },
+ {
+ "epoch": 6.823734729493892,
+ "grad_norm": 0.05058208107948303,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.030162334442139,
+ "step": 491
+ },
+ {
+ "epoch": 6.837696335078534,
+ "grad_norm": 0.05961710214614868,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 5.007053375244141,
+ "step": 492
+ },
+ {
+ "epoch": 6.851657940663176,
+ "grad_norm": 0.056259628385305405,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 5.020356178283691,
+ "step": 493
+ },
+ {
+ "epoch": 6.865619546247818,
+ "grad_norm": 0.048628658056259155,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 5.037652969360352,
+ "step": 494
+ },
+ {
+ "epoch": 6.879581151832461,
+ "grad_norm": 0.040836453437805176,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.990390777587891,
+ "step": 495
+ },
+ {
+ "epoch": 6.893542757417103,
+ "grad_norm": 0.04112066701054573,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 5.017822742462158,
+ "step": 496
+ },
+ {
+ "epoch": 6.907504363001745,
+ "grad_norm": 0.03956829756498337,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.993971824645996,
+ "step": 497
+ },
+ {
+ "epoch": 6.9214659685863875,
+ "grad_norm": 0.03887917473912239,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.982889175415039,
+ "step": 498
+ },
+ {
+ "epoch": 6.93542757417103,
+ "grad_norm": 0.031012587249279022,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.961717128753662,
+ "step": 499
+ },
+ {
+ "epoch": 6.949389179755672,
+ "grad_norm": 0.03443489596247673,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.9904046058654785,
+ "step": 500
+ },
+ {
+ "epoch": 6.963350785340314,
+ "grad_norm": 0.035175010561943054,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.983619213104248,
+ "step": 501
+ },
+ {
+ "epoch": 6.977312390924956,
+ "grad_norm": 0.031724050641059875,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 5.000522613525391,
+ "step": 502
+ },
+ {
+ "epoch": 6.991273996509599,
+ "grad_norm": 0.032037995755672455,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.947064399719238,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.01883271522819996,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 3.112166166305542,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6296830177307129,
+ "eval_runtime": 59.5495,
+ "eval_samples_per_second": 41.008,
+ "eval_steps_per_second": 0.655,
+ "step": 504
+ },
+ {
+ "epoch": 7.013961605584642,
+ "grad_norm": 0.031091563403606415,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.919820785522461,
+ "step": 505
+ },
+ {
+ "epoch": 7.027923211169284,
+ "grad_norm": 0.04074421525001526,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.945765972137451,
+ "step": 506
+ },
+ {
+ "epoch": 7.041884816753926,
+ "grad_norm": 0.04225052148103714,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.935208797454834,
+ "step": 507
+ },
+ {
+ "epoch": 7.055846422338569,
+ "grad_norm": 0.0399051196873188,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.951314926147461,
+ "step": 508
+ },
+ {
+ "epoch": 7.069808027923211,
+ "grad_norm": 0.04785890504717827,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.92843770980835,
+ "step": 509
+ },
+ {
+ "epoch": 7.0837696335078535,
+ "grad_norm": 0.04922301694750786,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.901020050048828,
+ "step": 510
+ },
+ {
+ "epoch": 7.097731239092496,
+ "grad_norm": 0.05307295173406601,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.961151599884033,
+ "step": 511
+ },
+ {
+ "epoch": 7.111692844677138,
+ "grad_norm": 0.07003362476825714,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.9427876472473145,
+ "step": 512
+ },
+ {
+ "epoch": 7.12565445026178,
+ "grad_norm": 0.06498776376247406,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.947838306427002,
+ "step": 513
+ },
+ {
+ "epoch": 7.139616055846423,
+ "grad_norm": 0.058723628520965576,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.938859939575195,
+ "step": 514
+ },
+ {
+ "epoch": 7.153577661431065,
+ "grad_norm": 0.06590746343135834,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.962221145629883,
+ "step": 515
+ },
+ {
+ "epoch": 7.167539267015707,
+ "grad_norm": 0.052714601159095764,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.931011199951172,
+ "step": 516
+ },
+ {
+ "epoch": 7.181500872600349,
+ "grad_norm": 0.04296804964542389,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.93586540222168,
+ "step": 517
+ },
+ {
+ "epoch": 7.195462478184991,
+ "grad_norm": 0.03699173778295517,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.910763263702393,
+ "step": 518
+ },
+ {
+ "epoch": 7.209424083769633,
+ "grad_norm": 0.0317913293838501,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.9256439208984375,
+ "step": 519
+ },
+ {
+ "epoch": 7.223385689354275,
+ "grad_norm": 0.030313925817608833,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.938329696655273,
+ "step": 520
+ },
+ {
+ "epoch": 7.237347294938918,
+ "grad_norm": 0.02879229187965393,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.924966812133789,
+ "step": 521
+ },
+ {
+ "epoch": 7.2513089005235605,
+ "grad_norm": 0.02689451351761818,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.9079742431640625,
+ "step": 522
+ },
+ {
+ "epoch": 7.265270506108203,
+ "grad_norm": 0.028403930366039276,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.913238525390625,
+ "step": 523
+ },
+ {
+ "epoch": 7.279232111692845,
+ "grad_norm": 0.025912854820489883,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.896944046020508,
+ "step": 524
+ },
+ {
+ "epoch": 7.293193717277487,
+ "grad_norm": 0.024899760261178017,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.884920120239258,
+ "step": 525
+ },
+ {
+ "epoch": 7.307155322862129,
+ "grad_norm": 0.026912573724985123,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.9181060791015625,
+ "step": 526
+ },
+ {
+ "epoch": 7.321116928446771,
+ "grad_norm": 0.029539721086621284,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.914867401123047,
+ "step": 527
+ },
+ {
+ "epoch": 7.335078534031414,
+ "grad_norm": 0.028527872636914253,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.907217979431152,
+ "step": 528
+ },
+ {
+ "epoch": 7.349040139616056,
+ "grad_norm": 0.025879016146063805,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.905783176422119,
+ "step": 529
+ },
+ {
+ "epoch": 7.363001745200698,
+ "grad_norm": 0.026395199820399284,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.880505561828613,
+ "step": 530
+ },
+ {
+ "epoch": 7.37696335078534,
+ "grad_norm": 0.025072351098060608,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.870449066162109,
+ "step": 531
+ },
+ {
+ "epoch": 7.390924956369982,
+ "grad_norm": 0.02751019410789013,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.888952255249023,
+ "step": 532
+ },
+ {
+ "epoch": 7.4048865619546245,
+ "grad_norm": 0.025810658931732178,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.894552230834961,
+ "step": 533
+ },
+ {
+ "epoch": 7.418848167539267,
+ "grad_norm": 0.02399870567023754,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.8768229484558105,
+ "step": 534
+ },
+ {
+ "epoch": 7.43280977312391,
+ "grad_norm": 0.024639368057250977,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.9030914306640625,
+ "step": 535
+ },
+ {
+ "epoch": 7.446771378708552,
+ "grad_norm": 0.02543291077017784,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.86467170715332,
+ "step": 536
+ },
+ {
+ "epoch": 7.460732984293194,
+ "grad_norm": 0.02702544629573822,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.888283729553223,
+ "step": 537
+ },
+ {
+ "epoch": 7.474694589877836,
+ "grad_norm": 0.028904743492603302,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.866081237792969,
+ "step": 538
+ },
+ {
+ "epoch": 7.488656195462478,
+ "grad_norm": 0.03055848367512226,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.890739440917969,
+ "step": 539
+ },
+ {
+ "epoch": 7.50261780104712,
+ "grad_norm": 0.03410084918141365,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.862918853759766,
+ "step": 540
+ },
+ {
+ "epoch": 7.516579406631763,
+ "grad_norm": 0.03738495334982872,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.864728927612305,
+ "step": 541
+ },
+ {
+ "epoch": 7.530541012216405,
+ "grad_norm": 0.037646882236003876,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.865185737609863,
+ "step": 542
+ },
+ {
+ "epoch": 7.544502617801047,
+ "grad_norm": 0.03418618440628052,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.866114139556885,
+ "step": 543
+ },
+ {
+ "epoch": 7.558464223385689,
+ "grad_norm": 0.036924004554748535,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.883186340332031,
+ "step": 544
+ },
+ {
+ "epoch": 7.5724258289703315,
+ "grad_norm": 0.03590598329901695,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.863725662231445,
+ "step": 545
+ },
+ {
+ "epoch": 7.5863874345549736,
+ "grad_norm": 0.037363454699516296,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.841030120849609,
+ "step": 546
+ },
+ {
+ "epoch": 7.600349040139616,
+ "grad_norm": 0.046537864953279495,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.873837471008301,
+ "step": 547
+ },
+ {
+ "epoch": 7.614310645724259,
+ "grad_norm": 0.057226769626140594,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.865679740905762,
+ "step": 548
+ },
+ {
+ "epoch": 7.628272251308901,
+ "grad_norm": 0.06175961345434189,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.888645648956299,
+ "step": 549
+ },
+ {
+ "epoch": 7.642233856893543,
+ "grad_norm": 0.05063694715499878,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.877828598022461,
+ "step": 550
+ },
+ {
+ "epoch": 7.656195462478185,
+ "grad_norm": 0.03950628638267517,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.88637638092041,
+ "step": 551
+ },
+ {
+ "epoch": 7.670157068062827,
+ "grad_norm": 0.04183337092399597,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.8597917556762695,
+ "step": 552
+ },
+ {
+ "epoch": 7.684118673647469,
+ "grad_norm": 0.04642999917268753,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.880380153656006,
+ "step": 553
+ },
+ {
+ "epoch": 7.698080279232112,
+ "grad_norm": 0.05333106964826584,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.853932857513428,
+ "step": 554
+ },
+ {
+ "epoch": 7.712041884816754,
+ "grad_norm": 0.04764975979924202,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.870235443115234,
+ "step": 555
+ },
+ {
+ "epoch": 7.726003490401396,
+ "grad_norm": 0.043072231113910675,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.822697639465332,
+ "step": 556
+ },
+ {
+ "epoch": 7.739965095986038,
+ "grad_norm": 0.03913857042789459,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.849332809448242,
+ "step": 557
+ },
+ {
+ "epoch": 7.7539267015706805,
+ "grad_norm": 0.034836407750844955,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.842280387878418,
+ "step": 558
+ },
+ {
+ "epoch": 7.767888307155323,
+ "grad_norm": 0.033344294875860214,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.837811470031738,
+ "step": 559
+ },
+ {
+ "epoch": 7.781849912739965,
+ "grad_norm": 0.03342701122164726,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.876040458679199,
+ "step": 560
+ },
+ {
+ "epoch": 7.795811518324607,
+ "grad_norm": 0.034638747572898865,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.830340385437012,
+ "step": 561
+ },
+ {
+ "epoch": 7.80977312390925,
+ "grad_norm": 0.03358135372400284,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.828607559204102,
+ "step": 562
+ },
+ {
+ "epoch": 7.823734729493892,
+ "grad_norm": 0.036001041531562805,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.827905654907227,
+ "step": 563
+ },
+ {
+ "epoch": 7.837696335078534,
+ "grad_norm": 0.038079630583524704,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.866955280303955,
+ "step": 564
+ },
+ {
+ "epoch": 7.851657940663176,
+ "grad_norm": 0.03728470206260681,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.828039169311523,
+ "step": 565
+ },
+ {
+ "epoch": 7.865619546247818,
+ "grad_norm": 0.03920375928282738,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.841631889343262,
+ "step": 566
+ },
+ {
+ "epoch": 7.879581151832461,
+ "grad_norm": 0.032353926450014114,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.861110210418701,
+ "step": 567
+ },
+ {
+ "epoch": 7.893542757417103,
+ "grad_norm": 0.03241422027349472,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.818441390991211,
+ "step": 568
+ },
+ {
+ "epoch": 7.907504363001745,
+ "grad_norm": 0.03879080340266228,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.823749542236328,
+ "step": 569
+ },
+ {
+ "epoch": 7.9214659685863875,
+ "grad_norm": 0.03486498072743416,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.830697059631348,
+ "step": 570
+ },
+ {
+ "epoch": 7.93542757417103,
+ "grad_norm": 0.03279713913798332,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.815736770629883,
+ "step": 571
+ },
+ {
+ "epoch": 7.949389179755672,
+ "grad_norm": 0.03313925117254257,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.8364033699035645,
+ "step": 572
+ },
+ {
+ "epoch": 7.963350785340314,
+ "grad_norm": 0.03405062481760979,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.816010475158691,
+ "step": 573
+ },
+ {
+ "epoch": 7.977312390924956,
+ "grad_norm": 0.03259236365556717,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.813413619995117,
+ "step": 574
+ },
+ {
+ "epoch": 7.991273996509599,
+ "grad_norm": 0.030408985912799835,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.833840370178223,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.022680889815092087,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 2.99565052986145,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6126086115837097,
+ "eval_runtime": 60.2662,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 576
+ },
+ {
+ "epoch": 8.013961605584642,
+ "grad_norm": 0.029732126742601395,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.7816314697265625,
+ "step": 577
+ },
+ {
+ "epoch": 8.027923211169284,
+ "grad_norm": 0.04357453063130379,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.7575225830078125,
+ "step": 578
+ },
+ {
+ "epoch": 8.041884816753926,
+ "grad_norm": 0.051903221756219864,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.802519798278809,
+ "step": 579
+ },
+ {
+ "epoch": 8.055846422338568,
+ "grad_norm": 0.0539076030254364,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.799004554748535,
+ "step": 580
+ },
+ {
+ "epoch": 8.06980802792321,
+ "grad_norm": 0.043642807751894,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.7896528244018555,
+ "step": 581
+ },
+ {
+ "epoch": 8.083769633507853,
+ "grad_norm": 0.0469554178416729,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.765430927276611,
+ "step": 582
+ },
+ {
+ "epoch": 8.097731239092496,
+ "grad_norm": 0.042060062289237976,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.784132957458496,
+ "step": 583
+ },
+ {
+ "epoch": 8.111692844677139,
+ "grad_norm": 0.035101912915706635,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.767234802246094,
+ "step": 584
+ },
+ {
+ "epoch": 8.12565445026178,
+ "grad_norm": 0.042576879262924194,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.762052536010742,
+ "step": 585
+ },
+ {
+ "epoch": 8.139616055846423,
+ "grad_norm": 0.04231419041752815,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.752150535583496,
+ "step": 586
+ },
+ {
+ "epoch": 8.153577661431065,
+ "grad_norm": 0.046912599354982376,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.771881103515625,
+ "step": 587
+ },
+ {
+ "epoch": 8.167539267015707,
+ "grad_norm": 0.051271144300699234,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.773284912109375,
+ "step": 588
+ },
+ {
+ "epoch": 8.181500872600349,
+ "grad_norm": 0.04655958339571953,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.769917964935303,
+ "step": 589
+ },
+ {
+ "epoch": 8.195462478184991,
+ "grad_norm": 0.04463687539100647,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.77938175201416,
+ "step": 590
+ },
+ {
+ "epoch": 8.209424083769633,
+ "grad_norm": 0.047959886491298676,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.785962104797363,
+ "step": 591
+ },
+ {
+ "epoch": 8.223385689354275,
+ "grad_norm": 0.04414986073970795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.786281585693359,
+ "step": 592
+ },
+ {
+ "epoch": 8.237347294938917,
+ "grad_norm": 0.044150348752737045,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.777070045471191,
+ "step": 593
+ },
+ {
+ "epoch": 8.25130890052356,
+ "grad_norm": 0.037002332508563995,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.77547550201416,
+ "step": 594
+ },
+ {
+ "epoch": 8.265270506108202,
+ "grad_norm": 0.03417367860674858,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.782101154327393,
+ "step": 595
+ },
+ {
+ "epoch": 8.279232111692846,
+ "grad_norm": 0.03069210797548294,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.745595455169678,
+ "step": 596
+ },
+ {
+ "epoch": 8.293193717277488,
+ "grad_norm": 0.03228718787431717,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.7663655281066895,
+ "step": 597
+ },
+ {
+ "epoch": 8.30715532286213,
+ "grad_norm": 0.028603695333003998,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.767274856567383,
+ "step": 598
+ },
+ {
+ "epoch": 8.321116928446772,
+ "grad_norm": 0.03098377026617527,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.7853684425354,
+ "step": 599
+ },
+ {
+ "epoch": 8.335078534031414,
+ "grad_norm": 0.029242439195513725,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.738029956817627,
+ "step": 600
+ },
+ {
+ "epoch": 8.349040139616056,
+ "grad_norm": 0.02751081995666027,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.76801872253418,
+ "step": 601
+ },
+ {
+ "epoch": 8.363001745200698,
+ "grad_norm": 0.02838125079870224,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.743884086608887,
+ "step": 602
+ },
+ {
+ "epoch": 8.37696335078534,
+ "grad_norm": 0.029203062877058983,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.746668815612793,
+ "step": 603
+ },
+ {
+ "epoch": 8.390924956369982,
+ "grad_norm": 0.02679944597184658,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.760764122009277,
+ "step": 604
+ },
+ {
+ "epoch": 8.404886561954624,
+ "grad_norm": 0.028955848887562752,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.7353668212890625,
+ "step": 605
+ },
+ {
+ "epoch": 8.418848167539267,
+ "grad_norm": 0.02880851738154888,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.766837120056152,
+ "step": 606
+ },
+ {
+ "epoch": 8.432809773123909,
+ "grad_norm": 0.02495497092604637,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.78959321975708,
+ "step": 607
+ },
+ {
+ "epoch": 8.44677137870855,
+ "grad_norm": 0.026064753532409668,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.736433029174805,
+ "step": 608
+ },
+ {
+ "epoch": 8.460732984293193,
+ "grad_norm": 0.027566388249397278,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.782382965087891,
+ "step": 609
+ },
+ {
+ "epoch": 8.474694589877837,
+ "grad_norm": 0.02967359498143196,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.735286712646484,
+ "step": 610
+ },
+ {
+ "epoch": 8.488656195462479,
+ "grad_norm": 0.03079676441848278,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.725196838378906,
+ "step": 611
+ },
+ {
+ "epoch": 8.502617801047121,
+ "grad_norm": 0.032531846314668655,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.719333648681641,
+ "step": 612
+ },
+ {
+ "epoch": 8.516579406631763,
+ "grad_norm": 0.03120352514088154,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.723845958709717,
+ "step": 613
+ },
+ {
+ "epoch": 8.530541012216405,
+ "grad_norm": 0.03328424692153931,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.7669782638549805,
+ "step": 614
+ },
+ {
+ "epoch": 8.544502617801047,
+ "grad_norm": 0.035601649433374405,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.714925765991211,
+ "step": 615
+ },
+ {
+ "epoch": 8.55846422338569,
+ "grad_norm": 0.03675796464085579,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.715317726135254,
+ "step": 616
+ },
+ {
+ "epoch": 8.572425828970331,
+ "grad_norm": 0.03701988235116005,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.724606990814209,
+ "step": 617
+ },
+ {
+ "epoch": 8.586387434554974,
+ "grad_norm": 0.036719705909490585,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.746044635772705,
+ "step": 618
+ },
+ {
+ "epoch": 8.600349040139616,
+ "grad_norm": 0.034756604582071304,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.74885368347168,
+ "step": 619
+ },
+ {
+ "epoch": 8.614310645724258,
+ "grad_norm": 0.03237922117114067,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.721858501434326,
+ "step": 620
+ },
+ {
+ "epoch": 8.6282722513089,
+ "grad_norm": 0.028420142829418182,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.733814239501953,
+ "step": 621
+ },
+ {
+ "epoch": 8.642233856893542,
+ "grad_norm": 0.026203038170933723,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.726088523864746,
+ "step": 622
+ },
+ {
+ "epoch": 8.656195462478184,
+ "grad_norm": 0.02627231739461422,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.727700233459473,
+ "step": 623
+ },
+ {
+ "epoch": 8.670157068062828,
+ "grad_norm": 0.02920261025428772,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.731454849243164,
+ "step": 624
+ },
+ {
+ "epoch": 8.68411867364747,
+ "grad_norm": 0.02792907878756523,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.734154224395752,
+ "step": 625
+ },
+ {
+ "epoch": 8.698080279232112,
+ "grad_norm": 0.026770325377583504,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.715337753295898,
+ "step": 626
+ },
+ {
+ "epoch": 8.712041884816754,
+ "grad_norm": 0.027346540242433548,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.716512680053711,
+ "step": 627
+ },
+ {
+ "epoch": 8.726003490401396,
+ "grad_norm": 0.031965773552656174,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.7072038650512695,
+ "step": 628
+ },
+ {
+ "epoch": 8.739965095986038,
+ "grad_norm": 0.03708025813102722,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.749785900115967,
+ "step": 629
+ },
+ {
+ "epoch": 8.75392670157068,
+ "grad_norm": 0.04846490919589996,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.735586166381836,
+ "step": 630
+ },
+ {
+ "epoch": 8.767888307155323,
+ "grad_norm": 0.053352758288383484,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.720524787902832,
+ "step": 631
+ },
+ {
+ "epoch": 8.781849912739965,
+ "grad_norm": 0.040559202432632446,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.719901084899902,
+ "step": 632
+ },
+ {
+ "epoch": 8.795811518324607,
+ "grad_norm": 0.03333032503724098,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.708139419555664,
+ "step": 633
+ },
+ {
+ "epoch": 8.809773123909249,
+ "grad_norm": 0.030532941222190857,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.728184223175049,
+ "step": 634
+ },
+ {
+ "epoch": 8.823734729493891,
+ "grad_norm": 0.03337649628520012,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.726020336151123,
+ "step": 635
+ },
+ {
+ "epoch": 8.837696335078533,
+ "grad_norm": 0.03352045267820358,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.7196855545043945,
+ "step": 636
+ },
+ {
+ "epoch": 8.851657940663177,
+ "grad_norm": 0.03662872314453125,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.732823371887207,
+ "step": 637
+ },
+ {
+ "epoch": 8.86561954624782,
+ "grad_norm": 0.037085745483636856,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.728443145751953,
+ "step": 638
+ },
+ {
+ "epoch": 8.879581151832461,
+ "grad_norm": 0.036635689437389374,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.713988780975342,
+ "step": 639
+ },
+ {
+ "epoch": 8.893542757417103,
+ "grad_norm": 0.03719659522175789,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.715950965881348,
+ "step": 640
+ },
+ {
+ "epoch": 8.907504363001745,
+ "grad_norm": 0.03719159588217735,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.7124924659729,
+ "step": 641
+ },
+ {
+ "epoch": 8.921465968586388,
+ "grad_norm": 0.04275721684098244,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.708404064178467,
+ "step": 642
+ },
+ {
+ "epoch": 8.93542757417103,
+ "grad_norm": 0.04204598814249039,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.715372085571289,
+ "step": 643
+ },
+ {
+ "epoch": 8.949389179755672,
+ "grad_norm": 0.03938259929418564,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.729745864868164,
+ "step": 644
+ },
+ {
+ "epoch": 8.963350785340314,
+ "grad_norm": 0.036226850003004074,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.733749866485596,
+ "step": 645
+ },
+ {
+ "epoch": 8.977312390924956,
+ "grad_norm": 0.03542874753475189,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.706825256347656,
+ "step": 646
+ },
+ {
+ "epoch": 8.991273996509598,
+ "grad_norm": 0.029055381193757057,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.718567848205566,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.01990380510687828,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 2.943183183670044,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.601699709892273,
+ "eval_runtime": 60.0221,
+ "eval_samples_per_second": 40.685,
+ "eval_steps_per_second": 0.65,
+ "step": 648
+ },
+ {
+ "epoch": 9.013961605584642,
+ "grad_norm": 0.030904127284884453,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.660327911376953,
+ "step": 649
+ },
+ {
+ "epoch": 9.027923211169284,
+ "grad_norm": 0.0441313236951828,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.701565742492676,
+ "step": 650
+ },
+ {
+ "epoch": 9.041884816753926,
+ "grad_norm": 0.047547727823257446,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.666440486907959,
+ "step": 651
+ },
+ {
+ "epoch": 9.055846422338568,
+ "grad_norm": 0.04549463838338852,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.660548210144043,
+ "step": 652
+ },
+ {
+ "epoch": 9.06980802792321,
+ "grad_norm": 0.050346389412879944,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.640229225158691,
+ "step": 653
+ },
+ {
+ "epoch": 9.083769633507853,
+ "grad_norm": 0.058505453169345856,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.673881530761719,
+ "step": 654
+ },
+ {
+ "epoch": 9.097731239092496,
+ "grad_norm": 0.051989227533340454,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.693943977355957,
+ "step": 655
+ },
+ {
+ "epoch": 9.111692844677139,
+ "grad_norm": 0.052564945071935654,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.644315242767334,
+ "step": 656
+ },
+ {
+ "epoch": 9.12565445026178,
+ "grad_norm": 0.05647841840982437,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.681768417358398,
+ "step": 657
+ },
+ {
+ "epoch": 9.139616055846423,
+ "grad_norm": 0.058640867471694946,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.676457405090332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153577661431065,
+ "grad_norm": 0.05386025458574295,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.659588813781738,
+ "step": 659
+ },
+ {
+ "epoch": 9.167539267015707,
+ "grad_norm": 0.048699021339416504,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.680031776428223,
+ "step": 660
+ },
+ {
+ "epoch": 9.181500872600349,
+ "grad_norm": 0.04374176263809204,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.672671318054199,
+ "step": 661
+ },
+ {
+ "epoch": 9.195462478184991,
+ "grad_norm": 0.042337436228990555,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.673431873321533,
+ "step": 662
+ },
+ {
+ "epoch": 9.209424083769633,
+ "grad_norm": 0.04564661160111427,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.632595062255859,
+ "step": 663
+ },
+ {
+ "epoch": 9.223385689354275,
+ "grad_norm": 0.04404835030436516,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.678891181945801,
+ "step": 664
+ },
+ {
+ "epoch": 9.237347294938917,
+ "grad_norm": 0.040642790496349335,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.633418560028076,
+ "step": 665
+ },
+ {
+ "epoch": 9.25130890052356,
+ "grad_norm": 0.036492060869932175,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.670109748840332,
+ "step": 666
+ },
+ {
+ "epoch": 9.265270506108202,
+ "grad_norm": 0.03735879808664322,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.608570098876953,
+ "step": 667
+ },
+ {
+ "epoch": 9.279232111692846,
+ "grad_norm": 0.034563641995191574,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.632505416870117,
+ "step": 668
+ },
+ {
+ "epoch": 9.293193717277488,
+ "grad_norm": 0.03522484749555588,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.6475629806518555,
+ "step": 669
+ },
+ {
+ "epoch": 9.30715532286213,
+ "grad_norm": 0.03216472268104553,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.627873420715332,
+ "step": 670
+ },
+ {
+ "epoch": 9.321116928446772,
+ "grad_norm": 0.029155759140849113,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.652706146240234,
+ "step": 671
+ },
+ {
+ "epoch": 9.335078534031414,
+ "grad_norm": 0.031060541048645973,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.644545078277588,
+ "step": 672
+ },
+ {
+ "epoch": 9.349040139616056,
+ "grad_norm": 0.034979186952114105,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.642645359039307,
+ "step": 673
+ },
+ {
+ "epoch": 9.363001745200698,
+ "grad_norm": 0.035258810967206955,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.647383689880371,
+ "step": 674
+ },
+ {
+ "epoch": 9.37696335078534,
+ "grad_norm": 0.031552575528621674,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.667779445648193,
+ "step": 675
+ },
+ {
+ "epoch": 9.390924956369982,
+ "grad_norm": 0.03193951025605202,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.63625431060791,
+ "step": 676
+ },
+ {
+ "epoch": 9.404886561954624,
+ "grad_norm": 0.029811711981892586,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.658000946044922,
+ "step": 677
+ },
+ {
+ "epoch": 9.418848167539267,
+ "grad_norm": 0.028464138507843018,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.6545515060424805,
+ "step": 678
+ },
+ {
+ "epoch": 9.432809773123909,
+ "grad_norm": 0.0315735749900341,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.641153335571289,
+ "step": 679
+ },
+ {
+ "epoch": 9.44677137870855,
+ "grad_norm": 0.033412106335163116,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.626931667327881,
+ "step": 680
+ },
+ {
+ "epoch": 9.460732984293193,
+ "grad_norm": 0.03757626190781593,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.671254634857178,
+ "step": 681
+ },
+ {
+ "epoch": 9.474694589877837,
+ "grad_norm": 0.03910793736577034,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.651954174041748,
+ "step": 682
+ },
+ {
+ "epoch": 9.488656195462479,
+ "grad_norm": 0.04015549272298813,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.613324165344238,
+ "step": 683
+ },
+ {
+ "epoch": 9.502617801047121,
+ "grad_norm": 0.040689144283533096,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.639243125915527,
+ "step": 684
+ },
+ {
+ "epoch": 9.516579406631763,
+ "grad_norm": 0.04075361415743828,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.649250030517578,
+ "step": 685
+ },
+ {
+ "epoch": 9.530541012216405,
+ "grad_norm": 0.03914607688784599,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.66444730758667,
+ "step": 686
+ },
+ {
+ "epoch": 9.544502617801047,
+ "grad_norm": 0.033253733068704605,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.647335052490234,
+ "step": 687
+ },
+ {
+ "epoch": 9.55846422338569,
+ "grad_norm": 0.03179008886218071,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.6727800369262695,
+ "step": 688
+ },
+ {
+ "epoch": 9.572425828970331,
+ "grad_norm": 0.03477025404572487,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.655778884887695,
+ "step": 689
+ },
+ {
+ "epoch": 9.586387434554974,
+ "grad_norm": 0.030094800516963005,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.680851936340332,
+ "step": 690
+ },
+ {
+ "epoch": 9.600349040139616,
+ "grad_norm": 0.030153103172779083,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.654638290405273,
+ "step": 691
+ },
+ {
+ "epoch": 9.614310645724258,
+ "grad_norm": 0.029009278863668442,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.625916481018066,
+ "step": 692
+ },
+ {
+ "epoch": 9.6282722513089,
+ "grad_norm": 0.027248440310359,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.656370162963867,
+ "step": 693
+ },
+ {
+ "epoch": 9.642233856893542,
+ "grad_norm": 0.027727462351322174,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.631473541259766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656195462478184,
+ "grad_norm": 0.026710307225584984,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.636524200439453,
+ "step": 695
+ },
+ {
+ "epoch": 9.670157068062828,
+ "grad_norm": 0.02726181596517563,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.629701614379883,
+ "step": 696
+ },
+ {
+ "epoch": 9.68411867364747,
+ "grad_norm": 0.027146652340888977,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.640240669250488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698080279232112,
+ "grad_norm": 0.027165887877345085,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.64462947845459,
+ "step": 698
+ },
+ {
+ "epoch": 9.712041884816754,
+ "grad_norm": 0.03025982715189457,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.6271820068359375,
+ "step": 699
+ },
+ {
+ "epoch": 9.726003490401396,
+ "grad_norm": 0.02837025746703148,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.629697799682617,
+ "step": 700
+ },
+ {
+ "epoch": 9.739965095986038,
+ "grad_norm": 0.026777995750308037,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.646146774291992,
+ "step": 701
+ },
+ {
+ "epoch": 9.75392670157068,
+ "grad_norm": 0.024626802653074265,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.641671657562256,
+ "step": 702
+ },
+ {
+ "epoch": 9.767888307155323,
+ "grad_norm": 0.025329895317554474,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.626345157623291,
+ "step": 703
+ },
+ {
+ "epoch": 9.781849912739965,
+ "grad_norm": 0.027804844081401825,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.621457099914551,
+ "step": 704
+ },
+ {
+ "epoch": 9.795811518324607,
+ "grad_norm": 0.028763437643647194,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.6124114990234375,
+ "step": 705
+ },
+ {
+ "epoch": 9.809773123909249,
+ "grad_norm": 0.030769916251301765,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.604399681091309,
+ "step": 706
+ },
+ {
+ "epoch": 9.823734729493891,
+ "grad_norm": 0.03060949593782425,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.640804290771484,
+ "step": 707
+ },
+ {
+ "epoch": 9.837696335078533,
+ "grad_norm": 0.029832778498530388,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.593376159667969,
+ "step": 708
+ },
+ {
+ "epoch": 9.851657940663177,
+ "grad_norm": 0.02952776849269867,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.619533538818359,
+ "step": 709
+ },
+ {
+ "epoch": 9.86561954624782,
+ "grad_norm": 0.02894212119281292,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.640516757965088,
+ "step": 710
+ },
+ {
+ "epoch": 9.879581151832461,
+ "grad_norm": 0.026435624808073044,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.6342997550964355,
+ "step": 711
+ },
+ {
+ "epoch": 9.893542757417103,
+ "grad_norm": 0.02591666206717491,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.630884170532227,
+ "step": 712
+ },
+ {
+ "epoch": 9.907504363001745,
+ "grad_norm": 0.02702309936285019,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.632085800170898,
+ "step": 713
+ },
+ {
+ "epoch": 9.921465968586388,
+ "grad_norm": 0.026766657829284668,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.6231231689453125,
+ "step": 714
+ },
+ {
+ "epoch": 9.93542757417103,
+ "grad_norm": 0.026666559278964996,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.601619243621826,
+ "step": 715
+ },
+ {
+ "epoch": 9.949389179755672,
+ "grad_norm": 0.026801064610481262,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.641724109649658,
+ "step": 716
+ },
+ {
+ "epoch": 9.963350785340314,
+ "grad_norm": 0.03152798116207123,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.642511367797852,
+ "step": 717
+ },
+ {
+ "epoch": 9.977312390924956,
+ "grad_norm": 0.035406943410634995,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.635779857635498,
+ "step": 718
+ },
+ {
+ "epoch": 9.991273996509598,
+ "grad_norm": 0.037504684180021286,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.629463195800781,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02263174019753933,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 2.876767635345459,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.5942904949188232,
+ "eval_runtime": 60.2086,
+ "eval_samples_per_second": 40.559,
+ "eval_steps_per_second": 0.648,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03707852587103844,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.5765533447265625,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.05454680323600769,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.576363563537598,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.04832114651799202,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.586915969848633,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.050469402223825455,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.5265092849731445,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0622565895318985,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.592833518981934,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0640949010848999,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.5855560302734375,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.06502196192741394,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.591236114501953,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.06167149916291237,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.56747579574585,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.06747374683618546,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.60452938079834,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.07210271060466766,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.561588764190674,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.0733448714017868,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.610766410827637,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.0731528028845787,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.589570045471191,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.0822535902261734,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.631077289581299,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.06921929866075516,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.628293991088867,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.07443586736917496,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.627321243286133,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.07153813540935516,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.641068935394287,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.069196417927742,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.615875720977783,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06483697891235352,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.6165313720703125,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.05482150986790657,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.599890232086182,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.05364019051194191,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.61989164352417,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.053216829895973206,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.564305305480957,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05740506947040558,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.582683563232422,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.05587657541036606,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.61264705657959,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.050252825021743774,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.648050308227539,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.050955042243003845,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.615367889404297,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04903746768832207,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.617378234863281,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04252410680055618,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.609130859375,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.03801053389906883,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.625149250030518,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.03422372043132782,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.606106281280518,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.030507879331707954,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.6125993728637695,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.0340375155210495,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.5690202713012695,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.030507374554872513,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.6042985916137695,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027383938431739807,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.585570812225342,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.028792303055524826,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.566786766052246,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02866457775235176,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.566376686096191,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.028657952323555946,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.581750869750977,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.028227943927049637,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.576788902282715,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.024757765233516693,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.59373664855957,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.027442635968327522,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.570978164672852,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.02590034157037735,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.571286201477051,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.025696570053696632,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.542209625244141,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.025478003546595573,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.5900421142578125,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.02563696913421154,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.544609546661377,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.025326019152998924,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.5808515548706055,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.024502627551555634,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.560579776763916,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.02687518671154976,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.567154884338379,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.025280235335230827,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.560011863708496,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.026266315951943398,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.5758442878723145,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.026058891788125038,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.583309173583984,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.02620471641421318,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.547737121582031,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02626362442970276,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.556955337524414,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.024503158405423164,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.580735206604004,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02580532431602478,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.564000129699707,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.024437161162495613,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.545594215393066,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.026664750650525093,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.56415319442749,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.02499818615615368,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.550689697265625,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.027514580637216568,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.5548248291015625,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.026799403131008148,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.548295021057129,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.02685476280748844,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.5660600662231445,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.026104001328349113,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.556082248687744,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.027479661628603935,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.5580244064331055,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.02638247236609459,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.550073623657227,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.025853842496871948,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.578088760375977,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.02636682614684105,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.58807373046875,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.02684398740530014,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.561017036437988,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.028538964688777924,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.548762321472168,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.028327809646725655,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.548969745635986,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.025131452828645706,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.564047813415527,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.026498543098568916,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.553997039794922,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.030081963166594505,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.5787787437438965,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.03201213851571083,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.566424369812012,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.020518779754638672,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8421616554260254,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.5899484157562256,
+ "eval_runtime": 60.2671,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03692612797021866,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.492715835571289,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.050258126109838486,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.480323791503906,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.04179152473807335,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.484371185302734,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.04160204157233238,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.4905524253845215,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.03974860534071922,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.503803253173828,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.037031810730695724,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.494629859924316,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.038762833923101425,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.495111465454102,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.0412563718855381,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.490113735198975,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.041371624916791916,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.490643501281738,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04151557758450508,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.49881649017334,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04037577286362648,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.492178916931152,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.03676290065050125,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.477460861206055,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.038055796176195145,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.509150981903076,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.041152238845825195,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.47133731842041,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.03728092834353447,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.508072376251221,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.035338856279850006,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.505619049072266,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.037813492119312286,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.509489059448242,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.03789111226797104,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.508646488189697,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.03880828619003296,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.52567720413208,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.03611127287149429,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.498889446258545,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.036903638392686844,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.498601913452148,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03620576485991478,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.490840911865234,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03555903211236,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.506166458129883,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.03441924974322319,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.498001575469971,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.0334339402616024,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.491082191467285,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.031418606638908386,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.527408599853516,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03073921985924244,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.514107704162598,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.029589885845780373,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.48323917388916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.02974501997232437,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.51338529586792,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.029247552156448364,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.504205226898193,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.03132032975554466,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.505674362182617,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03088037669658661,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.477198600769043,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.031054604798555374,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.498721122741699,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03293661028146744,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.527826309204102,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.03181406110525131,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.495999336242676,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.029380083084106445,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.530351638793945,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.029296347871422768,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.492367744445801,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.02958036959171295,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.493037700653076,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03078395500779152,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.47504186630249,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03280661255121231,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.518409729003906,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.03323954716324806,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.460753917694092,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.032254140824079514,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.474285125732422,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.03018898330628872,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.5162353515625,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.02802245132625103,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.538684844970703,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.02821732498705387,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.4850287437438965,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.029189350083470345,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.50770378112793,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.02851196378469467,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.499988555908203,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.02741030789911747,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.499852180480957,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.028708625584840775,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.465211868286133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.028519680723547935,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.502464294433594,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.027846841141581535,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.503248691558838,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030456233769655228,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.490391731262207,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029952052980661392,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.50056266784668,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.02925778552889824,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.49215030670166,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.029679695144295692,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.523841857910156,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.031178541481494904,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.499871730804443,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.031264498829841614,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.50546932220459,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03482109308242798,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.494480133056641,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.03341105580329895,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.532899856567383,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028623664751648903,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.5016279220581055,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028600146993994713,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.511589050292969,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.029127901419997215,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.502058982849121,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.027898237109184265,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.493182182312012,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.028683282434940338,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.519793510437012,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.02887023612856865,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.457289695739746,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.02811296097934246,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.49398946762085,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.028485111892223358,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.478344440460205,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02914245054125786,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.487666130065918,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.028030497953295708,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.518488883972168,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.027973609045147896,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.515271186828613,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.027458522468805313,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.490856647491455,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.01909293793141842,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.822120189666748,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5866912603378296,
+ "eval_runtime": 61.3185,
+ "eval_samples_per_second": 39.825,
+ "eval_steps_per_second": 0.636,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03683992475271225,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.438363075256348,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.05042923241853714,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.443560600280762,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.055914171040058136,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.422282695770264,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.057007886469364166,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.442644119262695,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.05999710038304329,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.44780969619751,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06472347676753998,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.485074996948242,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.06923902779817581,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.457515716552734,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.0710320770740509,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.465476989746094,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.06942574679851532,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.426226615905762,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.06867868453264236,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.438092231750488,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.0688796266913414,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.446664810180664,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.0690719485282898,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.456482887268066,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.06068912148475647,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.4730730056762695,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.05411486327648163,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.4596452713012695,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.056623928248882294,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.418197154998779,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.05634097009897232,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.482892036437988,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.05056057870388031,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.457589149475098,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.04460986703634262,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.451439380645752,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.045022401958703995,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.452855110168457,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.04054102674126625,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.417323112487793,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.03682882338762283,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.465713977813721,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.03528819605708122,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.4502339363098145,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.03568802773952484,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.422399520874023,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.0363088883459568,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.447678089141846,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.03460463136434555,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.437664985656738,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.034751344472169876,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.489151954650879,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.033658601343631744,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.4471635818481445,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.03433239087462425,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.455332279205322,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.032074104994535446,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.435157775878906,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.032168325036764145,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.435118675231934,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.03007678873836994,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.42781925201416,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03094375506043434,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.433105945587158,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.032186444848775864,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.383000373840332,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.030610304325819016,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.473257064819336,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.030257435515522957,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.453041076660156,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.0301323514431715,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.396578788757324,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.030934512615203857,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.445223808288574,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.029907869175076485,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.442287445068359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03163841366767883,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.434650421142578,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.029627414420247078,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.454374313354492,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.031026875600218773,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.413784027099609,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03099491260945797,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.458438873291016,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.02981731668114662,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.438560962677002,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.02927570603787899,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.44493293762207,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.029779717326164246,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.445121765136719,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.02909068390727043,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.4736127853393555,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.029217200353741646,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.434604644775391,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.029872911050915718,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.454591751098633,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.028742222115397453,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.468222141265869,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.030313383787870407,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.45759391784668,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.03014662116765976,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.446331977844238,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029481127858161926,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.478209495544434,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.028485266491770744,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.444445610046387,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.02857700176537037,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.430604934692383,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.029038602486252785,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.43759298324585,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.029576119035482407,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.474579811096191,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.02897879108786583,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.440483093261719,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.029799485579133034,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.4839277267456055,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.030710609629750252,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.466696739196777,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.030270954594016075,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.455197334289551,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.03170182928442955,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.454664707183838,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030684730038046837,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.437619209289551,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.030972573906183243,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.431796073913574,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028914963826537132,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.460618019104004,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.029810264706611633,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.407312393188477,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.029806939885020256,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.466520309448242,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.03020886518061161,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.427349090576172,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030598826706409454,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.427891731262207,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03192012757062912,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.448113918304443,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.030492989346385002,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.450060844421387,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03188254311680794,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.43007230758667,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02081947773694992,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.782536506652832,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5860040187835693,
+ "eval_runtime": 60.6725,
+ "eval_samples_per_second": 40.249,
+ "eval_steps_per_second": 0.643,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.036721356213092804,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.358441352844238,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.056633755564689636,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.361527919769287,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.06175022944808006,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.3827924728393555,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05159056559205055,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.354856491088867,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.04885871335864067,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.359966278076172,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.04407693073153496,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.364612579345703,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.04891050234436989,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.344021320343018,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.04643657058477402,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.366066932678223,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.05241481214761734,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.3390607833862305,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.05837438255548477,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.390261650085449,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.058482322841882706,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.3856201171875,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.06507553160190582,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.414223670959473,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.0575365275144577,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.402576446533203,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05055592209100723,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.389563083648682,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.047167956829071045,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.402750015258789,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.04473405331373215,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.358315467834473,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.042440690100193024,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.392551898956299,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04465719312429428,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.373558044433594,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.04337528720498085,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.37944221496582,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.04322915896773338,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.38493013381958,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.039873406291007996,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.389793395996094,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.0413178876042366,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.427964210510254,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.04567280039191246,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.404172897338867,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04154926910996437,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.391359329223633,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.03774876892566681,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.381457328796387,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03588850423693657,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.392526626586914,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.03757261857390404,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.39273738861084,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03849692642688751,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.409083366394043,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03787294402718544,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.3776044845581055,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.034496866166591644,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.413462162017822,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.03973205015063286,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.359318733215332,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.036192040890455246,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.392844200134277,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.03436632081866264,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.421384811401367,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.035658642649650574,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.386757850646973,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.03434896469116211,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.364978313446045,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03642290085554123,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.385326385498047,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.03427736461162567,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.406352996826172,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.034845270216464996,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.370953559875488,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03390975296497345,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.3799943923950195,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.03134961053729057,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.419156551361084,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.03512260690331459,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.404660701751709,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03319067507982254,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.388535022735596,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03113507106900215,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.403522968292236,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03559132665395737,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.382323265075684,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.03579750657081604,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.37620210647583,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.033675212413072586,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.411443710327148,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.03220067173242569,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.443470001220703,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031610481441020966,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.406038284301758,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.02958764135837555,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.41604471206665,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.0318237729370594,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4253387451171875,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03260074555873871,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.370792388916016,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.03359333053231239,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.397639274597168,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032516930252313614,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.4140777587890625,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.03203396871685982,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.408872604370117,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.03201935440301895,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.402975082397461,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.03330211713910103,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.394018173217773,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.0348181314766407,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.380587100982666,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.03174913674592972,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.372378349304199,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.029864462092518806,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.42018985748291,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.033312153071165085,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.385973930358887,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03203144669532776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.395167827606201,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.031578030437231064,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.3982062339782715,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.030885562300682068,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.400731086730957,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03235667943954468,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.389830589294434,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.032664328813552856,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.401277542114258,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.030354196205735207,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.422520637512207,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.031162403523921967,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.393032073974609,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.032187070697546005,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.464164733886719,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.03445067256689072,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.368991851806641,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03398612141609192,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.418752670288086,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.029730670154094696,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.41429328918457,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022440090775489807,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7315595149993896,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5853753089904785,
+ "eval_runtime": 60.7174,
+ "eval_samples_per_second": 40.219,
+ "eval_steps_per_second": 0.642,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.03754177689552307,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.304028511047363,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.051616180688142776,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.34659481048584,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055258627980947495,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.310445785522461,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.049489911645650864,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3416619300842285,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05377533286809921,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.3143310546875,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.05443088337779045,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.320600986480713,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.050222691148519516,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.334789752960205,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.051904670894145966,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.316756248474121,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.05376426503062248,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.274116516113281,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.053313180804252625,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.321177005767822,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.058234598487615585,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.3169264793396,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05730101093649864,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.307013511657715,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.058140840381383896,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.332961082458496,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05654815956950188,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.315462112426758,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.05129384249448776,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.338476181030273,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.047652702778577805,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.321232795715332,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.050912003964185715,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.318971633911133,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.0514022521674633,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.3334150314331055,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.05457981303334236,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.325802803039551,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04678885266184807,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.307281494140625,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.05048161372542381,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.299831390380859,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.05334261804819107,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.313231468200684,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.05055771768093109,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.36859130859375,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04955561086535454,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.301530361175537,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.04548008739948273,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.339517593383789,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.046585116535425186,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.333036422729492,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.0503716766834259,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.364710807800293,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.040569406002759933,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.369089126586914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.03754054754972458,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.3349432945251465,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03775954991579056,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.333134174346924,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.04066266492009163,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.354358673095703,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.040490977466106415,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.360981464385986,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.0426657609641552,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.357726097106934,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.0415605828166008,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.363457679748535,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.04020509123802185,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.372447967529297,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.04101516678929329,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.292452812194824,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.03546427935361862,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.33892297744751,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.03661105036735535,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.329778671264648,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03565476834774017,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.351012229919434,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.03591311350464821,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.376980781555176,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.03665630891919136,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.354767799377441,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.0371236577630043,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.374916076660156,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.03464849665760994,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.3713059425354,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03345152735710144,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.350670337677002,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.03357568010687828,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.3573713302612305,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03503486514091492,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.362038612365723,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.03455005958676338,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.3263773918151855,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.03173859417438507,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.351345062255859,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03301836550235748,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.398364067077637,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.033733855932950974,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.335268020629883,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034080155193805695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.371492385864258,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.034396376460790634,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.382076740264893,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.03303401917219162,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.36623477935791,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03297369182109833,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3148393630981445,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.034819211810827255,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.358110427856445,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.033389780670404434,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.367427825927734,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.032363444566726685,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.3813676834106445,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03509577363729477,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.386520862579346,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.034685730934143066,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.33738374710083,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.034755002707242966,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.360980987548828,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03461688384413719,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.396650314331055,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03392560034990311,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.356130123138428,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.03342147544026375,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.359145164489746,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.032852694392204285,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.3796162605285645,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03472859039902687,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.361110687255859,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03277098760008812,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.361811637878418,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.03341882303357124,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.369524955749512,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.032347049564123154,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.369096755981445,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03161193057894707,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.339844703674316,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.03322301432490349,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.376675605773926,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03419092297554016,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.346864223480225,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02461826801300049,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7236275672912598,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.5859372019767761,
+ "eval_runtime": 59.0139,
+ "eval_samples_per_second": 41.38,
+ "eval_steps_per_second": 0.661,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013961605584642,
+ "grad_norm": 0.039559684693813324,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 4.289841651916504,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027923211169284,
+ "grad_norm": 0.05554218962788582,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 4.255925178527832,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041884816753926,
+ "grad_norm": 0.05330996215343475,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 4.272125244140625,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055846422338568,
+ "grad_norm": 0.053247954696416855,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 4.290765285491943,
+ "step": 1084
+ },
+ {
+ "epoch": 15.06980802792321,
+ "grad_norm": 0.05471150949597359,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 4.271744728088379,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083769633507853,
+ "grad_norm": 0.05976719409227371,
+ "learning_rate": 0.000491285979372924,
+ "loss": 4.28203821182251,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097731239092496,
+ "grad_norm": 0.05369880795478821,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 4.275313377380371,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111692844677139,
+ "grad_norm": 0.052529267966747284,
+ "learning_rate": 0.000490873338088198,
+ "loss": 4.25678825378418,
+ "step": 1088
+ },
+ {
+ "epoch": 15.12565445026178,
+ "grad_norm": 0.05184522643685341,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 4.260510444641113,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139616055846423,
+ "grad_norm": 0.051606856286525726,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 4.283233642578125,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153577661431065,
+ "grad_norm": 0.05318903923034668,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 4.286595344543457,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167539267015707,
+ "grad_norm": 0.058525051921606064,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 4.282374382019043,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181500872600349,
+ "grad_norm": 0.054776228964328766,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 4.2913126945495605,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195462478184991,
+ "grad_norm": 0.04907633364200592,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 4.297187805175781,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209424083769633,
+ "grad_norm": 0.05163136497139931,
+ "learning_rate": 0.000489424334303338,
+ "loss": 4.2968645095825195,
+ "step": 1095
+ },
+ {
+ "epoch": 15.223385689354275,
+ "grad_norm": 0.053157489746809006,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 4.285839080810547,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237347294938917,
+ "grad_norm": 0.05251535773277283,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 4.311246871948242,
+ "step": 1097
+ },
+ {
+ "epoch": 15.25130890052356,
+ "grad_norm": 0.048785097897052765,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 4.300364971160889,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265270506108202,
+ "grad_norm": 0.050393108278512955,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 4.299890995025635,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279232111692846,
+ "grad_norm": 0.04894037917256355,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 4.312449932098389,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293193717277488,
+ "grad_norm": 0.05274952948093414,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 4.2889814376831055,
+ "step": 1101
+ },
+ {
+ "epoch": 15.30715532286213,
+ "grad_norm": 0.05946570262312889,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 4.2731170654296875,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321116928446772,
+ "grad_norm": 0.054245397448539734,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 4.272463798522949,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335078534031414,
+ "grad_norm": 0.042092181742191315,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 4.299953937530518,
+ "step": 1104
+ },
+ {
+ "epoch": 15.349040139616056,
+ "grad_norm": 0.040222618728876114,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 4.292524814605713,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363001745200698,
+ "grad_norm": 0.04298696666955948,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 4.31342887878418,
+ "step": 1106
+ },
+ {
+ "epoch": 15.37696335078534,
+ "grad_norm": 0.04271050915122032,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 4.277663707733154,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390924956369982,
+ "grad_norm": 0.042364370077848434,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 4.311102390289307,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404886561954624,
+ "grad_norm": 0.04343319311738014,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 4.311610221862793,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418848167539267,
+ "grad_norm": 0.041321512311697006,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 4.29677677154541,
+ "step": 1110
+ },
+ {
+ "epoch": 15.432809773123909,
+ "grad_norm": 0.04107942432165146,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 4.3163018226623535,
+ "step": 1111
+ },
+ {
+ "epoch": 15.44677137870855,
+ "grad_norm": 0.03844515234231949,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 4.311301231384277,
+ "step": 1112
+ },
+ {
+ "epoch": 15.460732984293193,
+ "grad_norm": 0.03731007128953934,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 4.31325626373291,
+ "step": 1113
+ },
+ {
+ "epoch": 15.474694589877837,
+ "grad_norm": 0.03807826340198517,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 4.296201705932617,
+ "step": 1114
+ },
+ {
+ "epoch": 15.488656195462479,
+ "grad_norm": 0.040137600153684616,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 4.320395469665527,
+ "step": 1115
+ },
+ {
+ "epoch": 15.502617801047121,
+ "grad_norm": 0.038484394550323486,
+ "learning_rate": 0.000485033394934934,
+ "loss": 4.313858985900879,
+ "step": 1116
+ },
+ {
+ "epoch": 15.516579406631763,
+ "grad_norm": 0.03603338822722435,
+ "learning_rate": 0.000484822676912974,
+ "loss": 4.297666549682617,
+ "step": 1117
+ },
+ {
+ "epoch": 15.530541012216405,
+ "grad_norm": 0.038647182285785675,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 4.305002689361572,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544502617801047,
+ "grad_norm": 0.03907139599323273,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 4.329583644866943,
+ "step": 1119
+ },
+ {
+ "epoch": 15.55846422338569,
+ "grad_norm": 0.03621622920036316,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 4.290896415710449,
+ "step": 1120
+ },
+ {
+ "epoch": 15.572425828970331,
+ "grad_norm": 0.037496887147426605,
+ "learning_rate": 0.000483978341612154,
+ "loss": 4.33336877822876,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586387434554974,
+ "grad_norm": 0.03778311610221863,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 4.285354137420654,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600349040139616,
+ "grad_norm": 0.03782997280359268,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 4.29697847366333,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614310645724258,
+ "grad_norm": 0.03665657714009285,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 4.320103168487549,
+ "step": 1124
+ },
+ {
+ "epoch": 15.6282722513089,
+ "grad_norm": 0.03796127438545227,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 4.317153453826904,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642233856893542,
+ "grad_norm": 0.036425214260816574,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 4.293642520904541,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656195462478184,
+ "grad_norm": 0.03629784658551216,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 4.291100025177002,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670157068062828,
+ "grad_norm": 0.03674762696027756,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 4.295144557952881,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68411867364747,
+ "grad_norm": 0.036291178315877914,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 4.313133239746094,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698080279232112,
+ "grad_norm": 0.03817860782146454,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 4.294533729553223,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712041884816754,
+ "grad_norm": 0.036819685250520706,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 4.299263954162598,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726003490401396,
+ "grad_norm": 0.035103823989629745,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 4.287391185760498,
+ "step": 1132
+ },
+ {
+ "epoch": 15.739965095986038,
+ "grad_norm": 0.035701096057891846,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 4.318216323852539,
+ "step": 1133
+ },
+ {
+ "epoch": 15.75392670157068,
+ "grad_norm": 0.03639856353402138,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 4.301465034484863,
+ "step": 1134
+ },
+ {
+ "epoch": 15.767888307155323,
+ "grad_norm": 0.03593013063073158,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 4.311532974243164,
+ "step": 1135
+ },
+ {
+ "epoch": 15.781849912739965,
+ "grad_norm": 0.03564770147204399,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 4.3400726318359375,
+ "step": 1136
+ },
+ {
+ "epoch": 15.795811518324607,
+ "grad_norm": 0.03598494827747345,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 4.30976676940918,
+ "step": 1137
+ },
+ {
+ "epoch": 15.809773123909249,
+ "grad_norm": 0.034391652792692184,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 4.328464508056641,
+ "step": 1138
+ },
+ {
+ "epoch": 15.823734729493891,
+ "grad_norm": 0.03445672616362572,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 4.310579299926758,
+ "step": 1139
+ },
+ {
+ "epoch": 15.837696335078533,
+ "grad_norm": 0.03643282502889633,
+ "learning_rate": 0.000479936075927473,
+ "loss": 4.320905685424805,
+ "step": 1140
+ },
+ {
+ "epoch": 15.851657940663177,
+ "grad_norm": 0.03489850088953972,
+ "learning_rate": 0.000479721889242305,
+ "loss": 4.317229270935059,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86561954624782,
+ "grad_norm": 0.03609171137213707,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 4.309203147888184,
+ "step": 1142
+ },
+ {
+ "epoch": 15.879581151832461,
+ "grad_norm": 0.03705338388681412,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 4.287387371063232,
+ "step": 1143
+ },
+ {
+ "epoch": 15.893542757417103,
+ "grad_norm": 0.03697717934846878,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 4.302860260009766,
+ "step": 1144
+ },
+ {
+ "epoch": 15.907504363001745,
+ "grad_norm": 0.035338208079338074,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 4.3152265548706055,
+ "step": 1145
+ },
+ {
+ "epoch": 15.921465968586388,
+ "grad_norm": 0.03548634052276611,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 4.327335357666016,
+ "step": 1146
+ },
+ {
+ "epoch": 15.93542757417103,
+ "grad_norm": 0.035512328147888184,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 4.3067803382873535,
+ "step": 1147
+ },
+ {
+ "epoch": 15.949389179755672,
+ "grad_norm": 0.03799289092421532,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 4.320701599121094,
+ "step": 1148
+ },
+ {
+ "epoch": 15.963350785340314,
+ "grad_norm": 0.03866339474916458,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 4.321087837219238,
+ "step": 1149
+ },
+ {
+ "epoch": 15.977312390924956,
+ "grad_norm": 0.036002084612846375,
+ "learning_rate": 0.000477787822648007,
+ "loss": 4.287045478820801,
+ "step": 1150
+ },
+ {
+ "epoch": 15.991273996509598,
+ "grad_norm": 0.03433441370725632,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 4.2562665939331055,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.02580358274281025,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 2.67783260345459,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.5870338082313538,
+ "eval_runtime": 60.0469,
+ "eval_samples_per_second": 40.668,
+ "eval_steps_per_second": 0.649,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013961605584644,
+ "grad_norm": 0.04105810821056366,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 4.224295616149902,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027923211169284,
+ "grad_norm": 0.049538832157850266,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 4.222184181213379,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041884816753928,
+ "grad_norm": 0.05089569091796875,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 4.2067060470581055,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05584642233857,
+ "grad_norm": 0.04600589722394943,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 4.238527774810791,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069808027923212,
+ "grad_norm": 0.043311409652233124,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 4.242960453033447,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083769633507853,
+ "grad_norm": 0.047897014766931534,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 4.199069023132324,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097731239092496,
+ "grad_norm": 0.05633604899048805,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 4.211665153503418,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111692844677137,
+ "grad_norm": 0.056110311299562454,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 4.20163631439209,
+ "step": 1160
+ },
+ {
+ "epoch": 16.12565445026178,
+ "grad_norm": 0.05426819995045662,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 4.21889591217041,
+ "step": 1161
+ },
+ {
+ "epoch": 16.13961605584642,
+ "grad_norm": 0.05489835515618324,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 4.220406532287598,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153577661431065,
+ "grad_norm": 0.053941354155540466,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 4.231909275054932,
+ "step": 1163
+ },
+ {
+ "epoch": 16.167539267015705,
+ "grad_norm": 0.058548808097839355,
+ "learning_rate": 0.000474756648870413,
+ "loss": 4.237003326416016,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18150087260035,
+ "grad_norm": 0.05370597913861275,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 4.240085124969482,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195462478184993,
+ "grad_norm": 0.05684991553425789,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 4.203842639923096,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209424083769633,
+ "grad_norm": 0.060745254158973694,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 4.247998237609863,
+ "step": 1167
+ },
+ {
+ "epoch": 16.223385689354277,
+ "grad_norm": 0.058296725153923035,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 4.226161003112793,
+ "step": 1168
+ },
+ {
+ "epoch": 16.237347294938917,
+ "grad_norm": 0.05310584232211113,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 4.229870796203613,
+ "step": 1169
+ },
+ {
+ "epoch": 16.25130890052356,
+ "grad_norm": 0.04710390046238899,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 4.240084648132324,
+ "step": 1170
+ },
+ {
+ "epoch": 16.2652705061082,
+ "grad_norm": 0.05003240332007408,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 4.240972518920898,
+ "step": 1171
+ },
+ {
+ "epoch": 16.279232111692846,
+ "grad_norm": 0.05552851781249046,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 4.248024940490723,
+ "step": 1172
+ },
+ {
+ "epoch": 16.293193717277486,
+ "grad_norm": 0.049967192113399506,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 4.279229164123535,
+ "step": 1173
+ },
+ {
+ "epoch": 16.30715532286213,
+ "grad_norm": 0.04883681237697601,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 4.269807815551758,
+ "step": 1174
+ },
+ {
+ "epoch": 16.32111692844677,
+ "grad_norm": 0.04931480810046196,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 4.232419490814209,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335078534031414,
+ "grad_norm": 0.048746511340141296,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 4.267911911010742,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349040139616054,
+ "grad_norm": 0.04678189381957054,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 4.262979507446289,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363001745200698,
+ "grad_norm": 0.048893529921770096,
+ "learning_rate": 0.000471698352726896,
+ "loss": 4.254408359527588,
+ "step": 1178
+ },
+ {
+ "epoch": 16.376963350785342,
+ "grad_norm": 0.045027732849121094,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 4.2445902824401855,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390924956369982,
+ "grad_norm": 0.04324870929121971,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 4.2685136795043945,
+ "step": 1180
+ },
+ {
+ "epoch": 16.404886561954626,
+ "grad_norm": 0.0470522940158844,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 4.237505912780762,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418848167539267,
+ "grad_norm": 0.04896962642669678,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 4.2523345947265625,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43280977312391,
+ "grad_norm": 0.04562787711620331,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 4.2192792892456055,
+ "step": 1183
+ },
+ {
+ "epoch": 16.44677137870855,
+ "grad_norm": 0.04414575919508934,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 4.250238418579102,
+ "step": 1184
+ },
+ {
+ "epoch": 16.460732984293195,
+ "grad_norm": 0.04363945871591568,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 4.256250381469727,
+ "step": 1185
+ },
+ {
+ "epoch": 16.474694589877835,
+ "grad_norm": 0.04159797728061676,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 4.274903297424316,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48865619546248,
+ "grad_norm": 0.04248544201254845,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 4.278222560882568,
+ "step": 1187
+ },
+ {
+ "epoch": 16.50261780104712,
+ "grad_norm": 0.04379890486598015,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 4.267505645751953,
+ "step": 1188
+ },
+ {
+ "epoch": 16.516579406631763,
+ "grad_norm": 0.04319411888718605,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 4.2607269287109375,
+ "step": 1189
+ },
+ {
+ "epoch": 16.530541012216403,
+ "grad_norm": 0.04143810644745827,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 4.267977714538574,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544502617801047,
+ "grad_norm": 0.03914840891957283,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 4.260498523712158,
+ "step": 1191
+ },
+ {
+ "epoch": 16.55846422338569,
+ "grad_norm": 0.04107543081045151,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 4.2550482749938965,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57242582897033,
+ "grad_norm": 0.041316065937280655,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 4.3075456619262695,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586387434554975,
+ "grad_norm": 0.04069342464208603,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 4.234392166137695,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600349040139616,
+ "grad_norm": 0.03816857561469078,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 4.251038551330566,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61431064572426,
+ "grad_norm": 0.04095374792814255,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 4.284790992736816,
+ "step": 1196
+ },
+ {
+ "epoch": 16.6282722513089,
+ "grad_norm": 0.03788160905241966,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 4.255716323852539,
+ "step": 1197
+ },
+ {
+ "epoch": 16.642233856893544,
+ "grad_norm": 0.03784061595797539,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 4.269606590270996,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656195462478184,
+ "grad_norm": 0.04087168723344803,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 4.231165885925293,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670157068062828,
+ "grad_norm": 0.040546853095293045,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 4.288422584533691,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68411867364747,
+ "grad_norm": 0.03777262568473816,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 4.277492523193359,
+ "step": 1201
+ },
+ {
+ "epoch": 16.698080279232112,
+ "grad_norm": 0.039857733994722366,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 4.239383697509766,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712041884816752,
+ "grad_norm": 0.04387792944908142,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 4.300265312194824,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726003490401396,
+ "grad_norm": 0.04214833676815033,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 4.255545616149902,
+ "step": 1204
+ },
+ {
+ "epoch": 16.739965095986037,
+ "grad_norm": 0.04290056228637695,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 4.275530815124512,
+ "step": 1205
+ },
+ {
+ "epoch": 16.75392670157068,
+ "grad_norm": 0.03975554555654526,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 4.282685279846191,
+ "step": 1206
+ },
+ {
+ "epoch": 16.767888307155324,
+ "grad_norm": 0.039464011788368225,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 4.285504341125488,
+ "step": 1207
+ },
+ {
+ "epoch": 16.781849912739965,
+ "grad_norm": 0.039714012295007706,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 4.273477554321289,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79581151832461,
+ "grad_norm": 0.03936111181974411,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 4.304205894470215,
+ "step": 1209
+ },
+ {
+ "epoch": 16.80977312390925,
+ "grad_norm": 0.03782064467668533,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 4.307382583618164,
+ "step": 1210
+ },
+ {
+ "epoch": 16.823734729493893,
+ "grad_norm": 0.040340717881917953,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 4.308620452880859,
+ "step": 1211
+ },
+ {
+ "epoch": 16.837696335078533,
+ "grad_norm": 0.0403442308306694,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 4.248626708984375,
+ "step": 1212
+ },
+ {
+ "epoch": 16.851657940663177,
+ "grad_norm": 0.038351960480213165,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 4.261248588562012,
+ "step": 1213
+ },
+ {
+ "epoch": 16.865619546247817,
+ "grad_norm": 0.03965333476662636,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 4.278291702270508,
+ "step": 1214
+ },
+ {
+ "epoch": 16.87958115183246,
+ "grad_norm": 0.040909092873334885,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 4.241847991943359,
+ "step": 1215
+ },
+ {
+ "epoch": 16.8935427574171,
+ "grad_norm": 0.038684047758579254,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 4.281346321105957,
+ "step": 1216
+ },
+ {
+ "epoch": 16.907504363001745,
+ "grad_norm": 0.03698103874921799,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 4.3085432052612305,
+ "step": 1217
+ },
+ {
+ "epoch": 16.921465968586386,
+ "grad_norm": 0.03908613696694374,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 4.24220085144043,
+ "step": 1218
+ },
+ {
+ "epoch": 16.93542757417103,
+ "grad_norm": 0.03873388469219208,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 4.282486915588379,
+ "step": 1219
+ },
+ {
+ "epoch": 16.949389179755673,
+ "grad_norm": 0.03709883987903595,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 4.279009819030762,
+ "step": 1220
+ },
+ {
+ "epoch": 16.963350785340314,
+ "grad_norm": 0.03937513753771782,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 4.273211479187012,
+ "step": 1221
+ },
+ {
+ "epoch": 16.977312390924958,
+ "grad_norm": 0.038478728383779526,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 4.281154155731201,
+ "step": 1222
+ },
+ {
+ "epoch": 16.991273996509598,
+ "grad_norm": 0.03913680836558342,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 4.269218444824219,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.02641749382019043,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 2.6905035972595215,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.5879416465759277,
+ "eval_runtime": 60.2807,
+ "eval_samples_per_second": 40.511,
+ "eval_steps_per_second": 0.647,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013961605584644,
+ "grad_norm": 0.048571839928627014,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 4.2008538246154785,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027923211169284,
+ "grad_norm": 0.0640900656580925,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 4.180826663970947,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041884816753928,
+ "grad_norm": 0.06458134204149246,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 4.199562072753906,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05584642233857,
+ "grad_norm": 0.0563119500875473,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 4.171557426452637,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069808027923212,
+ "grad_norm": 0.050914082676172256,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 4.183539390563965,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083769633507853,
+ "grad_norm": 0.053884491324424744,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 4.210450172424316,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097731239092496,
+ "grad_norm": 0.05301345884799957,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 4.1553192138671875,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111692844677137,
+ "grad_norm": 0.051138825714588165,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 4.158510208129883,
+ "step": 1232
+ },
+ {
+ "epoch": 17.12565445026178,
+ "grad_norm": 0.05355675891041756,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 4.167689323425293,
+ "step": 1233
+ },
+ {
+ "epoch": 17.13961605584642,
+ "grad_norm": 0.06062275916337967,
+ "learning_rate": 0.000459203665939781,
+ "loss": 4.170838356018066,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153577661431065,
+ "grad_norm": 0.058666616678237915,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 4.187441825866699,
+ "step": 1235
+ },
+ {
+ "epoch": 17.167539267015705,
+ "grad_norm": 0.057622168213129044,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 4.190746784210205,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18150087260035,
+ "grad_norm": 0.069090336561203,
+ "learning_rate": 0.000458522886595554,
+ "loss": 4.224946022033691,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195462478184993,
+ "grad_norm": 0.06552847474813461,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 4.204067230224609,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209424083769633,
+ "grad_norm": 0.06602177768945694,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 4.189286708831787,
+ "step": 1239
+ },
+ {
+ "epoch": 17.223385689354277,
+ "grad_norm": 0.06751416623592377,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 4.208927154541016,
+ "step": 1240
+ },
+ {
+ "epoch": 17.237347294938917,
+ "grad_norm": 0.0706794261932373,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 4.212923049926758,
+ "step": 1241
+ },
+ {
+ "epoch": 17.25130890052356,
+ "grad_norm": 0.05346836522221565,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 4.2221784591674805,
+ "step": 1242
+ },
+ {
+ "epoch": 17.2652705061082,
+ "grad_norm": 0.05395734682679176,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 4.196428298950195,
+ "step": 1243
+ },
+ {
+ "epoch": 17.279232111692846,
+ "grad_norm": 0.05496550351381302,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 4.208031177520752,
+ "step": 1244
+ },
+ {
+ "epoch": 17.293193717277486,
+ "grad_norm": 0.05320895463228226,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 4.232810974121094,
+ "step": 1245
+ },
+ {
+ "epoch": 17.30715532286213,
+ "grad_norm": 0.05064759403467178,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 4.207826614379883,
+ "step": 1246
+ },
+ {
+ "epoch": 17.32111692844677,
+ "grad_norm": 0.04287239536643028,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 4.195648670196533,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335078534031414,
+ "grad_norm": 0.04527301341295242,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 4.19954776763916,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349040139616054,
+ "grad_norm": 0.04421805590391159,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 4.248671054840088,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363001745200698,
+ "grad_norm": 0.04487175494432449,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 4.2072954177856445,
+ "step": 1250
+ },
+ {
+ "epoch": 17.376963350785342,
+ "grad_norm": 0.04415794834494591,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 4.222837448120117,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390924956369982,
+ "grad_norm": 0.04474013298749924,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 4.213104248046875,
+ "step": 1252
+ },
+ {
+ "epoch": 17.404886561954626,
+ "grad_norm": 0.042631037533283234,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 4.209726333618164,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418848167539267,
+ "grad_norm": 0.04424271360039711,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 4.202699661254883,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43280977312391,
+ "grad_norm": 0.04151812940835953,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 4.190732955932617,
+ "step": 1255
+ },
+ {
+ "epoch": 17.44677137870855,
+ "grad_norm": 0.041588131338357925,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 4.222360610961914,
+ "step": 1256
+ },
+ {
+ "epoch": 17.460732984293195,
+ "grad_norm": 0.042378414422273636,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 4.198424339294434,
+ "step": 1257
+ },
+ {
+ "epoch": 17.474694589877835,
+ "grad_norm": 0.041907574981451035,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 4.224804401397705,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48865619546248,
+ "grad_norm": 0.043774355202913284,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 4.2276611328125,
+ "step": 1259
+ },
+ {
+ "epoch": 17.50261780104712,
+ "grad_norm": 0.04238573834300041,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 4.228109359741211,
+ "step": 1260
+ },
+ {
+ "epoch": 17.516579406631763,
+ "grad_norm": 0.045351386070251465,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 4.208807468414307,
+ "step": 1261
+ },
+ {
+ "epoch": 17.530541012216403,
+ "grad_norm": 0.04300729185342789,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 4.257306098937988,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544502617801047,
+ "grad_norm": 0.041442908346652985,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 4.208004951477051,
+ "step": 1263
+ },
+ {
+ "epoch": 17.55846422338569,
+ "grad_norm": 0.04337337985634804,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 4.23393440246582,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57242582897033,
+ "grad_norm": 0.04335462674498558,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 4.228087425231934,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586387434554975,
+ "grad_norm": 0.03987223282456398,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 4.213971138000488,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600349040139616,
+ "grad_norm": 0.041760094463825226,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 4.248366355895996,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61431064572426,
+ "grad_norm": 0.04420867934823036,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 4.278865814208984,
+ "step": 1268
+ },
+ {
+ "epoch": 17.6282722513089,
+ "grad_norm": 0.04244982823729515,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 4.223642349243164,
+ "step": 1269
+ },
+ {
+ "epoch": 17.642233856893544,
+ "grad_norm": 0.04280664771795273,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 4.231023788452148,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656195462478184,
+ "grad_norm": 0.04294133558869362,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 4.222380638122559,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670157068062828,
+ "grad_norm": 0.04179249703884125,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 4.215641975402832,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68411867364747,
+ "grad_norm": 0.04038798063993454,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 4.228580474853516,
+ "step": 1273
+ },
+ {
+ "epoch": 17.698080279232112,
+ "grad_norm": 0.04211528226733208,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 4.245234489440918,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712041884816752,
+ "grad_norm": 0.04102989658713341,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 4.218495845794678,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726003490401396,
+ "grad_norm": 0.04038692265748978,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 4.2466511726379395,
+ "step": 1276
+ },
+ {
+ "epoch": 17.739965095986037,
+ "grad_norm": 0.03958791121840477,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 4.237762451171875,
+ "step": 1277
+ },
+ {
+ "epoch": 17.75392670157068,
+ "grad_norm": 0.04129907861351967,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 4.238739013671875,
+ "step": 1278
+ },
+ {
+ "epoch": 17.767888307155324,
+ "grad_norm": 0.042581796646118164,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 4.235558032989502,
+ "step": 1279
+ },
+ {
+ "epoch": 17.781849912739965,
+ "grad_norm": 0.04339978098869324,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 4.226593971252441,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79581151832461,
+ "grad_norm": 0.03808023780584335,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 4.238213062286377,
+ "step": 1281
+ },
+ {
+ "epoch": 17.80977312390925,
+ "grad_norm": 0.04022229090332985,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 4.214224815368652,
+ "step": 1282
+ },
+ {
+ "epoch": 17.823734729493893,
+ "grad_norm": 0.04379156231880188,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 4.253068447113037,
+ "step": 1283
+ },
+ {
+ "epoch": 17.837696335078533,
+ "grad_norm": 0.04232964664697647,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 4.249903202056885,
+ "step": 1284
+ },
+ {
+ "epoch": 17.851657940663177,
+ "grad_norm": 0.040980905294418335,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 4.252640724182129,
+ "step": 1285
+ },
+ {
+ "epoch": 17.865619546247817,
+ "grad_norm": 0.04015526548027992,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 4.234862327575684,
+ "step": 1286
+ },
+ {
+ "epoch": 17.87958115183246,
+ "grad_norm": 0.04273828864097595,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 4.216028213500977,
+ "step": 1287
+ },
+ {
+ "epoch": 17.8935427574171,
+ "grad_norm": 0.04210563376545906,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 4.23359489440918,
+ "step": 1288
+ },
+ {
+ "epoch": 17.907504363001745,
+ "grad_norm": 0.038766685873270035,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 4.220360279083252,
+ "step": 1289
+ },
+ {
+ "epoch": 17.921465968586386,
+ "grad_norm": 0.03921609744429588,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 4.253805160522461,
+ "step": 1290
+ },
+ {
+ "epoch": 17.93542757417103,
+ "grad_norm": 0.04071544483304024,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 4.211221694946289,
+ "step": 1291
+ },
+ {
+ "epoch": 17.949389179755673,
+ "grad_norm": 0.03996877372264862,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 4.241739273071289,
+ "step": 1292
+ },
+ {
+ "epoch": 17.963350785340314,
+ "grad_norm": 0.03878853842616081,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 4.237001419067383,
+ "step": 1293
+ },
+ {
+ "epoch": 17.977312390924958,
+ "grad_norm": 0.03889666125178337,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 4.244929790496826,
+ "step": 1294
+ },
+ {
+ "epoch": 17.991273996509598,
+ "grad_norm": 0.03876441717147827,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 4.236083030700684,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.027286000549793243,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 2.636739730834961,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.5896694660186768,
+ "eval_runtime": 60.4742,
+ "eval_samples_per_second": 40.381,
+ "eval_steps_per_second": 0.645,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013961605584644,
+ "grad_norm": 0.04279542714357376,
+ "learning_rate": 0.000444677103403854,
+ "loss": 4.136401653289795,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027923211169284,
+ "grad_norm": 0.051685452461242676,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 4.159584045410156,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041884816753928,
+ "grad_norm": 0.05411457270383835,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 4.160367965698242,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05584642233857,
+ "grad_norm": 0.048574138432741165,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 4.134189128875732,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069808027923212,
+ "grad_norm": 0.04528478533029556,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 4.1022725105285645,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083769633507853,
+ "grad_norm": 0.04474090039730072,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 4.16152286529541,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097731239092496,
+ "grad_norm": 0.04882791265845299,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 4.1553144454956055,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111692844677137,
+ "grad_norm": 0.04831654578447342,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 4.147022724151611,
+ "step": 1304
+ },
+ {
+ "epoch": 18.12565445026178,
+ "grad_norm": 0.047365427017211914,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 4.139405250549316,
+ "step": 1305
+ },
+ {
+ "epoch": 18.13961605584642,
+ "grad_norm": 0.0514012910425663,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 4.115568161010742,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153577661431065,
+ "grad_norm": 0.05225161090493202,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 4.166429042816162,
+ "step": 1307
+ },
+ {
+ "epoch": 18.167539267015705,
+ "grad_norm": 0.05575249716639519,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 4.1414289474487305,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18150087260035,
+ "grad_norm": 0.05405383184552193,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 4.138680934906006,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195462478184993,
+ "grad_norm": 0.05185031518340111,
+ "learning_rate": 0.000441621628895411,
+ "loss": 4.159243583679199,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209424083769633,
+ "grad_norm": 0.05070916935801506,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 4.1844892501831055,
+ "step": 1311
+ },
+ {
+ "epoch": 18.223385689354277,
+ "grad_norm": 0.05037331581115723,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 4.132781028747559,
+ "step": 1312
+ },
+ {
+ "epoch": 18.237347294938917,
+ "grad_norm": 0.05082111805677414,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 4.143031120300293,
+ "step": 1313
+ },
+ {
+ "epoch": 18.25130890052356,
+ "grad_norm": 0.04742169380187988,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 4.137580871582031,
+ "step": 1314
+ },
+ {
+ "epoch": 18.2652705061082,
+ "grad_norm": 0.049042731523513794,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 4.159721374511719,
+ "step": 1315
+ },
+ {
+ "epoch": 18.279232111692846,
+ "grad_norm": 0.045649368315935135,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 4.144008636474609,
+ "step": 1316
+ },
+ {
+ "epoch": 18.293193717277486,
+ "grad_norm": 0.0473787859082222,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 4.164243698120117,
+ "step": 1317
+ },
+ {
+ "epoch": 18.30715532286213,
+ "grad_norm": 0.047626055777072906,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 4.164441108703613,
+ "step": 1318
+ },
+ {
+ "epoch": 18.32111692844677,
+ "grad_norm": 0.04738573357462883,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 4.180338382720947,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335078534031414,
+ "grad_norm": 0.04712498560547829,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 4.146285533905029,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349040139616054,
+ "grad_norm": 0.045505691319704056,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 4.162572860717773,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363001745200698,
+ "grad_norm": 0.0489528588950634,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 4.160312175750732,
+ "step": 1322
+ },
+ {
+ "epoch": 18.376963350785342,
+ "grad_norm": 0.047517575323581696,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 4.164361953735352,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390924956369982,
+ "grad_norm": 0.044214535504579544,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 4.182154655456543,
+ "step": 1324
+ },
+ {
+ "epoch": 18.404886561954626,
+ "grad_norm": 0.043775733560323715,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 4.130522727966309,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418848167539267,
+ "grad_norm": 0.044364530593156815,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 4.151247024536133,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43280977312391,
+ "grad_norm": 0.046320103108882904,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 4.204159736633301,
+ "step": 1327
+ },
+ {
+ "epoch": 18.44677137870855,
+ "grad_norm": 0.04802565276622772,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 4.182550430297852,
+ "step": 1328
+ },
+ {
+ "epoch": 18.460732984293195,
+ "grad_norm": 0.049340926110744476,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 4.163658142089844,
+ "step": 1329
+ },
+ {
+ "epoch": 18.474694589877835,
+ "grad_norm": 0.046701837331056595,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 4.170445442199707,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48865619546248,
+ "grad_norm": 0.04746273159980774,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 4.220232963562012,
+ "step": 1331
+ },
+ {
+ "epoch": 18.50261780104712,
+ "grad_norm": 0.04644489288330078,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 4.190027236938477,
+ "step": 1332
+ },
+ {
+ "epoch": 18.516579406631763,
+ "grad_norm": 0.04394262284040451,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 4.159570693969727,
+ "step": 1333
+ },
+ {
+ "epoch": 18.530541012216403,
+ "grad_norm": 0.04609063267707825,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 4.1657586097717285,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544502617801047,
+ "grad_norm": 0.0450211800634861,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 4.184865474700928,
+ "step": 1335
+ },
+ {
+ "epoch": 18.55846422338569,
+ "grad_norm": 0.0475253090262413,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 4.185857772827148,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57242582897033,
+ "grad_norm": 0.04332203418016434,
+ "learning_rate": 0.000435215814386134,
+ "loss": 4.160165309906006,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586387434554975,
+ "grad_norm": 0.0427570678293705,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 4.1727495193481445,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600349040139616,
+ "grad_norm": 0.04651174694299698,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 4.204797744750977,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61431064572426,
+ "grad_norm": 0.04324701428413391,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 4.186931610107422,
+ "step": 1340
+ },
+ {
+ "epoch": 18.6282722513089,
+ "grad_norm": 0.044644393026828766,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 4.209554195404053,
+ "step": 1341
+ },
+ {
+ "epoch": 18.642233856893544,
+ "grad_norm": 0.042982228100299835,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 4.205570220947266,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656195462478184,
+ "grad_norm": 0.04385090991854668,
+ "learning_rate": 0.000433781621332752,
+ "loss": 4.177936553955078,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670157068062828,
+ "grad_norm": 0.04771510139107704,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 4.1795806884765625,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68411867364747,
+ "grad_norm": 0.04551433399319649,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 4.192194938659668,
+ "step": 1345
+ },
+ {
+ "epoch": 18.698080279232112,
+ "grad_norm": 0.043501488864421844,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 4.177781105041504,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712041884816752,
+ "grad_norm": 0.04394034296274185,
+ "learning_rate": 0.000432823382695327,
+ "loss": 4.206354141235352,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726003490401396,
+ "grad_norm": 0.04911153391003609,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 4.205874919891357,
+ "step": 1348
+ },
+ {
+ "epoch": 18.739965095986037,
+ "grad_norm": 0.04602045193314552,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 4.1947021484375,
+ "step": 1349
+ },
+ {
+ "epoch": 18.75392670157068,
+ "grad_norm": 0.04393904283642769,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 4.188337802886963,
+ "step": 1350
+ },
+ {
+ "epoch": 18.767888307155324,
+ "grad_norm": 0.04526393488049507,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 4.175235748291016,
+ "step": 1351
+ },
+ {
+ "epoch": 18.781849912739965,
+ "grad_norm": 0.041930247098207474,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 4.210986137390137,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79581151832461,
+ "grad_norm": 0.044262565672397614,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 4.176301002502441,
+ "step": 1353
+ },
+ {
+ "epoch": 18.80977312390925,
+ "grad_norm": 0.044190581887960434,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 4.1888837814331055,
+ "step": 1354
+ },
+ {
+ "epoch": 18.823734729493893,
+ "grad_norm": 0.045371655374765396,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 4.182244777679443,
+ "step": 1355
+ },
+ {
+ "epoch": 18.837696335078533,
+ "grad_norm": 0.04302282631397247,
+ "learning_rate": 0.000430661245733797,
+ "loss": 4.189350128173828,
+ "step": 1356
+ },
+ {
+ "epoch": 18.851657940663177,
+ "grad_norm": 0.04402372986078262,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 4.233372211456299,
+ "step": 1357
+ },
+ {
+ "epoch": 18.865619546247817,
+ "grad_norm": 0.04397045075893402,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 4.21682071685791,
+ "step": 1358
+ },
+ {
+ "epoch": 18.87958115183246,
+ "grad_norm": 0.04319734871387482,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 4.204183578491211,
+ "step": 1359
+ },
+ {
+ "epoch": 18.8935427574171,
+ "grad_norm": 0.04600509628653526,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 4.206524848937988,
+ "step": 1360
+ },
+ {
+ "epoch": 18.907504363001745,
+ "grad_norm": 0.04364705830812454,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 4.213656425476074,
+ "step": 1361
+ },
+ {
+ "epoch": 18.921465968586386,
+ "grad_norm": 0.044004205614328384,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 4.1841583251953125,
+ "step": 1362
+ },
+ {
+ "epoch": 18.93542757417103,
+ "grad_norm": 0.04480595886707306,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 4.166214942932129,
+ "step": 1363
+ },
+ {
+ "epoch": 18.949389179755673,
+ "grad_norm": 0.04260887950658798,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 4.198726654052734,
+ "step": 1364
+ },
+ {
+ "epoch": 18.963350785340314,
+ "grad_norm": 0.04513927176594734,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 4.2122392654418945,
+ "step": 1365
+ },
+ {
+ "epoch": 18.977312390924958,
+ "grad_norm": 0.04286042973399162,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 4.192383766174316,
+ "step": 1366
+ },
+ {
+ "epoch": 18.991273996509598,
+ "grad_norm": 0.04287346079945564,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 4.221860408782959,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.030670644715428352,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 2.624823570251465,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.5910831689834595,
+ "eval_runtime": 60.5517,
+ "eval_samples_per_second": 40.329,
+ "eval_steps_per_second": 0.644,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013961605584644,
+ "grad_norm": 0.046196095645427704,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 4.104501724243164,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027923211169284,
+ "grad_norm": 0.05348595231771469,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 4.109098434448242,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041884816753928,
+ "grad_norm": 0.05256905406713486,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 4.115832328796387,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05584642233857,
+ "grad_norm": 0.05241430550813675,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 4.079610824584961,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069808027923212,
+ "grad_norm": 0.04983370006084442,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 4.095335006713867,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083769633507853,
+ "grad_norm": 0.04900007322430611,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 4.093306541442871,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097731239092496,
+ "grad_norm": 0.04995666816830635,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 4.111114501953125,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111692844677137,
+ "grad_norm": 0.04875681549310684,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 4.091944694519043,
+ "step": 1376
+ },
+ {
+ "epoch": 19.12565445026178,
+ "grad_norm": 0.05283811315894127,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 4.112549304962158,
+ "step": 1377
+ },
+ {
+ "epoch": 19.13961605584642,
+ "grad_norm": 0.04801992326974869,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 4.108818054199219,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153577661431065,
+ "grad_norm": 0.04751930385828018,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 4.089441299438477,
+ "step": 1379
+ },
+ {
+ "epoch": 19.167539267015705,
+ "grad_norm": 0.045653726905584335,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 4.07722282409668,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18150087260035,
+ "grad_norm": 0.045616209506988525,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 4.097578048706055,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195462478184993,
+ "grad_norm": 0.04929365590214729,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 4.082221984863281,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209424083769633,
+ "grad_norm": 0.04915834963321686,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 4.123330116271973,
+ "step": 1383
+ },
+ {
+ "epoch": 19.223385689354277,
+ "grad_norm": 0.0470132976770401,
+ "learning_rate": 0.000423881957237287,
+ "loss": 4.113927841186523,
+ "step": 1384
+ },
+ {
+ "epoch": 19.237347294938917,
+ "grad_norm": 0.04695259779691696,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 4.096814155578613,
+ "step": 1385
+ },
+ {
+ "epoch": 19.25130890052356,
+ "grad_norm": 0.04730398207902908,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 4.122525215148926,
+ "step": 1386
+ },
+ {
+ "epoch": 19.2652705061082,
+ "grad_norm": 0.050295207649469376,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 4.117993354797363,
+ "step": 1387
+ },
+ {
+ "epoch": 19.279232111692846,
+ "grad_norm": 0.05082492157816887,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 4.14082145690918,
+ "step": 1388
+ },
+ {
+ "epoch": 19.293193717277486,
+ "grad_norm": 0.04960232973098755,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 4.139887809753418,
+ "step": 1389
+ },
+ {
+ "epoch": 19.30715532286213,
+ "grad_norm": 0.04851173236966133,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 4.138492584228516,
+ "step": 1390
+ },
+ {
+ "epoch": 19.32111692844677,
+ "grad_norm": 0.047838468104600906,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 4.118064880371094,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335078534031414,
+ "grad_norm": 0.05148220434784889,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 4.136533737182617,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349040139616054,
+ "grad_norm": 0.053177181631326675,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 4.1125168800354,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363001745200698,
+ "grad_norm": 0.04861171543598175,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 4.1437153816223145,
+ "step": 1394
+ },
+ {
+ "epoch": 19.376963350785342,
+ "grad_norm": 0.05194924771785736,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 4.097813606262207,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390924956369982,
+ "grad_norm": 0.049609847366809845,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 4.068940162658691,
+ "step": 1396
+ },
+ {
+ "epoch": 19.404886561954626,
+ "grad_norm": 0.047042325139045715,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 4.121333122253418,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418848167539267,
+ "grad_norm": 0.047646693885326385,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 4.114943981170654,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43280977312391,
+ "grad_norm": 0.04676460474729538,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 4.139252185821533,
+ "step": 1399
+ },
+ {
+ "epoch": 19.44677137870855,
+ "grad_norm": 0.04905956983566284,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 4.135970115661621,
+ "step": 1400
+ },
+ {
+ "epoch": 19.460732984293195,
+ "grad_norm": 0.050025034695863724,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 4.15662956237793,
+ "step": 1401
+ },
+ {
+ "epoch": 19.474694589877835,
+ "grad_norm": 0.04740425944328308,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 4.140295505523682,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48865619546248,
+ "grad_norm": 0.046904709190130234,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 4.14083194732666,
+ "step": 1403
+ },
+ {
+ "epoch": 19.50261780104712,
+ "grad_norm": 0.048393651843070984,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 4.161823272705078,
+ "step": 1404
+ },
+ {
+ "epoch": 19.516579406631763,
+ "grad_norm": 0.04766222462058067,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 4.151817798614502,
+ "step": 1405
+ },
+ {
+ "epoch": 19.530541012216403,
+ "grad_norm": 0.04797709733247757,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 4.173160552978516,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544502617801047,
+ "grad_norm": 0.05012711137533188,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 4.161989688873291,
+ "step": 1407
+ },
+ {
+ "epoch": 19.55846422338569,
+ "grad_norm": 0.04726000875234604,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 4.136838436126709,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57242582897033,
+ "grad_norm": 0.046251531690359116,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 4.1526994705200195,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586387434554975,
+ "grad_norm": 0.047958049923181534,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 4.13817024230957,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600349040139616,
+ "grad_norm": 0.046333350241184235,
+ "learning_rate": 0.000417272427439646,
+ "loss": 4.160187721252441,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61431064572426,
+ "grad_norm": 0.0488678440451622,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 4.166190147399902,
+ "step": 1412
+ },
+ {
+ "epoch": 19.6282722513089,
+ "grad_norm": 0.04784323647618294,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 4.164414405822754,
+ "step": 1413
+ },
+ {
+ "epoch": 19.642233856893544,
+ "grad_norm": 0.05064673349261284,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 4.149814128875732,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656195462478184,
+ "grad_norm": 0.05072085186839104,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 4.167759895324707,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670157068062828,
+ "grad_norm": 0.0499395951628685,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 4.162792205810547,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68411867364747,
+ "grad_norm": 0.046766918152570724,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 4.1416521072387695,
+ "step": 1417
+ },
+ {
+ "epoch": 19.698080279232112,
+ "grad_norm": 0.04523273929953575,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 4.161751747131348,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712041884816752,
+ "grad_norm": 0.046106480062007904,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 4.166394233703613,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726003490401396,
+ "grad_norm": 0.04666793718934059,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 4.124961853027344,
+ "step": 1420
+ },
+ {
+ "epoch": 19.739965095986037,
+ "grad_norm": 0.047407012432813644,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 4.1794328689575195,
+ "step": 1421
+ },
+ {
+ "epoch": 19.75392670157068,
+ "grad_norm": 0.04690873995423317,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 4.160114765167236,
+ "step": 1422
+ },
+ {
+ "epoch": 19.767888307155324,
+ "grad_norm": 0.04589436575770378,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 4.137969970703125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.781849912739965,
+ "grad_norm": 0.04641961678862572,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 4.161533355712891,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79581151832461,
+ "grad_norm": 0.045799050480127335,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 4.187039375305176,
+ "step": 1425
+ },
+ {
+ "epoch": 19.80977312390925,
+ "grad_norm": 0.04881894588470459,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 4.158107757568359,
+ "step": 1426
+ },
+ {
+ "epoch": 19.823734729493893,
+ "grad_norm": 0.04506978392601013,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 4.143360137939453,
+ "step": 1427
+ },
+ {
+ "epoch": 19.837696335078533,
+ "grad_norm": 0.04504861310124397,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 4.174564361572266,
+ "step": 1428
+ },
+ {
+ "epoch": 19.851657940663177,
+ "grad_norm": 0.04639774188399315,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 4.162899971008301,
+ "step": 1429
+ },
+ {
+ "epoch": 19.865619546247817,
+ "grad_norm": 0.047444432973861694,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 4.143390655517578,
+ "step": 1430
+ },
+ {
+ "epoch": 19.87958115183246,
+ "grad_norm": 0.04461613669991493,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 4.171284198760986,
+ "step": 1431
+ },
+ {
+ "epoch": 19.8935427574171,
+ "grad_norm": 0.046383798122406006,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 4.14809513092041,
+ "step": 1432
+ },
+ {
+ "epoch": 19.907504363001745,
+ "grad_norm": 0.04642366245388985,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 4.178903579711914,
+ "step": 1433
+ },
+ {
+ "epoch": 19.921465968586386,
+ "grad_norm": 0.0451771542429924,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 4.171825408935547,
+ "step": 1434
+ },
+ {
+ "epoch": 19.93542757417103,
+ "grad_norm": 0.04767893627285957,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 4.169270038604736,
+ "step": 1435
+ },
+ {
+ "epoch": 19.949389179755673,
+ "grad_norm": 0.04673412814736366,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 4.175506591796875,
+ "step": 1436
+ },
+ {
+ "epoch": 19.963350785340314,
+ "grad_norm": 0.04465145245194435,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 4.149707794189453,
+ "step": 1437
+ },
+ {
+ "epoch": 19.977312390924958,
+ "grad_norm": 0.04734131321310997,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 4.166044235229492,
+ "step": 1438
+ },
+ {
+ "epoch": 19.991273996509598,
+ "grad_norm": 0.04706178605556488,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 4.152081489562988,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.031556785106658936,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 2.60109543800354,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013961605584644,
+ "grad_norm": 0.05152537673711777,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 4.084830284118652,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027923211169284,
+ "grad_norm": 0.05918274074792862,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 4.045302391052246,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041884816753928,
+ "grad_norm": 0.05093073844909668,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 4.070842742919922,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05584642233857,
+ "grad_norm": 0.05095985531806946,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 4.083273410797119,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069808027923212,
+ "grad_norm": 0.05502687394618988,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 4.072626113891602,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083769633507853,
+ "grad_norm": 0.05037364363670349,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 4.075478553771973,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097731239092496,
+ "grad_norm": 0.04983789846301079,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 4.066207408905029,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111692844677137,
+ "grad_norm": 0.049478255212306976,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 4.069969654083252,
+ "step": 1448
+ },
+ {
+ "epoch": 20.12565445026178,
+ "grad_norm": 0.051651593297719955,
+ "learning_rate": 0.000407857329622967,
+ "loss": 4.050713062286377,
+ "step": 1449
+ },
+ {
+ "epoch": 20.13961605584642,
+ "grad_norm": 0.05436555668711662,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 4.069849967956543,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153577661431065,
+ "grad_norm": 0.055190350860357285,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 4.081624984741211,
+ "step": 1451
+ },
+ {
+ "epoch": 20.167539267015705,
+ "grad_norm": 0.053246211260557175,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 4.076076507568359,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18150087260035,
+ "grad_norm": 0.051757775247097015,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 4.049139022827148,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195462478184993,
+ "grad_norm": 0.05064203217625618,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 4.068076133728027,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209424083769633,
+ "grad_norm": 0.05114085227251053,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 4.092901229858398,
+ "step": 1455
+ },
+ {
+ "epoch": 20.223385689354277,
+ "grad_norm": 0.047997843474149704,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 4.077084541320801,
+ "step": 1456
+ },
+ {
+ "epoch": 20.237347294938917,
+ "grad_norm": 0.04880451038479805,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 4.059187889099121,
+ "step": 1457
+ },
+ {
+ "epoch": 20.25130890052356,
+ "grad_norm": 0.05036284029483795,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 4.102115154266357,
+ "step": 1458
+ },
+ {
+ "epoch": 20.2652705061082,
+ "grad_norm": 0.04916979745030403,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 4.098356246948242,
+ "step": 1459
+ },
+ {
+ "epoch": 20.279232111692846,
+ "grad_norm": 0.04713984206318855,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 4.057218074798584,
+ "step": 1460
+ },
+ {
+ "epoch": 20.293193717277486,
+ "grad_norm": 0.04982447996735573,
+ "learning_rate": 0.000404858275823277,
+ "loss": 4.070363998413086,
+ "step": 1461
+ },
+ {
+ "epoch": 20.30715532286213,
+ "grad_norm": 0.0475616492331028,
+ "learning_rate": 0.000404607816444578,
+ "loss": 4.074314117431641,
+ "step": 1462
+ },
+ {
+ "epoch": 20.32111692844677,
+ "grad_norm": 0.046537093818187714,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 4.079501152038574,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335078534031414,
+ "grad_norm": 0.04889644682407379,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 4.0731425285339355,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349040139616054,
+ "grad_norm": 0.05095166340470314,
+ "learning_rate": 0.000403855947934478,
+ "loss": 4.08765983581543,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363001745200698,
+ "grad_norm": 0.049677830189466476,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 4.128052711486816,
+ "step": 1466
+ },
+ {
+ "epoch": 20.376963350785342,
+ "grad_norm": 0.05221942439675331,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 4.094407081604004,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390924956369982,
+ "grad_norm": 0.052042894065380096,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 4.088282585144043,
+ "step": 1468
+ },
+ {
+ "epoch": 20.404886561954626,
+ "grad_norm": 0.04822041094303131,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 4.1070990562438965,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418848167539267,
+ "grad_norm": 0.047207772731781006,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 4.102565765380859,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43280977312391,
+ "grad_norm": 0.049633774906396866,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 4.097216606140137,
+ "step": 1471
+ },
+ {
+ "epoch": 20.44677137870855,
+ "grad_norm": 0.05135185644030571,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 4.095193862915039,
+ "step": 1472
+ },
+ {
+ "epoch": 20.460732984293195,
+ "grad_norm": 0.05215826258063316,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 4.085207939147949,
+ "step": 1473
+ },
+ {
+ "epoch": 20.474694589877835,
+ "grad_norm": 0.050653353333473206,
+ "learning_rate": 0.000401595976318565,
+ "loss": 4.085142612457275,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48865619546248,
+ "grad_norm": 0.04795852303504944,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 4.1158270835876465,
+ "step": 1475
+ },
+ {
+ "epoch": 20.50261780104712,
+ "grad_norm": 0.047416940331459045,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 4.096989154815674,
+ "step": 1476
+ },
+ {
+ "epoch": 20.516579406631763,
+ "grad_norm": 0.049753669649362564,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 4.118251800537109,
+ "step": 1477
+ },
+ {
+ "epoch": 20.530541012216403,
+ "grad_norm": 0.04889055714011192,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 4.081287384033203,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544502617801047,
+ "grad_norm": 0.04857790097594261,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 4.084896564483643,
+ "step": 1479
+ },
+ {
+ "epoch": 20.55846422338569,
+ "grad_norm": 0.047250498086214066,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 4.113656997680664,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57242582897033,
+ "grad_norm": 0.048608481884002686,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 4.113809108734131,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586387434554975,
+ "grad_norm": 0.048189081251621246,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 4.113949775695801,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600349040139616,
+ "grad_norm": 0.049929242581129074,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 4.104221820831299,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61431064572426,
+ "grad_norm": 0.048482850193977356,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 4.091039180755615,
+ "step": 1484
+ },
+ {
+ "epoch": 20.6282722513089,
+ "grad_norm": 0.04810645803809166,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 4.130088806152344,
+ "step": 1485
+ },
+ {
+ "epoch": 20.642233856893544,
+ "grad_norm": 0.04745674133300781,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 4.106161117553711,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656195462478184,
+ "grad_norm": 0.048513609915971756,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 4.119222164154053,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670157068062828,
+ "grad_norm": 0.04506325349211693,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 4.115543842315674,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68411867364747,
+ "grad_norm": 0.04726741835474968,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 4.084345817565918,
+ "step": 1489
+ },
+ {
+ "epoch": 20.698080279232112,
+ "grad_norm": 0.04873545467853546,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 4.116580009460449,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712041884816752,
+ "grad_norm": 0.051252465695142746,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 4.129733085632324,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726003490401396,
+ "grad_norm": 0.0489567406475544,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 4.103477954864502,
+ "step": 1492
+ },
+ {
+ "epoch": 20.739965095986037,
+ "grad_norm": 0.04539056867361069,
+ "learning_rate": 0.000396803974909638,
+ "loss": 4.1327667236328125,
+ "step": 1493
+ },
+ {
+ "epoch": 20.75392670157068,
+ "grad_norm": 0.04954095557332039,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 4.084133625030518,
+ "step": 1494
+ },
+ {
+ "epoch": 20.767888307155324,
+ "grad_norm": 0.05041591823101044,
+ "learning_rate": 0.000396297942393265,
+ "loss": 4.12288761138916,
+ "step": 1495
+ },
+ {
+ "epoch": 20.781849912739965,
+ "grad_norm": 0.046549517661333084,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 4.1259918212890625,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79581151832461,
+ "grad_norm": 0.04704999923706055,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 4.102563858032227,
+ "step": 1497
+ },
+ {
+ "epoch": 20.80977312390925,
+ "grad_norm": 0.05025320500135422,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 4.135492324829102,
+ "step": 1498
+ },
+ {
+ "epoch": 20.823734729493893,
+ "grad_norm": 0.04559441655874252,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 4.118548393249512,
+ "step": 1499
+ },
+ {
+ "epoch": 20.837696335078533,
+ "grad_norm": 0.04913092032074928,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 4.129302024841309,
+ "step": 1500
+ },
+ {
+ "epoch": 20.851657940663177,
+ "grad_norm": 0.04846089705824852,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 4.166059970855713,
+ "step": 1501
+ },
+ {
+ "epoch": 20.865619546247817,
+ "grad_norm": 0.04710553213953972,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 4.104712009429932,
+ "step": 1502
+ },
+ {
+ "epoch": 20.87958115183246,
+ "grad_norm": 0.04896444454789162,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 4.140282154083252,
+ "step": 1503
+ },
+ {
+ "epoch": 20.8935427574171,
+ "grad_norm": 0.0469740591943264,
+ "learning_rate": 0.000394017098798633,
+ "loss": 4.125306129455566,
+ "step": 1504
+ },
+ {
+ "epoch": 20.907504363001745,
+ "grad_norm": 0.04638523980975151,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 4.133161544799805,
+ "step": 1505
+ },
+ {
+ "epoch": 20.921465968586386,
+ "grad_norm": 0.04814442619681358,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 4.093092918395996,
+ "step": 1506
+ },
+ {
+ "epoch": 20.93542757417103,
+ "grad_norm": 0.049911532551050186,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 4.1001715660095215,
+ "step": 1507
+ },
+ {
+ "epoch": 20.949389179755673,
+ "grad_norm": 0.0446670837700367,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 4.140350341796875,
+ "step": 1508
+ },
+ {
+ "epoch": 20.963350785340314,
+ "grad_norm": 0.04789397493004799,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 4.163155555725098,
+ "step": 1509
+ },
+ {
+ "epoch": 20.977312390924958,
+ "grad_norm": 0.050148315727710724,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 4.118380069732666,
+ "step": 1510
+ },
+ {
+ "epoch": 20.991273996509598,
+ "grad_norm": 0.047446198761463165,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 4.116507530212402,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.03415544703602791,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 2.5738282203674316,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.5933796167373657,
+ "eval_runtime": 60.4011,
+ "eval_samples_per_second": 40.43,
+ "eval_steps_per_second": 0.646,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013961605584644,
+ "grad_norm": 0.052276045083999634,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 4.0301923751831055,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027923211169284,
+ "grad_norm": 0.06514342874288559,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 4.031280517578125,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041884816753928,
+ "grad_norm": 0.05948134884238243,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 4.060680389404297,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05584642233857,
+ "grad_norm": 0.05251640826463699,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 4.0218400955200195,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069808027923212,
+ "grad_norm": 0.05426032841205597,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 4.026187896728516,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083769633507853,
+ "grad_norm": 0.052583567798137665,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 4.016767501831055,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097731239092496,
+ "grad_norm": 0.04966135695576668,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 4.043508529663086,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111692844677137,
+ "grad_norm": 0.053838398307561874,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 4.054832458496094,
+ "step": 1520
+ },
+ {
+ "epoch": 21.12565445026178,
+ "grad_norm": 0.05304184556007385,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 4.03280782699585,
+ "step": 1521
+ },
+ {
+ "epoch": 21.13961605584642,
+ "grad_norm": 0.05252664163708687,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 4.039841175079346,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153577661431065,
+ "grad_norm": 0.054593686014413834,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 4.048308372497559,
+ "step": 1523
+ },
+ {
+ "epoch": 21.167539267015705,
+ "grad_norm": 0.050722211599349976,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 4.051860332489014,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18150087260035,
+ "grad_norm": 0.05248095095157623,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 4.056846618652344,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195462478184993,
+ "grad_norm": 0.05166107788681984,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 4.045178413391113,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209424083769633,
+ "grad_norm": 0.04987792298197746,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 4.014827728271484,
+ "step": 1527
+ },
+ {
+ "epoch": 21.223385689354277,
+ "grad_norm": 0.05316167324781418,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 4.033143997192383,
+ "step": 1528
+ },
+ {
+ "epoch": 21.237347294938917,
+ "grad_norm": 0.05228257551789284,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 4.033842086791992,
+ "step": 1529
+ },
+ {
+ "epoch": 21.25130890052356,
+ "grad_norm": 0.049349382519721985,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 4.025982856750488,
+ "step": 1530
+ },
+ {
+ "epoch": 21.2652705061082,
+ "grad_norm": 0.055104926228523254,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 4.059151649475098,
+ "step": 1531
+ },
+ {
+ "epoch": 21.279232111692846,
+ "grad_norm": 0.052629340440034866,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 4.032320976257324,
+ "step": 1532
+ },
+ {
+ "epoch": 21.293193717277486,
+ "grad_norm": 0.05396415665745735,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 4.0425567626953125,
+ "step": 1533
+ },
+ {
+ "epoch": 21.30715532286213,
+ "grad_norm": 0.054417286068201065,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 4.03759241104126,
+ "step": 1534
+ },
+ {
+ "epoch": 21.32111692844677,
+ "grad_norm": 0.061248697340488434,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 4.0441083908081055,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335078534031414,
+ "grad_norm": 0.05934970825910568,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 4.050718307495117,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349040139616054,
+ "grad_norm": 0.052805181592702866,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 4.069460868835449,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363001745200698,
+ "grad_norm": 0.052820611745119095,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 4.063657283782959,
+ "step": 1538
+ },
+ {
+ "epoch": 21.376963350785342,
+ "grad_norm": 0.05043052136898041,
+ "learning_rate": 0.00038509205478744,
+ "loss": 4.029552459716797,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390924956369982,
+ "grad_norm": 0.04828709363937378,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 4.05656623840332,
+ "step": 1540
+ },
+ {
+ "epoch": 21.404886561954626,
+ "grad_norm": 0.049760788679122925,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 4.066338539123535,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418848167539267,
+ "grad_norm": 0.05037159472703934,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 4.06976842880249,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43280977312391,
+ "grad_norm": 0.05276317894458771,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 4.062975883483887,
+ "step": 1543
+ },
+ {
+ "epoch": 21.44677137870855,
+ "grad_norm": 0.05571872368454933,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 4.065622329711914,
+ "step": 1544
+ },
+ {
+ "epoch": 21.460732984293195,
+ "grad_norm": 0.05438469722867012,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 4.074963569641113,
+ "step": 1545
+ },
+ {
+ "epoch": 21.474694589877835,
+ "grad_norm": 0.053415946662425995,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 4.080055236816406,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48865619546248,
+ "grad_norm": 0.05216296389698982,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 4.091355800628662,
+ "step": 1547
+ },
+ {
+ "epoch": 21.50261780104712,
+ "grad_norm": 0.051212459802627563,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 4.058528900146484,
+ "step": 1548
+ },
+ {
+ "epoch": 21.516579406631763,
+ "grad_norm": 0.05033959075808525,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 4.064615726470947,
+ "step": 1549
+ },
+ {
+ "epoch": 21.530541012216403,
+ "grad_norm": 0.04665188118815422,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 4.059698104858398,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544502617801047,
+ "grad_norm": 0.050221361219882965,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 4.068614959716797,
+ "step": 1551
+ },
+ {
+ "epoch": 21.55846422338569,
+ "grad_norm": 0.05089551955461502,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 4.0833892822265625,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57242582897033,
+ "grad_norm": 0.051966603845357895,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 4.083354949951172,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586387434554975,
+ "grad_norm": 0.04981927201151848,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 4.083340644836426,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600349040139616,
+ "grad_norm": 0.05141765624284744,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 4.066590309143066,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61431064572426,
+ "grad_norm": 0.048985838890075684,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 4.088237762451172,
+ "step": 1556
+ },
+ {
+ "epoch": 21.6282722513089,
+ "grad_norm": 0.05013054236769676,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 4.081377983093262,
+ "step": 1557
+ },
+ {
+ "epoch": 21.642233856893544,
+ "grad_norm": 0.05170975625514984,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 4.054303169250488,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656195462478184,
+ "grad_norm": 0.048257142305374146,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 4.103672504425049,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670157068062828,
+ "grad_norm": 0.049942679703235626,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 4.029951572418213,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68411867364747,
+ "grad_norm": 0.05083315819501877,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 4.048798084259033,
+ "step": 1561
+ },
+ {
+ "epoch": 21.698080279232112,
+ "grad_norm": 0.04984128847718239,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 4.079655647277832,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712041884816752,
+ "grad_norm": 0.04985307902097702,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 4.071877956390381,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726003490401396,
+ "grad_norm": 0.051137227565050125,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 4.076539993286133,
+ "step": 1564
+ },
+ {
+ "epoch": 21.739965095986037,
+ "grad_norm": 0.04968905821442604,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 4.084498882293701,
+ "step": 1565
+ },
+ {
+ "epoch": 21.75392670157068,
+ "grad_norm": 0.049119509756565094,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 4.066773414611816,
+ "step": 1566
+ },
+ {
+ "epoch": 21.767888307155324,
+ "grad_norm": 0.05107425898313522,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 4.065309047698975,
+ "step": 1567
+ },
+ {
+ "epoch": 21.781849912739965,
+ "grad_norm": 0.049012504518032074,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 4.062110900878906,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79581151832461,
+ "grad_norm": 0.04972011595964432,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 4.0630903244018555,
+ "step": 1569
+ },
+ {
+ "epoch": 21.80977312390925,
+ "grad_norm": 0.04822700098156929,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 4.106127738952637,
+ "step": 1570
+ },
+ {
+ "epoch": 21.823734729493893,
+ "grad_norm": 0.04808267205953598,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 4.085200309753418,
+ "step": 1571
+ },
+ {
+ "epoch": 21.837696335078533,
+ "grad_norm": 0.048644762486219406,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 4.0856170654296875,
+ "step": 1572
+ },
+ {
+ "epoch": 21.851657940663177,
+ "grad_norm": 0.05232283100485802,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 4.126387119293213,
+ "step": 1573
+ },
+ {
+ "epoch": 21.865619546247817,
+ "grad_norm": 0.05095643922686577,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 4.108254432678223,
+ "step": 1574
+ },
+ {
+ "epoch": 21.87958115183246,
+ "grad_norm": 0.045972101390361786,
+ "learning_rate": 0.000375827577564042,
+ "loss": 4.058515548706055,
+ "step": 1575
+ },
+ {
+ "epoch": 21.8935427574171,
+ "grad_norm": 0.04974910616874695,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 4.089197158813477,
+ "step": 1576
+ },
+ {
+ "epoch": 21.907504363001745,
+ "grad_norm": 0.04890159144997597,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 4.1078410148620605,
+ "step": 1577
+ },
+ {
+ "epoch": 21.921465968586386,
+ "grad_norm": 0.04738525673747063,
+ "learning_rate": 0.000375051971546236,
+ "loss": 4.1219282150268555,
+ "step": 1578
+ },
+ {
+ "epoch": 21.93542757417103,
+ "grad_norm": 0.04713096842169762,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 4.10295295715332,
+ "step": 1579
+ },
+ {
+ "epoch": 21.949389179755673,
+ "grad_norm": 0.04834369197487831,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 4.102747917175293,
+ "step": 1580
+ },
+ {
+ "epoch": 21.963350785340314,
+ "grad_norm": 0.047073688358068466,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 4.08024787902832,
+ "step": 1581
+ },
+ {
+ "epoch": 21.977312390924958,
+ "grad_norm": 0.04568164795637131,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 4.078930854797363,
+ "step": 1582
+ },
+ {
+ "epoch": 21.991273996509598,
+ "grad_norm": 0.04822581633925438,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 4.089601516723633,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.03507270663976669,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 2.5586814880371094,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.595272958278656,
+ "eval_runtime": 60.7797,
+ "eval_samples_per_second": 40.178,
+ "eval_steps_per_second": 0.642,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013961605584644,
+ "grad_norm": 0.05232533439993858,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 4.011311054229736,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027923211169284,
+ "grad_norm": 0.058515798300504684,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 4.0049638748168945,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041884816753928,
+ "grad_norm": 0.055082619190216064,
+ "learning_rate": 0.000372722041257003,
+ "loss": 3.9904184341430664,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05584642233857,
+ "grad_norm": 0.05213981494307518,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 4.003454685211182,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069808027923212,
+ "grad_norm": 0.051619648933410645,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 3.9900388717651367,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083769633507853,
+ "grad_norm": 0.055212799459695816,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 3.973494052886963,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097731239092496,
+ "grad_norm": 0.05262092128396034,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 4.00933837890625,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111692844677137,
+ "grad_norm": 0.05012505501508713,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 4.01861047744751,
+ "step": 1592
+ },
+ {
+ "epoch": 22.12565445026178,
+ "grad_norm": 0.049657098948955536,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 4.011143684387207,
+ "step": 1593
+ },
+ {
+ "epoch": 22.13961605584642,
+ "grad_norm": 0.05187768116593361,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 3.9922969341278076,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153577661431065,
+ "grad_norm": 0.05053112655878067,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 4.007021903991699,
+ "step": 1595
+ },
+ {
+ "epoch": 22.167539267015705,
+ "grad_norm": 0.048542529344558716,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 3.9966416358947754,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18150087260035,
+ "grad_norm": 0.050972841680049896,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 3.9803225994110107,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195462478184993,
+ "grad_norm": 0.05001165345311165,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 4.0140509605407715,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209424083769633,
+ "grad_norm": 0.05100536718964577,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 3.9906058311462402,
+ "step": 1599
+ },
+ {
+ "epoch": 22.223385689354277,
+ "grad_norm": 0.052977193146944046,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 4.012383460998535,
+ "step": 1600
+ },
+ {
+ "epoch": 22.237347294938917,
+ "grad_norm": 0.051219623535871506,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 3.982931613922119,
+ "step": 1601
+ },
+ {
+ "epoch": 22.25130890052356,
+ "grad_norm": 0.05275029316544533,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 4.032669544219971,
+ "step": 1602
+ },
+ {
+ "epoch": 22.2652705061082,
+ "grad_norm": 0.04781945049762726,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 3.9933102130889893,
+ "step": 1603
+ },
+ {
+ "epoch": 22.279232111692846,
+ "grad_norm": 0.054087597876787186,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 4.0075201988220215,
+ "step": 1604
+ },
+ {
+ "epoch": 22.293193717277486,
+ "grad_norm": 0.05485915020108223,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 4.004299640655518,
+ "step": 1605
+ },
+ {
+ "epoch": 22.30715532286213,
+ "grad_norm": 0.05148882791399956,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 4.044161796569824,
+ "step": 1606
+ },
+ {
+ "epoch": 22.32111692844677,
+ "grad_norm": 0.051912397146224976,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 3.9777069091796875,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335078534031414,
+ "grad_norm": 0.052630942314863205,
+ "learning_rate": 0.000367268105959126,
+ "loss": 4.029580116271973,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349040139616054,
+ "grad_norm": 0.0512809231877327,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 4.022092342376709,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363001745200698,
+ "grad_norm": 0.05094102770090103,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 4.018102169036865,
+ "step": 1610
+ },
+ {
+ "epoch": 22.376963350785342,
+ "grad_norm": 0.04972759634256363,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 4.027868270874023,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390924956369982,
+ "grad_norm": 0.05072534456849098,
+ "learning_rate": 0.000366226612810024,
+ "loss": 3.995988368988037,
+ "step": 1612
+ },
+ {
+ "epoch": 22.404886561954626,
+ "grad_norm": 0.050623781979084015,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 4.019711494445801,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418848167539267,
+ "grad_norm": 0.05035007372498512,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 4.007514953613281,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43280977312391,
+ "grad_norm": 0.054426416754722595,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 4.0167131423950195,
+ "step": 1615
+ },
+ {
+ "epoch": 22.44677137870855,
+ "grad_norm": 0.05365690588951111,
+ "learning_rate": 0.000365184304613104,
+ "loss": 4.0313920974731445,
+ "step": 1616
+ },
+ {
+ "epoch": 22.460732984293195,
+ "grad_norm": 0.0509161539375782,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 4.022063255310059,
+ "step": 1617
+ },
+ {
+ "epoch": 22.474694589877835,
+ "grad_norm": 0.05309660732746124,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 4.017317771911621,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48865619546248,
+ "grad_norm": 0.0488961897790432,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 4.028288841247559,
+ "step": 1619
+ },
+ {
+ "epoch": 22.50261780104712,
+ "grad_norm": 0.04982095956802368,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 4.007115840911865,
+ "step": 1620
+ },
+ {
+ "epoch": 22.516579406631763,
+ "grad_norm": 0.05136589705944061,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 4.028022766113281,
+ "step": 1621
+ },
+ {
+ "epoch": 22.530541012216403,
+ "grad_norm": 0.04865279048681259,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 4.015928268432617,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544502617801047,
+ "grad_norm": 0.05131891369819641,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 4.041997909545898,
+ "step": 1623
+ },
+ {
+ "epoch": 22.55846422338569,
+ "grad_norm": 0.05063068866729736,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 4.042117118835449,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57242582897033,
+ "grad_norm": 0.049348752945661545,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 4.012609481811523,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586387434554975,
+ "grad_norm": 0.05123874917626381,
+ "learning_rate": 0.000362575056487632,
+ "loss": 4.03358793258667,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600349040139616,
+ "grad_norm": 0.04921206086874008,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 4.039857864379883,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61431064572426,
+ "grad_norm": 0.05011922866106033,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 4.015835762023926,
+ "step": 1628
+ },
+ {
+ "epoch": 22.6282722513089,
+ "grad_norm": 0.049950532615184784,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 4.068299293518066,
+ "step": 1629
+ },
+ {
+ "epoch": 22.642233856893544,
+ "grad_norm": 0.04962172359228134,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 4.03205680847168,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656195462478184,
+ "grad_norm": 0.05059773847460747,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 4.065027236938477,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670157068062828,
+ "grad_norm": 0.04986768960952759,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 4.01540470123291,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68411867364747,
+ "grad_norm": 0.050884976983070374,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 4.060809135437012,
+ "step": 1633
+ },
+ {
+ "epoch": 22.698080279232112,
+ "grad_norm": 0.052478861063718796,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 4.088409423828125,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712041884816752,
+ "grad_norm": 0.05118529871106148,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 4.071645736694336,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726003490401396,
+ "grad_norm": 0.05059075728058815,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 4.028438568115234,
+ "step": 1636
+ },
+ {
+ "epoch": 22.739965095986037,
+ "grad_norm": 0.049222562462091446,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 4.02166223526001,
+ "step": 1637
+ },
+ {
+ "epoch": 22.75392670157068,
+ "grad_norm": 0.049673549830913544,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 4.0511860847473145,
+ "step": 1638
+ },
+ {
+ "epoch": 22.767888307155324,
+ "grad_norm": 0.04954935610294342,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 4.036063194274902,
+ "step": 1639
+ },
+ {
+ "epoch": 22.781849912739965,
+ "grad_norm": 0.05110877379775047,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 4.081101417541504,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79581151832461,
+ "grad_norm": 0.049009695649147034,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 4.042314529418945,
+ "step": 1641
+ },
+ {
+ "epoch": 22.80977312390925,
+ "grad_norm": 0.0496281161904335,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 4.028315544128418,
+ "step": 1642
+ },
+ {
+ "epoch": 22.823734729493893,
+ "grad_norm": 0.04931195452809334,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 4.018457412719727,
+ "step": 1643
+ },
+ {
+ "epoch": 22.837696335078533,
+ "grad_norm": 0.05006880313158035,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 4.0709991455078125,
+ "step": 1644
+ },
+ {
+ "epoch": 22.851657940663177,
+ "grad_norm": 0.04956946149468422,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 4.065585136413574,
+ "step": 1645
+ },
+ {
+ "epoch": 22.865619546247817,
+ "grad_norm": 0.048182398080825806,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 4.0522589683532715,
+ "step": 1646
+ },
+ {
+ "epoch": 22.87958115183246,
+ "grad_norm": 0.04964807257056236,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 4.06989860534668,
+ "step": 1647
+ },
+ {
+ "epoch": 22.8935427574171,
+ "grad_norm": 0.049299128353595734,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 4.082858085632324,
+ "step": 1648
+ },
+ {
+ "epoch": 22.907504363001745,
+ "grad_norm": 0.04893320053815842,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 4.047144889831543,
+ "step": 1649
+ },
+ {
+ "epoch": 22.921465968586386,
+ "grad_norm": 0.04772866889834404,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 4.043060302734375,
+ "step": 1650
+ },
+ {
+ "epoch": 22.93542757417103,
+ "grad_norm": 0.049417730420827866,
+ "learning_rate": 0.000356031397755195,
+ "loss": 4.031186580657959,
+ "step": 1651
+ },
+ {
+ "epoch": 22.949389179755673,
+ "grad_norm": 0.049944665282964706,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 4.074260711669922,
+ "step": 1652
+ },
+ {
+ "epoch": 22.963350785340314,
+ "grad_norm": 0.04856271669268608,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 4.054103851318359,
+ "step": 1653
+ },
+ {
+ "epoch": 22.977312390924958,
+ "grad_norm": 0.0470513179898262,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 4.041568756103516,
+ "step": 1654
+ },
+ {
+ "epoch": 22.991273996509598,
+ "grad_norm": 0.049018729478120804,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 4.066161632537842,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.034226156771183014,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 2.5147223472595215,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.5970026850700378,
+ "eval_runtime": 60.0787,
+ "eval_samples_per_second": 40.647,
+ "eval_steps_per_second": 0.649,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013961605584644,
+ "grad_norm": 0.05091085657477379,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 3.95741868019104,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027923211169284,
+ "grad_norm": 0.05874285474419594,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 3.966327667236328,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041884816753928,
+ "grad_norm": 0.05394266918301582,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.937232732772827,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05584642233857,
+ "grad_norm": 0.05120540037751198,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 3.9637370109558105,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069808027923212,
+ "grad_norm": 0.05310065299272537,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 3.998631477355957,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083769633507853,
+ "grad_norm": 0.05127466097474098,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.949101448059082,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097731239092496,
+ "grad_norm": 0.05201594531536102,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 3.986267566680908,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111692844677137,
+ "grad_norm": 0.051927369087934494,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 3.9604103565216064,
+ "step": 1664
+ },
+ {
+ "epoch": 23.12565445026178,
+ "grad_norm": 0.05125141143798828,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.9339547157287598,
+ "step": 1665
+ },
+ {
+ "epoch": 23.13961605584642,
+ "grad_norm": 0.05241088196635246,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 3.971261978149414,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153577661431065,
+ "grad_norm": 0.050563592463731766,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.9450082778930664,
+ "step": 1667
+ },
+ {
+ "epoch": 23.167539267015705,
+ "grad_norm": 0.05204763635993004,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 3.9696707725524902,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18150087260035,
+ "grad_norm": 0.05518610402941704,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 3.9816057682037354,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195462478184993,
+ "grad_norm": 0.05336211621761322,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.932252883911133,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209424083769633,
+ "grad_norm": 0.052204642444849014,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 3.9818131923675537,
+ "step": 1671
+ },
+ {
+ "epoch": 23.223385689354277,
+ "grad_norm": 0.05299545079469681,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.9488978385925293,
+ "step": 1672
+ },
+ {
+ "epoch": 23.237347294938917,
+ "grad_norm": 0.05143733322620392,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 4.00959587097168,
+ "step": 1673
+ },
+ {
+ "epoch": 23.25130890052356,
+ "grad_norm": 0.05186324194073677,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.9529542922973633,
+ "step": 1674
+ },
+ {
+ "epoch": 23.2652705061082,
+ "grad_norm": 0.05334410071372986,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 3.96693754196167,
+ "step": 1675
+ },
+ {
+ "epoch": 23.279232111692846,
+ "grad_norm": 0.05345796048641205,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 3.9674174785614014,
+ "step": 1676
+ },
+ {
+ "epoch": 23.293193717277486,
+ "grad_norm": 0.050988901406526566,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 3.9698638916015625,
+ "step": 1677
+ },
+ {
+ "epoch": 23.30715532286213,
+ "grad_norm": 0.056467678397893906,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 3.9645955562591553,
+ "step": 1678
+ },
+ {
+ "epoch": 23.32111692844677,
+ "grad_norm": 0.05554647371172905,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 4.0036091804504395,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335078534031414,
+ "grad_norm": 0.05607818067073822,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 3.973773717880249,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349040139616054,
+ "grad_norm": 0.05734477937221527,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 3.987692356109619,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363001745200698,
+ "grad_norm": 0.052175361663103104,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 3.9761803150177,
+ "step": 1682
+ },
+ {
+ "epoch": 23.376963350785342,
+ "grad_norm": 0.053368814289569855,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 3.965020179748535,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390924956369982,
+ "grad_norm": 0.05505364388227463,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 3.987494468688965,
+ "step": 1684
+ },
+ {
+ "epoch": 23.404886561954626,
+ "grad_norm": 0.0524514764547348,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 3.9624319076538086,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418848167539267,
+ "grad_norm": 0.05681568384170532,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 4.008373737335205,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43280977312391,
+ "grad_norm": 0.05344909429550171,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 3.9840354919433594,
+ "step": 1687
+ },
+ {
+ "epoch": 23.44677137870855,
+ "grad_norm": 0.052657466381788254,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 3.992262363433838,
+ "step": 1688
+ },
+ {
+ "epoch": 23.460732984293195,
+ "grad_norm": 0.05526747554540634,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 3.97529935836792,
+ "step": 1689
+ },
+ {
+ "epoch": 23.474694589877835,
+ "grad_norm": 0.05492551624774933,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 4.004708766937256,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48865619546248,
+ "grad_norm": 0.053742099553346634,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 3.994016647338867,
+ "step": 1691
+ },
+ {
+ "epoch": 23.50261780104712,
+ "grad_norm": 0.05363722890615463,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 3.9942378997802734,
+ "step": 1692
+ },
+ {
+ "epoch": 23.516579406631763,
+ "grad_norm": 0.05167972669005394,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 3.998253345489502,
+ "step": 1693
+ },
+ {
+ "epoch": 23.530541012216403,
+ "grad_norm": 0.0518590584397316,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 3.9850645065307617,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544502617801047,
+ "grad_norm": 0.05521136894822121,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 4.02669620513916,
+ "step": 1695
+ },
+ {
+ "epoch": 23.55846422338569,
+ "grad_norm": 0.054986029863357544,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 4.022308826446533,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57242582897033,
+ "grad_norm": 0.05405082181096077,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 4.021487236022949,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586387434554975,
+ "grad_norm": 0.051257841289043427,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 3.9724135398864746,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600349040139616,
+ "grad_norm": 0.05109119415283203,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 3.978914499282837,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61431064572426,
+ "grad_norm": 0.05303420498967171,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 4.027509689331055,
+ "step": 1700
+ },
+ {
+ "epoch": 23.6282722513089,
+ "grad_norm": 0.055543459951877594,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 3.9750852584838867,
+ "step": 1701
+ },
+ {
+ "epoch": 23.642233856893544,
+ "grad_norm": 0.05068540200591087,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 4.029600143432617,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656195462478184,
+ "grad_norm": 0.04991193488240242,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 3.9901509284973145,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670157068062828,
+ "grad_norm": 0.05403713881969452,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 4.017679214477539,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68411867364747,
+ "grad_norm": 0.05203867331147194,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 4.008406639099121,
+ "step": 1705
+ },
+ {
+ "epoch": 23.698080279232112,
+ "grad_norm": 0.05106597766280174,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 4.009983062744141,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712041884816752,
+ "grad_norm": 0.04976251348853111,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 4.007230281829834,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726003490401396,
+ "grad_norm": 0.05199650675058365,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 3.9902141094207764,
+ "step": 1708
+ },
+ {
+ "epoch": 23.739965095986037,
+ "grad_norm": 0.05285118520259857,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 3.9966492652893066,
+ "step": 1709
+ },
+ {
+ "epoch": 23.75392670157068,
+ "grad_norm": 0.051937371492385864,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 3.962397336959839,
+ "step": 1710
+ },
+ {
+ "epoch": 23.767888307155324,
+ "grad_norm": 0.05179062858223915,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 4.027141571044922,
+ "step": 1711
+ },
+ {
+ "epoch": 23.781849912739965,
+ "grad_norm": 0.049667056649923325,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 4.024991512298584,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79581151832461,
+ "grad_norm": 0.05204658955335617,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 4.002647876739502,
+ "step": 1713
+ },
+ {
+ "epoch": 23.80977312390925,
+ "grad_norm": 0.05069981515407562,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 4.035422325134277,
+ "step": 1714
+ },
+ {
+ "epoch": 23.823734729493893,
+ "grad_norm": 0.05044785887002945,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 4.042465686798096,
+ "step": 1715
+ },
+ {
+ "epoch": 23.837696335078533,
+ "grad_norm": 0.053949885070323944,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 4.002717971801758,
+ "step": 1716
+ },
+ {
+ "epoch": 23.851657940663177,
+ "grad_norm": 0.05139258876442909,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 4.026609897613525,
+ "step": 1717
+ },
+ {
+ "epoch": 23.865619546247817,
+ "grad_norm": 0.05316372960805893,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 4.019758224487305,
+ "step": 1718
+ },
+ {
+ "epoch": 23.87958115183246,
+ "grad_norm": 0.050791382789611816,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 3.9859633445739746,
+ "step": 1719
+ },
+ {
+ "epoch": 23.8935427574171,
+ "grad_norm": 0.05124847590923309,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 4.038055419921875,
+ "step": 1720
+ },
+ {
+ "epoch": 23.907504363001745,
+ "grad_norm": 0.0495096817612648,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 4.02764368057251,
+ "step": 1721
+ },
+ {
+ "epoch": 23.921465968586386,
+ "grad_norm": 0.05205567181110382,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 4.003867149353027,
+ "step": 1722
+ },
+ {
+ "epoch": 23.93542757417103,
+ "grad_norm": 0.05233347788453102,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 4.024157524108887,
+ "step": 1723
+ },
+ {
+ "epoch": 23.949389179755673,
+ "grad_norm": 0.05126349255442619,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 4.05349063873291,
+ "step": 1724
+ },
+ {
+ "epoch": 23.963350785340314,
+ "grad_norm": 0.050006967037916183,
+ "learning_rate": 0.000336518346307424,
+ "loss": 4.024704933166504,
+ "step": 1725
+ },
+ {
+ "epoch": 23.977312390924958,
+ "grad_norm": 0.052343279123306274,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 4.008549213409424,
+ "step": 1726
+ },
+ {
+ "epoch": 23.991273996509598,
+ "grad_norm": 0.05126652494072914,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 4.002162933349609,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.03560106083750725,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 2.511049270629883,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.5985628962516785,
+ "eval_runtime": 59.0712,
+ "eval_samples_per_second": 41.34,
+ "eval_steps_per_second": 0.66,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013961605584644,
+ "grad_norm": 0.05205906555056572,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.936516284942627,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027923211169284,
+ "grad_norm": 0.056368257850408554,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.95200777053833,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041884816753928,
+ "grad_norm": 0.05075085908174515,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.940810441970825,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05584642233857,
+ "grad_norm": 0.05119713768362999,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.90130615234375,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069808027923212,
+ "grad_norm": 0.05254847928881645,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.9491095542907715,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083769633507853,
+ "grad_norm": 0.053025662899017334,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.9371917247772217,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097731239092496,
+ "grad_norm": 0.05402665585279465,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 3.9651317596435547,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111692844677137,
+ "grad_norm": 0.0519571453332901,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.9189062118530273,
+ "step": 1736
+ },
+ {
+ "epoch": 24.12565445026178,
+ "grad_norm": 0.05436019226908684,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.948184013366699,
+ "step": 1737
+ },
+ {
+ "epoch": 24.13961605584642,
+ "grad_norm": 0.056572459638118744,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.9076738357543945,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153577661431065,
+ "grad_norm": 0.05553183704614639,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.933659076690674,
+ "step": 1739
+ },
+ {
+ "epoch": 24.167539267015705,
+ "grad_norm": 0.05554374307394028,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.941751480102539,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18150087260035,
+ "grad_norm": 0.05451614409685135,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.928593635559082,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195462478184993,
+ "grad_norm": 0.055458780378103256,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 3.961820602416992,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209424083769633,
+ "grad_norm": 0.054980456829071045,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 3.9736528396606445,
+ "step": 1743
+ },
+ {
+ "epoch": 24.223385689354277,
+ "grad_norm": 0.05538604408502579,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.935093641281128,
+ "step": 1744
+ },
+ {
+ "epoch": 24.237347294938917,
+ "grad_norm": 0.05580678582191467,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.9240689277648926,
+ "step": 1745
+ },
+ {
+ "epoch": 24.25130890052356,
+ "grad_norm": 0.052294690161943436,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.937718152999878,
+ "step": 1746
+ },
+ {
+ "epoch": 24.2652705061082,
+ "grad_norm": 0.052018795162439346,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.942049503326416,
+ "step": 1747
+ },
+ {
+ "epoch": 24.279232111692846,
+ "grad_norm": 0.0550091527402401,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.9215564727783203,
+ "step": 1748
+ },
+ {
+ "epoch": 24.293193717277486,
+ "grad_norm": 0.05447716638445854,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.9154953956604004,
+ "step": 1749
+ },
+ {
+ "epoch": 24.30715532286213,
+ "grad_norm": 0.05300029739737511,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 3.9642622470855713,
+ "step": 1750
+ },
+ {
+ "epoch": 24.32111692844677,
+ "grad_norm": 0.05312013998627663,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.938897132873535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335078534031414,
+ "grad_norm": 0.053737837821245193,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.94972562789917,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349040139616054,
+ "grad_norm": 0.05087469890713692,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.9301164150238037,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363001745200698,
+ "grad_norm": 0.05179893225431442,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.9393205642700195,
+ "step": 1754
+ },
+ {
+ "epoch": 24.376963350785342,
+ "grad_norm": 0.05277438461780548,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 3.960968494415283,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390924956369982,
+ "grad_norm": 0.05090474337339401,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 3.9516637325286865,
+ "step": 1756
+ },
+ {
+ "epoch": 24.404886561954626,
+ "grad_norm": 0.052786506712436676,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.9332518577575684,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418848167539267,
+ "grad_norm": 0.052095942199230194,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.9108662605285645,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43280977312391,
+ "grad_norm": 0.05310450494289398,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 3.9746041297912598,
+ "step": 1759
+ },
+ {
+ "epoch": 24.44677137870855,
+ "grad_norm": 0.05078566074371338,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.93316650390625,
+ "step": 1760
+ },
+ {
+ "epoch": 24.460732984293195,
+ "grad_norm": 0.0519949309527874,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.9489870071411133,
+ "step": 1761
+ },
+ {
+ "epoch": 24.474694589877835,
+ "grad_norm": 0.05239349976181984,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 3.955118417739868,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48865619546248,
+ "grad_norm": 0.05193360522389412,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 3.9707555770874023,
+ "step": 1763
+ },
+ {
+ "epoch": 24.50261780104712,
+ "grad_norm": 0.05397362262010574,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 3.973053455352783,
+ "step": 1764
+ },
+ {
+ "epoch": 24.516579406631763,
+ "grad_norm": 0.05379117652773857,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 3.957289695739746,
+ "step": 1765
+ },
+ {
+ "epoch": 24.530541012216403,
+ "grad_norm": 0.0526256300508976,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 3.9595577716827393,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544502617801047,
+ "grad_norm": 0.05377693101763725,
+ "learning_rate": 0.000325372061241796,
+ "loss": 3.97849178314209,
+ "step": 1767
+ },
+ {
+ "epoch": 24.55846422338569,
+ "grad_norm": 0.05182809755206108,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 3.9714043140411377,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57242582897033,
+ "grad_norm": 0.05364793911576271,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 3.9582200050354004,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586387434554975,
+ "grad_norm": 0.051558300852775574,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 3.9747936725616455,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600349040139616,
+ "grad_norm": 0.053487639874219894,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 3.9556500911712646,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61431064572426,
+ "grad_norm": 0.05146970972418785,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.94858455657959,
+ "step": 1772
+ },
+ {
+ "epoch": 24.6282722513089,
+ "grad_norm": 0.05099836364388466,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 3.9716954231262207,
+ "step": 1773
+ },
+ {
+ "epoch": 24.642233856893544,
+ "grad_norm": 0.054634734988212585,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 3.974881887435913,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656195462478184,
+ "grad_norm": 0.054601747542619705,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 3.9746241569519043,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670157068062828,
+ "grad_norm": 0.052188221365213394,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 3.957106113433838,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68411867364747,
+ "grad_norm": 0.055673446506261826,
+ "learning_rate": 0.000322712903929477,
+ "loss": 3.9803967475891113,
+ "step": 1777
+ },
+ {
+ "epoch": 24.698080279232112,
+ "grad_norm": 0.05397328734397888,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 3.967238426208496,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712041884816752,
+ "grad_norm": 0.05384642630815506,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 3.949392795562744,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726003490401396,
+ "grad_norm": 0.05452995374798775,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.9432482719421387,
+ "step": 1780
+ },
+ {
+ "epoch": 24.739965095986037,
+ "grad_norm": 0.052430473268032074,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 3.9944653511047363,
+ "step": 1781
+ },
+ {
+ "epoch": 24.75392670157068,
+ "grad_norm": 0.05514240264892578,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 3.97713565826416,
+ "step": 1782
+ },
+ {
+ "epoch": 24.767888307155324,
+ "grad_norm": 0.05189919099211693,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 3.983150005340576,
+ "step": 1783
+ },
+ {
+ "epoch": 24.781849912739965,
+ "grad_norm": 0.05590776726603508,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 3.9548301696777344,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79581151832461,
+ "grad_norm": 0.05564190074801445,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 3.9725170135498047,
+ "step": 1785
+ },
+ {
+ "epoch": 24.80977312390925,
+ "grad_norm": 0.05121326446533203,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 3.9722347259521484,
+ "step": 1786
+ },
+ {
+ "epoch": 24.823734729493893,
+ "grad_norm": 0.05444202199578285,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 3.9634029865264893,
+ "step": 1787
+ },
+ {
+ "epoch": 24.837696335078533,
+ "grad_norm": 0.05242442339658737,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 3.973270893096924,
+ "step": 1788
+ },
+ {
+ "epoch": 24.851657940663177,
+ "grad_norm": 0.05497463420033455,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 3.9851887226104736,
+ "step": 1789
+ },
+ {
+ "epoch": 24.865619546247817,
+ "grad_norm": 0.05482025071978569,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 4.001110076904297,
+ "step": 1790
+ },
+ {
+ "epoch": 24.87958115183246,
+ "grad_norm": 0.054101862013339996,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 3.9782891273498535,
+ "step": 1791
+ },
+ {
+ "epoch": 24.8935427574171,
+ "grad_norm": 0.05538944527506828,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 3.973541498184204,
+ "step": 1792
+ },
+ {
+ "epoch": 24.907504363001745,
+ "grad_norm": 0.0502893440425396,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 3.9875426292419434,
+ "step": 1793
+ },
+ {
+ "epoch": 24.921465968586386,
+ "grad_norm": 0.05576152354478836,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.947035551071167,
+ "step": 1794
+ },
+ {
+ "epoch": 24.93542757417103,
+ "grad_norm": 0.05356278270483017,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 3.9683666229248047,
+ "step": 1795
+ },
+ {
+ "epoch": 24.949389179755673,
+ "grad_norm": 0.05377693101763725,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 3.9973440170288086,
+ "step": 1796
+ },
+ {
+ "epoch": 24.963350785340314,
+ "grad_norm": 0.05334382876753807,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 3.9843404293060303,
+ "step": 1797
+ },
+ {
+ "epoch": 24.977312390924958,
+ "grad_norm": 0.05514024198055267,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 3.9920568466186523,
+ "step": 1798
+ },
+ {
+ "epoch": 24.991273996509598,
+ "grad_norm": 0.05631054937839508,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 4.005859375,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.03653077408671379,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 2.503864049911499,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 0.5995591878890991,
+ "eval_runtime": 59.5213,
+ "eval_samples_per_second": 41.027,
+ "eval_steps_per_second": 0.655,
+ "step": 1800
+ },
+ {
+ "epoch": 25.013961605584644,
+ "grad_norm": 0.0576021745800972,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 3.894834518432617,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027923211169284,
+ "grad_norm": 0.05842140316963196,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 3.8879270553588867,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041884816753928,
+ "grad_norm": 0.051821205765008926,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 3.9011340141296387,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05584642233857,
+ "grad_norm": 0.053872160613536835,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 3.877183437347412,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069808027923212,
+ "grad_norm": 0.05487574636936188,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 3.9126815795898438,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083769633507853,
+ "grad_norm": 0.054469428956508636,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 3.894768238067627,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097731239092496,
+ "grad_norm": 0.05124446377158165,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 3.916809320449829,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111692844677137,
+ "grad_norm": 0.05193263292312622,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 3.917140483856201,
+ "step": 1808
+ },
+ {
+ "epoch": 25.12565445026178,
+ "grad_norm": 0.0519125871360302,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 3.903704881668091,
+ "step": 1809
+ },
+ {
+ "epoch": 25.13961605584642,
+ "grad_norm": 0.052205223590135574,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 3.892639636993408,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153577661431065,
+ "grad_norm": 0.052036914974451065,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 3.895559787750244,
+ "step": 1811
+ },
+ {
+ "epoch": 25.167539267015705,
+ "grad_norm": 0.054616156965494156,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 3.8976142406463623,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18150087260035,
+ "grad_norm": 0.05603514611721039,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 3.914311170578003,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195462478184993,
+ "grad_norm": 0.05170242115855217,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 3.9141643047332764,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209424083769633,
+ "grad_norm": 0.05476640164852142,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 3.8952436447143555,
+ "step": 1815
+ },
+ {
+ "epoch": 25.223385689354277,
+ "grad_norm": 0.055114105343818665,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 3.872978687286377,
+ "step": 1816
+ },
+ {
+ "epoch": 25.237347294938917,
+ "grad_norm": 0.051593828946352005,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 3.893911600112915,
+ "step": 1817
+ },
+ {
+ "epoch": 25.25130890052356,
+ "grad_norm": 0.05363611876964569,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 3.9062986373901367,
+ "step": 1818
+ },
+ {
+ "epoch": 25.2652705061082,
+ "grad_norm": 0.054656051099300385,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 3.9145383834838867,
+ "step": 1819
+ },
+ {
+ "epoch": 25.279232111692846,
+ "grad_norm": 0.05251947045326233,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 3.912965774536133,
+ "step": 1820
+ },
+ {
+ "epoch": 25.293193717277486,
+ "grad_norm": 0.05575991049408913,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 3.901510238647461,
+ "step": 1821
+ },
+ {
+ "epoch": 25.30715532286213,
+ "grad_norm": 0.05540676787495613,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 3.8659887313842773,
+ "step": 1822
+ },
+ {
+ "epoch": 25.32111692844677,
+ "grad_norm": 0.0577131025493145,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 3.884077548980713,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335078534031414,
+ "grad_norm": 0.055947501212358475,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 3.9274377822875977,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349040139616054,
+ "grad_norm": 0.05328478664159775,
+ "learning_rate": 0.000309930621834911,
+ "loss": 3.8783111572265625,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363001745200698,
+ "grad_norm": 0.057696059346199036,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 3.903721570968628,
+ "step": 1826
+ },
+ {
+ "epoch": 25.376963350785342,
+ "grad_norm": 0.05717543140053749,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 3.9259400367736816,
+ "step": 1827
+ },
+ {
+ "epoch": 25.390924956369982,
+ "grad_norm": 0.05328287556767464,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 3.9450511932373047,
+ "step": 1828
+ },
+ {
+ "epoch": 25.404886561954626,
+ "grad_norm": 0.05414140224456787,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 3.9231765270233154,
+ "step": 1829
+ },
+ {
+ "epoch": 25.418848167539267,
+ "grad_norm": 0.05535200983285904,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 3.8928630352020264,
+ "step": 1830
+ },
+ {
+ "epoch": 25.43280977312391,
+ "grad_norm": 0.05468215420842171,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 3.933401584625244,
+ "step": 1831
+ },
+ {
+ "epoch": 25.44677137870855,
+ "grad_norm": 0.056624677032232285,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 3.922624111175537,
+ "step": 1832
+ },
+ {
+ "epoch": 25.460732984293195,
+ "grad_norm": 0.0545354001224041,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 3.916515827178955,
+ "step": 1833
+ },
+ {
+ "epoch": 25.474694589877835,
+ "grad_norm": 0.053979165852069855,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 3.891812324523926,
+ "step": 1834
+ },
+ {
+ "epoch": 25.48865619546248,
+ "grad_norm": 0.05619875714182854,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 3.959524631500244,
+ "step": 1835
+ },
+ {
+ "epoch": 25.50261780104712,
+ "grad_norm": 0.053561773151159286,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 3.9109320640563965,
+ "step": 1836
+ },
+ {
+ "epoch": 25.516579406631763,
+ "grad_norm": 0.05566735938191414,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 3.910492420196533,
+ "step": 1837
+ },
+ {
+ "epoch": 25.530541012216403,
+ "grad_norm": 0.05757030472159386,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 3.9716763496398926,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544502617801047,
+ "grad_norm": 0.05121897906064987,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 3.9114646911621094,
+ "step": 1839
+ },
+ {
+ "epoch": 25.55846422338569,
+ "grad_norm": 0.05664774775505066,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 3.9465560913085938,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57242582897033,
+ "grad_norm": 0.0569215826690197,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 3.908374309539795,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586387434554975,
+ "grad_norm": 0.05345458909869194,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 3.9336414337158203,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600349040139616,
+ "grad_norm": 0.0565933920443058,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 3.9478759765625,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61431064572426,
+ "grad_norm": 0.05887870490550995,
+ "learning_rate": 0.000304866093757771,
+ "loss": 3.9018349647521973,
+ "step": 1844
+ },
+ {
+ "epoch": 25.6282722513089,
+ "grad_norm": 0.05515308678150177,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 3.9446825981140137,
+ "step": 1845
+ },
+ {
+ "epoch": 25.642233856893544,
+ "grad_norm": 0.058602213859558105,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 3.9497885704040527,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656195462478184,
+ "grad_norm": 0.053806282579898834,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 3.9247193336486816,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670157068062828,
+ "grad_norm": 0.05495952442288399,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 3.912296772003174,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68411867364747,
+ "grad_norm": 0.05479473993182182,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 3.9471306800842285,
+ "step": 1849
+ },
+ {
+ "epoch": 25.698080279232112,
+ "grad_norm": 0.053319111466407776,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 3.955942153930664,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712041884816752,
+ "grad_norm": 0.05832696333527565,
+ "learning_rate": 0.000303,
+ "loss": 3.959169626235962,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726003490401396,
+ "grad_norm": 0.057019371539354324,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 3.9682188034057617,
+ "step": 1852
+ },
+ {
+ "epoch": 25.739965095986037,
+ "grad_norm": 0.05099530518054962,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 3.921377658843994,
+ "step": 1853
+ },
+ {
+ "epoch": 25.75392670157068,
+ "grad_norm": 0.05466439574956894,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 3.946341037750244,
+ "step": 1854
+ },
+ {
+ "epoch": 25.767888307155324,
+ "grad_norm": 0.05742308124899864,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 3.948031425476074,
+ "step": 1855
+ },
+ {
+ "epoch": 25.781849912739965,
+ "grad_norm": 0.057581741362810135,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 3.91162109375,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79581151832461,
+ "grad_norm": 0.05520833283662796,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 3.937495231628418,
+ "step": 1857
+ },
+ {
+ "epoch": 25.80977312390925,
+ "grad_norm": 0.05763327330350876,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 3.9565534591674805,
+ "step": 1858
+ },
+ {
+ "epoch": 25.823734729493893,
+ "grad_norm": 0.0544707253575325,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 3.965190887451172,
+ "step": 1859
+ },
+ {
+ "epoch": 25.837696335078533,
+ "grad_norm": 0.05445104092359543,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 3.938063859939575,
+ "step": 1860
+ },
+ {
+ "epoch": 25.851657940663177,
+ "grad_norm": 0.05505113676190376,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 3.928392171859741,
+ "step": 1861
+ },
+ {
+ "epoch": 25.865619546247817,
+ "grad_norm": 0.05574212223291397,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 3.959397077560425,
+ "step": 1862
+ },
+ {
+ "epoch": 25.87958115183246,
+ "grad_norm": 0.053950466215610504,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 3.9532437324523926,
+ "step": 1863
+ },
+ {
+ "epoch": 25.8935427574171,
+ "grad_norm": 0.05495546758174896,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 3.946746826171875,
+ "step": 1864
+ },
+ {
+ "epoch": 25.907504363001745,
+ "grad_norm": 0.05328625068068504,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 3.9494776725769043,
+ "step": 1865
+ },
+ {
+ "epoch": 25.921465968586386,
+ "grad_norm": 0.05593104660511017,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 3.959989547729492,
+ "step": 1866
+ },
+ {
+ "epoch": 25.93542757417103,
+ "grad_norm": 0.05442969873547554,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 3.952371597290039,
+ "step": 1867
+ },
+ {
+ "epoch": 25.949389179755673,
+ "grad_norm": 0.05446415767073631,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 3.9538888931274414,
+ "step": 1868
+ },
+ {
+ "epoch": 25.963350785340314,
+ "grad_norm": 0.054761797189712524,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 3.9386215209960938,
+ "step": 1869
+ },
+ {
+ "epoch": 25.977312390924958,
+ "grad_norm": 0.054201845079660416,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 3.9252963066101074,
+ "step": 1870
+ },
+ {
+ "epoch": 25.991273996509598,
+ "grad_norm": 0.05364517495036125,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 3.9733197689056396,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.038914378732442856,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 2.4261250495910645,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 0.6009622812271118,
+ "eval_runtime": 60.6864,
+ "eval_samples_per_second": 40.24,
+ "eval_steps_per_second": 0.643,
+ "step": 1872
+ },
+ {
+ "epoch": 26.013961605584644,
+ "grad_norm": 0.05345558747649193,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 3.8902673721313477,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027923211169284,
+ "grad_norm": 0.05672457069158554,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 3.8676247596740723,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041884816753928,
+ "grad_norm": 0.051981374621391296,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 3.872048854827881,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05584642233857,
+ "grad_norm": 0.05262405797839165,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 3.8582329750061035,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069808027923212,
+ "grad_norm": 0.054473262280225754,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 3.8352112770080566,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083769633507853,
+ "grad_norm": 0.053167302161455154,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 3.881760835647583,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097731239092496,
+ "grad_norm": 0.053720805794000626,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 3.891775608062744,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111692844677137,
+ "grad_norm": 0.055102936923503876,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 3.8872156143188477,
+ "step": 1880
+ },
+ {
+ "epoch": 26.12565445026178,
+ "grad_norm": 0.05108853057026863,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 3.8625407218933105,
+ "step": 1881
+ },
+ {
+ "epoch": 26.13961605584642,
+ "grad_norm": 0.05526706948876381,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 3.886127233505249,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153577661431065,
+ "grad_norm": 0.05458226427435875,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 3.8722105026245117,
+ "step": 1883
+ },
+ {
+ "epoch": 26.167539267015705,
+ "grad_norm": 0.05342113599181175,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 3.853677749633789,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18150087260035,
+ "grad_norm": 0.053665641695261,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 3.849315881729126,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195462478184993,
+ "grad_norm": 0.054922688752412796,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 3.8599777221679688,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209424083769633,
+ "grad_norm": 0.05412537604570389,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 3.8767571449279785,
+ "step": 1887
+ },
+ {
+ "epoch": 26.223385689354277,
+ "grad_norm": 0.05467838793992996,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 3.8436965942382812,
+ "step": 1888
+ },
+ {
+ "epoch": 26.237347294938917,
+ "grad_norm": 0.054886411875486374,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 3.8796300888061523,
+ "step": 1889
+ },
+ {
+ "epoch": 26.25130890052356,
+ "grad_norm": 0.05437029153108597,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 3.858133316040039,
+ "step": 1890
+ },
+ {
+ "epoch": 26.2652705061082,
+ "grad_norm": 0.05273836478590965,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 3.839015483856201,
+ "step": 1891
+ },
+ {
+ "epoch": 26.279232111692846,
+ "grad_norm": 0.0546451173722744,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 3.8723373413085938,
+ "step": 1892
+ },
+ {
+ "epoch": 26.293193717277486,
+ "grad_norm": 0.05240490287542343,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 3.877549886703491,
+ "step": 1893
+ },
+ {
+ "epoch": 26.30715532286213,
+ "grad_norm": 0.05420667678117752,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 3.8643171787261963,
+ "step": 1894
+ },
+ {
+ "epoch": 26.32111692844677,
+ "grad_norm": 0.053440142422914505,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 3.8681447505950928,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335078534031414,
+ "grad_norm": 0.052396632730960846,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 3.8691887855529785,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349040139616054,
+ "grad_norm": 0.05318925902247429,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 3.897003650665283,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363001745200698,
+ "grad_norm": 0.05328533053398132,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 3.9079437255859375,
+ "step": 1898
+ },
+ {
+ "epoch": 26.376963350785342,
+ "grad_norm": 0.05196928232908249,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 3.8392016887664795,
+ "step": 1899
+ },
+ {
+ "epoch": 26.390924956369982,
+ "grad_norm": 0.053326766937971115,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 3.8930301666259766,
+ "step": 1900
+ },
+ {
+ "epoch": 26.404886561954626,
+ "grad_norm": 0.052926234900951385,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 3.866997718811035,
+ "step": 1901
+ },
+ {
+ "epoch": 26.418848167539267,
+ "grad_norm": 0.05289287120103836,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 3.896690607070923,
+ "step": 1902
+ },
+ {
+ "epoch": 26.43280977312391,
+ "grad_norm": 0.052162300795316696,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 3.8642401695251465,
+ "step": 1903
+ },
+ {
+ "epoch": 26.44677137870855,
+ "grad_norm": 0.05495629832148552,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 3.8982696533203125,
+ "step": 1904
+ },
+ {
+ "epoch": 26.460732984293195,
+ "grad_norm": 0.055102866142988205,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 3.8725972175598145,
+ "step": 1905
+ },
+ {
+ "epoch": 26.474694589877835,
+ "grad_norm": 0.05537770688533783,
+ "learning_rate": 0.000288343693342466,
+ "loss": 3.9050111770629883,
+ "step": 1906
+ },
+ {
+ "epoch": 26.48865619546248,
+ "grad_norm": 0.05490661785006523,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 3.90586256980896,
+ "step": 1907
+ },
+ {
+ "epoch": 26.50261780104712,
+ "grad_norm": 0.054393451660871506,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 3.8841705322265625,
+ "step": 1908
+ },
+ {
+ "epoch": 26.516579406631763,
+ "grad_norm": 0.0553058385848999,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 3.883500576019287,
+ "step": 1909
+ },
+ {
+ "epoch": 26.530541012216403,
+ "grad_norm": 0.053976885974407196,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 3.894948959350586,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544502617801047,
+ "grad_norm": 0.05290692299604416,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 3.8880579471588135,
+ "step": 1911
+ },
+ {
+ "epoch": 26.55846422338569,
+ "grad_norm": 0.05570613220334053,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 3.9035487174987793,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57242582897033,
+ "grad_norm": 0.054136309772729874,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 3.9137563705444336,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586387434554975,
+ "grad_norm": 0.05328850820660591,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 3.87838077545166,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600349040139616,
+ "grad_norm": 0.05169905349612236,
+ "learning_rate": 0.000285947841605349,
+ "loss": 3.882963180541992,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61431064572426,
+ "grad_norm": 0.05304642394185066,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 3.9250073432922363,
+ "step": 1916
+ },
+ {
+ "epoch": 26.6282722513089,
+ "grad_norm": 0.051766350865364075,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 3.8915140628814697,
+ "step": 1917
+ },
+ {
+ "epoch": 26.642233856893544,
+ "grad_norm": 0.054485417902469635,
+ "learning_rate": 0.000285149463934261,
+ "loss": 3.8973851203918457,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656195462478184,
+ "grad_norm": 0.05437815189361572,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 3.9054274559020996,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670157068062828,
+ "grad_norm": 0.05402572825551033,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 3.9060683250427246,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68411867364747,
+ "grad_norm": 0.05332810431718826,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 3.9053144454956055,
+ "step": 1921
+ },
+ {
+ "epoch": 26.698080279232112,
+ "grad_norm": 0.0557575598359108,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 3.9026694297790527,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712041884816752,
+ "grad_norm": 0.05650033429265022,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 3.8795413970947266,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726003490401396,
+ "grad_norm": 0.054456304758787155,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 3.932950496673584,
+ "step": 1924
+ },
+ {
+ "epoch": 26.739965095986037,
+ "grad_norm": 0.05676668882369995,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 3.878704071044922,
+ "step": 1925
+ },
+ {
+ "epoch": 26.75392670157068,
+ "grad_norm": 0.05875959247350693,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 3.920717239379883,
+ "step": 1926
+ },
+ {
+ "epoch": 26.767888307155324,
+ "grad_norm": 0.054490379989147186,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 3.912592887878418,
+ "step": 1927
+ },
+ {
+ "epoch": 26.781849912739965,
+ "grad_norm": 0.055268481373786926,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 3.900418758392334,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79581151832461,
+ "grad_norm": 0.05617086961865425,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 3.899095296859741,
+ "step": 1929
+ },
+ {
+ "epoch": 26.80977312390925,
+ "grad_norm": 0.05310523882508278,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 3.917501449584961,
+ "step": 1930
+ },
+ {
+ "epoch": 26.823734729493893,
+ "grad_norm": 0.05476578325033188,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 3.867847442626953,
+ "step": 1931
+ },
+ {
+ "epoch": 26.837696335078533,
+ "grad_norm": 0.05456431582570076,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 3.8760366439819336,
+ "step": 1932
+ },
+ {
+ "epoch": 26.851657940663177,
+ "grad_norm": 0.055269334465265274,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 3.910797119140625,
+ "step": 1933
+ },
+ {
+ "epoch": 26.865619546247817,
+ "grad_norm": 0.054999858140945435,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 3.946951389312744,
+ "step": 1934
+ },
+ {
+ "epoch": 26.87958115183246,
+ "grad_norm": 0.05520344525575638,
+ "learning_rate": 0.000280627938758204,
+ "loss": 3.9161229133605957,
+ "step": 1935
+ },
+ {
+ "epoch": 26.8935427574171,
+ "grad_norm": 0.05451727285981178,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 3.908015727996826,
+ "step": 1936
+ },
+ {
+ "epoch": 26.907504363001745,
+ "grad_norm": 0.05561735853552818,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 3.944608449935913,
+ "step": 1937
+ },
+ {
+ "epoch": 26.921465968586386,
+ "grad_norm": 0.054743457585573196,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 3.906559467315674,
+ "step": 1938
+ },
+ {
+ "epoch": 26.93542757417103,
+ "grad_norm": 0.05497661978006363,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 3.900981903076172,
+ "step": 1939
+ },
+ {
+ "epoch": 26.949389179755673,
+ "grad_norm": 0.05452290549874306,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 3.910825252532959,
+ "step": 1940
+ },
+ {
+ "epoch": 26.963350785340314,
+ "grad_norm": 0.05409001186490059,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 3.921509265899658,
+ "step": 1941
+ },
+ {
+ "epoch": 26.977312390924958,
+ "grad_norm": 0.05467002093791962,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 3.9039864540100098,
+ "step": 1942
+ },
+ {
+ "epoch": 26.991273996509598,
+ "grad_norm": 0.05442725867033005,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 3.9153316020965576,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.039271868765354156,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 2.451029062271118,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 0.6021912097930908,
+ "eval_runtime": 61.3232,
+ "eval_samples_per_second": 39.822,
+ "eval_steps_per_second": 0.636,
+ "step": 1944
+ },
+ {
+ "epoch": 27.013961605584644,
+ "grad_norm": 0.05251244455575943,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 3.8143270015716553,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027923211169284,
+ "grad_norm": 0.056699808686971664,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 3.8331453800201416,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041884816753928,
+ "grad_norm": 0.055446673184633255,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 3.833841323852539,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05584642233857,
+ "grad_norm": 0.053186994045972824,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 3.82643985748291,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069808027923212,
+ "grad_norm": 0.05483868718147278,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 3.8209316730499268,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083769633507853,
+ "grad_norm": 0.05387664586305618,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 3.8402364253997803,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097731239092496,
+ "grad_norm": 0.05422234162688255,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 3.8125908374786377,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111692844677137,
+ "grad_norm": 0.05522442236542702,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 3.830599784851074,
+ "step": 1952
+ },
+ {
+ "epoch": 27.12565445026178,
+ "grad_norm": 0.05163128301501274,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 3.833247184753418,
+ "step": 1953
+ },
+ {
+ "epoch": 27.13961605584642,
+ "grad_norm": 0.053523529320955276,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 3.812887191772461,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153577661431065,
+ "grad_norm": 0.05362863466143608,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 3.8325610160827637,
+ "step": 1955
+ },
+ {
+ "epoch": 27.167539267015705,
+ "grad_norm": 0.054255858063697815,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 3.807152271270752,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18150087260035,
+ "grad_norm": 0.05259740352630615,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 3.843714475631714,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195462478184993,
+ "grad_norm": 0.05457160621881485,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 3.847249984741211,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209424083769633,
+ "grad_norm": 0.05517619848251343,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 3.864114284515381,
+ "step": 1959
+ },
+ {
+ "epoch": 27.223385689354277,
+ "grad_norm": 0.052946750074625015,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 3.8219656944274902,
+ "step": 1960
+ },
+ {
+ "epoch": 27.237347294938917,
+ "grad_norm": 0.053545452654361725,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 3.8502726554870605,
+ "step": 1961
+ },
+ {
+ "epoch": 27.25130890052356,
+ "grad_norm": 0.054657287895679474,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 3.837273120880127,
+ "step": 1962
+ },
+ {
+ "epoch": 27.2652705061082,
+ "grad_norm": 0.05385086312890053,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 3.837573528289795,
+ "step": 1963
+ },
+ {
+ "epoch": 27.279232111692846,
+ "grad_norm": 0.05512048304080963,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 3.842432737350464,
+ "step": 1964
+ },
+ {
+ "epoch": 27.293193717277486,
+ "grad_norm": 0.05286296457052231,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 3.845123767852783,
+ "step": 1965
+ },
+ {
+ "epoch": 27.30715532286213,
+ "grad_norm": 0.05396866053342819,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 3.836874485015869,
+ "step": 1966
+ },
+ {
+ "epoch": 27.32111692844677,
+ "grad_norm": 0.053341615945100784,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 3.8860960006713867,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335078534031414,
+ "grad_norm": 0.0531761609017849,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 3.828686237335205,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349040139616054,
+ "grad_norm": 0.054875023663043976,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 3.83782696723938,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363001745200698,
+ "grad_norm": 0.057710494846105576,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 3.8841819763183594,
+ "step": 1970
+ },
+ {
+ "epoch": 27.376963350785342,
+ "grad_norm": 0.05443769320845604,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 3.856564521789551,
+ "step": 1971
+ },
+ {
+ "epoch": 27.390924956369982,
+ "grad_norm": 0.05571551248431206,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 3.847541570663452,
+ "step": 1972
+ },
+ {
+ "epoch": 27.404886561954626,
+ "grad_norm": 0.05770037695765495,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 3.843390941619873,
+ "step": 1973
+ },
+ {
+ "epoch": 27.418848167539267,
+ "grad_norm": 0.05414504557847977,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 3.8786306381225586,
+ "step": 1974
+ },
+ {
+ "epoch": 27.43280977312391,
+ "grad_norm": 0.058058444410562515,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 3.8680999279022217,
+ "step": 1975
+ },
+ {
+ "epoch": 27.44677137870855,
+ "grad_norm": 0.056312717497348785,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 3.858879566192627,
+ "step": 1976
+ },
+ {
+ "epoch": 27.460732984293195,
+ "grad_norm": 0.055812690407037735,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 3.870741844177246,
+ "step": 1977
+ },
+ {
+ "epoch": 27.474694589877835,
+ "grad_norm": 0.057005349546670914,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 3.8850440979003906,
+ "step": 1978
+ },
+ {
+ "epoch": 27.48865619546248,
+ "grad_norm": 0.055214181542396545,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 3.8485918045043945,
+ "step": 1979
+ },
+ {
+ "epoch": 27.50261780104712,
+ "grad_norm": 0.054300058633089066,
+ "learning_rate": 0.00026868712586269,
+ "loss": 3.8759729862213135,
+ "step": 1980
+ },
+ {
+ "epoch": 27.516579406631763,
+ "grad_norm": 0.05370737239718437,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 3.8498623371124268,
+ "step": 1981
+ },
+ {
+ "epoch": 27.530541012216403,
+ "grad_norm": 0.05481775104999542,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 3.8408432006835938,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544502617801047,
+ "grad_norm": 0.05360998213291168,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 3.8800129890441895,
+ "step": 1983
+ },
+ {
+ "epoch": 27.55846422338569,
+ "grad_norm": 0.05510537698864937,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 3.8329358100891113,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57242582897033,
+ "grad_norm": 0.057950302958488464,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 3.870562791824341,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586387434554975,
+ "grad_norm": 0.0546487458050251,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 3.85231351852417,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600349040139616,
+ "grad_norm": 0.056359753012657166,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 3.8782615661621094,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61431064572426,
+ "grad_norm": 0.05409953370690346,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 3.8209736347198486,
+ "step": 1988
+ },
+ {
+ "epoch": 27.6282722513089,
+ "grad_norm": 0.05549091845750809,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 3.9041504859924316,
+ "step": 1989
+ },
+ {
+ "epoch": 27.642233856893544,
+ "grad_norm": 0.05437365546822548,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 3.8530654907226562,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656195462478184,
+ "grad_norm": 0.055456943809986115,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 3.862886428833008,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670157068062828,
+ "grad_norm": 0.05614335089921951,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 3.867595672607422,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68411867364747,
+ "grad_norm": 0.05374857038259506,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 3.8731391429901123,
+ "step": 1993
+ },
+ {
+ "epoch": 27.698080279232112,
+ "grad_norm": 0.05544714257121086,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 3.837608575820923,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712041884816752,
+ "grad_norm": 0.05495234578847885,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 3.8587915897369385,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726003490401396,
+ "grad_norm": 0.05354582890868187,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 3.8618052005767822,
+ "step": 1996
+ },
+ {
+ "epoch": 27.739965095986037,
+ "grad_norm": 0.05556262657046318,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 3.8630597591400146,
+ "step": 1997
+ },
+ {
+ "epoch": 27.75392670157068,
+ "grad_norm": 0.05576452240347862,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 3.8570914268493652,
+ "step": 1998
+ },
+ {
+ "epoch": 27.767888307155324,
+ "grad_norm": 0.053394027054309845,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 3.868783950805664,
+ "step": 1999
+ },
+ {
+ "epoch": 27.781849912739965,
+ "grad_norm": 0.05506644397974014,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 3.871555805206299,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79581151832461,
+ "grad_norm": 0.054367583245038986,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 3.8487513065338135,
+ "step": 2001
+ },
+ {
+ "epoch": 27.80977312390925,
+ "grad_norm": 0.05482787266373634,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 3.88199520111084,
+ "step": 2002
+ },
+ {
+ "epoch": 27.823734729493893,
+ "grad_norm": 0.05416601523756981,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 3.8876938819885254,
+ "step": 2003
+ },
+ {
+ "epoch": 27.837696335078533,
+ "grad_norm": 0.05612047761678696,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 3.8772940635681152,
+ "step": 2004
+ },
+ {
+ "epoch": 27.851657940663177,
+ "grad_norm": 0.05662519112229347,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 3.8429012298583984,
+ "step": 2005
+ },
+ {
+ "epoch": 27.865619546247817,
+ "grad_norm": 0.055557794868946075,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 3.8712010383605957,
+ "step": 2006
+ },
+ {
+ "epoch": 27.87958115183246,
+ "grad_norm": 0.054365914314985275,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 3.884854793548584,
+ "step": 2007
+ },
+ {
+ "epoch": 27.8935427574171,
+ "grad_norm": 0.05622334033250809,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 3.853909492492676,
+ "step": 2008
+ },
+ {
+ "epoch": 27.907504363001745,
+ "grad_norm": 0.05513857677578926,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 3.888460159301758,
+ "step": 2009
+ },
+ {
+ "epoch": 27.921465968586386,
+ "grad_norm": 0.05537945404648781,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 3.891343593597412,
+ "step": 2010
+ },
+ {
+ "epoch": 27.93542757417103,
+ "grad_norm": 0.05649508908390999,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 3.8807168006896973,
+ "step": 2011
+ },
+ {
+ "epoch": 27.949389179755673,
+ "grad_norm": 0.054261550307273865,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 3.8827662467956543,
+ "step": 2012
+ },
+ {
+ "epoch": 27.963350785340314,
+ "grad_norm": 0.05615224689245224,
+ "learning_rate": 0.00025996500713765,
+ "loss": 3.8720929622650146,
+ "step": 2013
+ },
+ {
+ "epoch": 27.977312390924958,
+ "grad_norm": 0.05571882426738739,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 3.8978285789489746,
+ "step": 2014
+ },
+ {
+ "epoch": 27.991273996509598,
+ "grad_norm": 0.05477411672472954,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 3.875284194946289,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.03992476686835289,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 2.404466152191162,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 0.6039577126502991,
+ "eval_runtime": 59.8223,
+ "eval_samples_per_second": 40.821,
+ "eval_steps_per_second": 0.652,
+ "step": 2016
+ },
+ {
+ "epoch": 28.013961605584644,
+ "grad_norm": 0.05395985022187233,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 3.78012752532959,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027923211169284,
+ "grad_norm": 0.05445432290434837,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 3.794297218322754,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041884816753928,
+ "grad_norm": 0.05048619955778122,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 3.8059635162353516,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05584642233857,
+ "grad_norm": 0.054649483412504196,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 3.8080105781555176,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069808027923212,
+ "grad_norm": 0.054571982473134995,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 3.8171639442443848,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083769633507853,
+ "grad_norm": 0.05476690083742142,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 3.7937488555908203,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097731239092496,
+ "grad_norm": 0.05586248263716698,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 3.804347515106201,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111692844677137,
+ "grad_norm": 0.05466577783226967,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 3.8114442825317383,
+ "step": 2024
+ },
+ {
+ "epoch": 28.12565445026178,
+ "grad_norm": 0.05342315137386322,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 3.8107991218566895,
+ "step": 2025
+ },
+ {
+ "epoch": 28.13961605584642,
+ "grad_norm": 0.05318617820739746,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 3.8390064239501953,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153577661431065,
+ "grad_norm": 0.05371803045272827,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 3.8021721839904785,
+ "step": 2027
+ },
+ {
+ "epoch": 28.167539267015705,
+ "grad_norm": 0.05319817736744881,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 3.844275712966919,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18150087260035,
+ "grad_norm": 0.05202893167734146,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 3.8008670806884766,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195462478184993,
+ "grad_norm": 0.054026514291763306,
+ "learning_rate": 0.000255486047794226,
+ "loss": 3.81990909576416,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209424083769633,
+ "grad_norm": 0.05195610597729683,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 3.8142457008361816,
+ "step": 2031
+ },
+ {
+ "epoch": 28.223385689354277,
+ "grad_norm": 0.05212428793311119,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 3.8139138221740723,
+ "step": 2032
+ },
+ {
+ "epoch": 28.237347294938917,
+ "grad_norm": 0.05503816530108452,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 3.8072667121887207,
+ "step": 2033
+ },
+ {
+ "epoch": 28.25130890052356,
+ "grad_norm": 0.05305452272295952,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 3.8146510124206543,
+ "step": 2034
+ },
+ {
+ "epoch": 28.2652705061082,
+ "grad_norm": 0.05301898717880249,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 3.7936806678771973,
+ "step": 2035
+ },
+ {
+ "epoch": 28.279232111692846,
+ "grad_norm": 0.0542520172894001,
+ "learning_rate": 0.000253907826333243,
+ "loss": 3.7920427322387695,
+ "step": 2036
+ },
+ {
+ "epoch": 28.293193717277486,
+ "grad_norm": 0.05349896103143692,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 3.819732189178467,
+ "step": 2037
+ },
+ {
+ "epoch": 28.30715532286213,
+ "grad_norm": 0.05463109537959099,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 3.804321765899658,
+ "step": 2038
+ },
+ {
+ "epoch": 28.32111692844677,
+ "grad_norm": 0.05445708706974983,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 3.8128511905670166,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335078534031414,
+ "grad_norm": 0.05655837804079056,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 3.798435688018799,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349040139616054,
+ "grad_norm": 0.05543001368641853,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 3.779137134552002,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363001745200698,
+ "grad_norm": 0.05458657816052437,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 3.833167552947998,
+ "step": 2042
+ },
+ {
+ "epoch": 28.376963350785342,
+ "grad_norm": 0.056820500642061234,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 3.8207151889801025,
+ "step": 2043
+ },
+ {
+ "epoch": 28.390924956369982,
+ "grad_norm": 0.05464114993810654,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 3.788215160369873,
+ "step": 2044
+ },
+ {
+ "epoch": 28.404886561954626,
+ "grad_norm": 0.053662411868572235,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 3.8059287071228027,
+ "step": 2045
+ },
+ {
+ "epoch": 28.418848167539267,
+ "grad_norm": 0.055799756199121475,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 3.8050360679626465,
+ "step": 2046
+ },
+ {
+ "epoch": 28.43280977312391,
+ "grad_norm": 0.05412062630057335,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 3.8110790252685547,
+ "step": 2047
+ },
+ {
+ "epoch": 28.44677137870855,
+ "grad_norm": 0.05672532692551613,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 3.841888904571533,
+ "step": 2048
+ },
+ {
+ "epoch": 28.460732984293195,
+ "grad_norm": 0.05622367933392525,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 3.8200669288635254,
+ "step": 2049
+ },
+ {
+ "epoch": 28.474694589877835,
+ "grad_norm": 0.0544997975230217,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 3.8031818866729736,
+ "step": 2050
+ },
+ {
+ "epoch": 28.48865619546248,
+ "grad_norm": 0.058867089450359344,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 3.856843948364258,
+ "step": 2051
+ },
+ {
+ "epoch": 28.50261780104712,
+ "grad_norm": 0.055324845016002655,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 3.807135820388794,
+ "step": 2052
+ },
+ {
+ "epoch": 28.516579406631763,
+ "grad_norm": 0.055414266884326935,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 3.8380215167999268,
+ "step": 2053
+ },
+ {
+ "epoch": 28.530541012216403,
+ "grad_norm": 0.05787660554051399,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 3.831721782684326,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544502617801047,
+ "grad_norm": 0.05551844462752342,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 3.8572306632995605,
+ "step": 2055
+ },
+ {
+ "epoch": 28.55846422338569,
+ "grad_norm": 0.05452537164092064,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 3.840963840484619,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57242582897033,
+ "grad_norm": 0.05530945956707001,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 3.840442419052124,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586387434554975,
+ "grad_norm": 0.05631886050105095,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 3.7851438522338867,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600349040139616,
+ "grad_norm": 0.05489915981888771,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 3.7870893478393555,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61431064572426,
+ "grad_norm": 0.055331867188215256,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 3.8441572189331055,
+ "step": 2060
+ },
+ {
+ "epoch": 28.6282722513089,
+ "grad_norm": 0.05580870062112808,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 3.834148645401001,
+ "step": 2061
+ },
+ {
+ "epoch": 28.642233856893544,
+ "grad_norm": 0.05681769549846649,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 3.8349759578704834,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656195462478184,
+ "grad_norm": 0.05484442040324211,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 3.798792600631714,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670157068062828,
+ "grad_norm": 0.05657925084233284,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 3.853915214538574,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68411867364747,
+ "grad_norm": 0.05419609695672989,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 3.8423995971679688,
+ "step": 2065
+ },
+ {
+ "epoch": 28.698080279232112,
+ "grad_norm": 0.05700303241610527,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 3.822547674179077,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712041884816752,
+ "grad_norm": 0.055846985429525375,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 3.830044984817505,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726003490401396,
+ "grad_norm": 0.05754954740405083,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 3.868316173553467,
+ "step": 2068
+ },
+ {
+ "epoch": 28.739965095986037,
+ "grad_norm": 0.056265417486429214,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 3.8297243118286133,
+ "step": 2069
+ },
+ {
+ "epoch": 28.75392670157068,
+ "grad_norm": 0.05429232493042946,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 3.872836112976074,
+ "step": 2070
+ },
+ {
+ "epoch": 28.767888307155324,
+ "grad_norm": 0.05589764565229416,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 3.838311195373535,
+ "step": 2071
+ },
+ {
+ "epoch": 28.781849912739965,
+ "grad_norm": 0.05460003390908241,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 3.8316597938537598,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79581151832461,
+ "grad_norm": 0.054542504251003265,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 3.816542625427246,
+ "step": 2073
+ },
+ {
+ "epoch": 28.80977312390925,
+ "grad_norm": 0.0549996979534626,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 3.796245574951172,
+ "step": 2074
+ },
+ {
+ "epoch": 28.823734729493893,
+ "grad_norm": 0.05521940439939499,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 3.8112473487854004,
+ "step": 2075
+ },
+ {
+ "epoch": 28.837696335078533,
+ "grad_norm": 0.05690402165055275,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 3.8242125511169434,
+ "step": 2076
+ },
+ {
+ "epoch": 28.851657940663177,
+ "grad_norm": 0.055383820086717606,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 3.84938383102417,
+ "step": 2077
+ },
+ {
+ "epoch": 28.865619546247817,
+ "grad_norm": 0.05612736567854881,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 3.8383970260620117,
+ "step": 2078
+ },
+ {
+ "epoch": 28.87958115183246,
+ "grad_norm": 0.05723752826452255,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 3.828171730041504,
+ "step": 2079
+ },
+ {
+ "epoch": 28.8935427574171,
+ "grad_norm": 0.0576927475631237,
+ "learning_rate": 0.000242380656502223,
+ "loss": 3.8756160736083984,
+ "step": 2080
+ },
+ {
+ "epoch": 28.907504363001745,
+ "grad_norm": 0.05609026178717613,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 3.8703267574310303,
+ "step": 2081
+ },
+ {
+ "epoch": 28.921465968586386,
+ "grad_norm": 0.057038433849811554,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 3.8566956520080566,
+ "step": 2082
+ },
+ {
+ "epoch": 28.93542757417103,
+ "grad_norm": 0.05683749541640282,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 3.8409035205841064,
+ "step": 2083
+ },
+ {
+ "epoch": 28.949389179755673,
+ "grad_norm": 0.05682830885052681,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 3.8435707092285156,
+ "step": 2084
+ },
+ {
+ "epoch": 28.963350785340314,
+ "grad_norm": 0.05645477771759033,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 3.8142364025115967,
+ "step": 2085
+ },
+ {
+ "epoch": 28.977312390924958,
+ "grad_norm": 0.05519694834947586,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 3.830415725708008,
+ "step": 2086
+ },
+ {
+ "epoch": 28.991273996509598,
+ "grad_norm": 0.05785970389842987,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 3.8413705825805664,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.04018386825919151,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 2.3790979385375977,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 0.6047304272651672,
+ "eval_runtime": 61.1049,
+ "eval_samples_per_second": 39.964,
+ "eval_steps_per_second": 0.638,
+ "step": 2088
+ },
+ {
+ "epoch": 29.013961605584644,
+ "grad_norm": 0.058019861578941345,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 3.793783664703369,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027923211169284,
+ "grad_norm": 0.058003105223178864,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 3.783708095550537,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041884816753928,
+ "grad_norm": 0.05485067516565323,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 3.7829298973083496,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05584642233857,
+ "grad_norm": 0.056009069085121155,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 3.7747392654418945,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069808027923212,
+ "grad_norm": 0.05480542406439781,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 3.774707317352295,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083769633507853,
+ "grad_norm": 0.05384726822376251,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 3.75443959236145,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097731239092496,
+ "grad_norm": 0.053818102926015854,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 3.7790050506591797,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111692844677137,
+ "grad_norm": 0.05595344305038452,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 3.74898624420166,
+ "step": 2096
+ },
+ {
+ "epoch": 29.12565445026178,
+ "grad_norm": 0.05394545570015907,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 3.7751331329345703,
+ "step": 2097
+ },
+ {
+ "epoch": 29.13961605584642,
+ "grad_norm": 0.05346408113837242,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 3.7671167850494385,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153577661431065,
+ "grad_norm": 0.0541236512362957,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 3.749098062515259,
+ "step": 2099
+ },
+ {
+ "epoch": 29.167539267015705,
+ "grad_norm": 0.05369208753108978,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 3.7778515815734863,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18150087260035,
+ "grad_norm": 0.055362965911626816,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 3.747697591781616,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195462478184993,
+ "grad_norm": 0.05389911308884621,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 3.7759058475494385,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209424083769633,
+ "grad_norm": 0.05315527692437172,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 3.7832705974578857,
+ "step": 2103
+ },
+ {
+ "epoch": 29.223385689354277,
+ "grad_norm": 0.05377589911222458,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 3.7743210792541504,
+ "step": 2104
+ },
+ {
+ "epoch": 29.237347294938917,
+ "grad_norm": 0.05307800695300102,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 3.7654497623443604,
+ "step": 2105
+ },
+ {
+ "epoch": 29.25130890052356,
+ "grad_norm": 0.056719157844781876,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 3.794179916381836,
+ "step": 2106
+ },
+ {
+ "epoch": 29.2652705061082,
+ "grad_norm": 0.053281862288713455,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 3.7620739936828613,
+ "step": 2107
+ },
+ {
+ "epoch": 29.279232111692846,
+ "grad_norm": 0.05470574274659157,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 3.788421869277954,
+ "step": 2108
+ },
+ {
+ "epoch": 29.293193717277486,
+ "grad_norm": 0.05293853580951691,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 3.7655067443847656,
+ "step": 2109
+ },
+ {
+ "epoch": 29.30715532286213,
+ "grad_norm": 0.05524440482258797,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 3.7701616287231445,
+ "step": 2110
+ },
+ {
+ "epoch": 29.32111692844677,
+ "grad_norm": 0.05508396402001381,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 3.7993319034576416,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335078534031414,
+ "grad_norm": 0.05559603124856949,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 3.7799668312072754,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349040139616054,
+ "grad_norm": 0.05530122295022011,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 3.751032829284668,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363001745200698,
+ "grad_norm": 0.05408896878361702,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 3.7761611938476562,
+ "step": 2114
+ },
+ {
+ "epoch": 29.376963350785342,
+ "grad_norm": 0.05530737340450287,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 3.745713233947754,
+ "step": 2115
+ },
+ {
+ "epoch": 29.390924956369982,
+ "grad_norm": 0.05424542352557182,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 3.8041834831237793,
+ "step": 2116
+ },
+ {
+ "epoch": 29.404886561954626,
+ "grad_norm": 0.054800380021333694,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 3.7995352745056152,
+ "step": 2117
+ },
+ {
+ "epoch": 29.418848167539267,
+ "grad_norm": 0.05685900151729584,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 3.7970776557922363,
+ "step": 2118
+ },
+ {
+ "epoch": 29.43280977312391,
+ "grad_norm": 0.05555475875735283,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 3.7950801849365234,
+ "step": 2119
+ },
+ {
+ "epoch": 29.44677137870855,
+ "grad_norm": 0.05670571327209473,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 3.7967746257781982,
+ "step": 2120
+ },
+ {
+ "epoch": 29.460732984293195,
+ "grad_norm": 0.05598503723740578,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 3.809349298477173,
+ "step": 2121
+ },
+ {
+ "epoch": 29.474694589877835,
+ "grad_norm": 0.05501085892319679,
+ "learning_rate": 0.000231465389734324,
+ "loss": 3.803968906402588,
+ "step": 2122
+ },
+ {
+ "epoch": 29.48865619546248,
+ "grad_norm": 0.05519728362560272,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 3.7615163326263428,
+ "step": 2123
+ },
+ {
+ "epoch": 29.50261780104712,
+ "grad_norm": 0.055866893380880356,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 3.7939207553863525,
+ "step": 2124
+ },
+ {
+ "epoch": 29.516579406631763,
+ "grad_norm": 0.05460986867547035,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 3.7740321159362793,
+ "step": 2125
+ },
+ {
+ "epoch": 29.530541012216403,
+ "grad_norm": 0.055752526968717575,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 3.82424259185791,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544502617801047,
+ "grad_norm": 0.056792002171278,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 3.8006062507629395,
+ "step": 2127
+ },
+ {
+ "epoch": 29.55846422338569,
+ "grad_norm": 0.054986681789159775,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 3.7713897228240967,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57242582897033,
+ "grad_norm": 0.054215673357248306,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 3.8005690574645996,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586387434554975,
+ "grad_norm": 0.05700439214706421,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 3.825161933898926,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600349040139616,
+ "grad_norm": 0.055078551173210144,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 3.786716938018799,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61431064572426,
+ "grad_norm": 0.055569086223840714,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 3.8020172119140625,
+ "step": 2132
+ },
+ {
+ "epoch": 29.6282722513089,
+ "grad_norm": 0.0548536442220211,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 3.8038570880889893,
+ "step": 2133
+ },
+ {
+ "epoch": 29.642233856893544,
+ "grad_norm": 0.05604325234889984,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 3.796006441116333,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656195462478184,
+ "grad_norm": 0.05614005774259567,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 3.7938876152038574,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670157068062828,
+ "grad_norm": 0.05761241167783737,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 3.813568592071533,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68411867364747,
+ "grad_norm": 0.05579913780093193,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 3.826862335205078,
+ "step": 2137
+ },
+ {
+ "epoch": 29.698080279232112,
+ "grad_norm": 0.05784283205866814,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 3.808281421661377,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712041884816752,
+ "grad_norm": 0.05596453696489334,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 3.8119921684265137,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726003490401396,
+ "grad_norm": 0.05699361115694046,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 3.822702407836914,
+ "step": 2140
+ },
+ {
+ "epoch": 29.739965095986037,
+ "grad_norm": 0.05375700071454048,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 3.78023099899292,
+ "step": 2141
+ },
+ {
+ "epoch": 29.75392670157068,
+ "grad_norm": 0.05588206648826599,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 3.791456699371338,
+ "step": 2142
+ },
+ {
+ "epoch": 29.767888307155324,
+ "grad_norm": 0.05579312890768051,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 3.799713134765625,
+ "step": 2143
+ },
+ {
+ "epoch": 29.781849912739965,
+ "grad_norm": 0.05713854730129242,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 3.8211703300476074,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79581151832461,
+ "grad_norm": 0.05645165592432022,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 3.7947046756744385,
+ "step": 2145
+ },
+ {
+ "epoch": 29.80977312390925,
+ "grad_norm": 0.056201402097940445,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 3.811009407043457,
+ "step": 2146
+ },
+ {
+ "epoch": 29.823734729493893,
+ "grad_norm": 0.05659479275345802,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 3.8009982109069824,
+ "step": 2147
+ },
+ {
+ "epoch": 29.837696335078533,
+ "grad_norm": 0.05797034874558449,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 3.7994565963745117,
+ "step": 2148
+ },
+ {
+ "epoch": 29.851657940663177,
+ "grad_norm": 0.05734949931502342,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 3.7696728706359863,
+ "step": 2149
+ },
+ {
+ "epoch": 29.865619546247817,
+ "grad_norm": 0.060396358370780945,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 3.8269689083099365,
+ "step": 2150
+ },
+ {
+ "epoch": 29.87958115183246,
+ "grad_norm": 0.057348065078258514,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 3.7925050258636475,
+ "step": 2151
+ },
+ {
+ "epoch": 29.8935427574171,
+ "grad_norm": 0.058563247323036194,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 3.82887864112854,
+ "step": 2152
+ },
+ {
+ "epoch": 29.907504363001745,
+ "grad_norm": 0.06011585891246796,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 3.8061296939849854,
+ "step": 2153
+ },
+ {
+ "epoch": 29.921465968586386,
+ "grad_norm": 0.058440592139959335,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 3.792586326599121,
+ "step": 2154
+ },
+ {
+ "epoch": 29.93542757417103,
+ "grad_norm": 0.057050157338380814,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 3.7885212898254395,
+ "step": 2155
+ },
+ {
+ "epoch": 29.949389179755673,
+ "grad_norm": 0.05920996889472008,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 3.813566207885742,
+ "step": 2156
+ },
+ {
+ "epoch": 29.963350785340314,
+ "grad_norm": 0.057791322469711304,
+ "learning_rate": 0.00022244633283095,
+ "loss": 3.8399810791015625,
+ "step": 2157
+ },
+ {
+ "epoch": 29.977312390924958,
+ "grad_norm": 0.05644001066684723,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 3.79464054107666,
+ "step": 2158
+ },
+ {
+ "epoch": 29.991273996509598,
+ "grad_norm": 0.06021320819854736,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 3.804086208343506,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.040621865540742874,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 2.386565923690796,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 0.6063579320907593,
+ "eval_runtime": 59.5345,
+ "eval_samples_per_second": 41.018,
+ "eval_steps_per_second": 0.655,
+ "step": 2160
+ },
+ {
+ "epoch": 30.013961605584644,
+ "grad_norm": 0.05961514264345169,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 3.756401777267456,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027923211169284,
+ "grad_norm": 0.057169776409864426,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 3.7327475547790527,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041884816753928,
+ "grad_norm": 0.056084226816892624,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 3.7365498542785645,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05584642233857,
+ "grad_norm": 0.059099674224853516,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 3.7383875846862793,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069808027923212,
+ "grad_norm": 0.05632825568318367,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 3.7642762660980225,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083769633507853,
+ "grad_norm": 0.056937988847494125,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 3.7614428997039795,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097731239092496,
+ "grad_norm": 0.05722609534859657,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 3.7394845485687256,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111692844677137,
+ "grad_norm": 0.056204989552497864,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 3.7473866939544678,
+ "step": 2168
+ },
+ {
+ "epoch": 30.12565445026178,
+ "grad_norm": 0.05640130862593651,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 3.7504067420959473,
+ "step": 2169
+ },
+ {
+ "epoch": 30.13961605584642,
+ "grad_norm": 0.05625637248158455,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 3.754812717437744,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153577661431065,
+ "grad_norm": 0.05596087500452995,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 3.741854429244995,
+ "step": 2171
+ },
+ {
+ "epoch": 30.167539267015705,
+ "grad_norm": 0.054819896817207336,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 3.724705934524536,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18150087260035,
+ "grad_norm": 0.05644344910979271,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 3.7609610557556152,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195462478184993,
+ "grad_norm": 0.05366845801472664,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 3.7557997703552246,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209424083769633,
+ "grad_norm": 0.05517663434147835,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 3.7494120597839355,
+ "step": 2175
+ },
+ {
+ "epoch": 30.223385689354277,
+ "grad_norm": 0.05463553965091705,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 3.7434799671173096,
+ "step": 2176
+ },
+ {
+ "epoch": 30.237347294938917,
+ "grad_norm": 0.05352703109383583,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 3.7435178756713867,
+ "step": 2177
+ },
+ {
+ "epoch": 30.25130890052356,
+ "grad_norm": 0.053156204521656036,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 3.7648849487304688,
+ "step": 2178
+ },
+ {
+ "epoch": 30.2652705061082,
+ "grad_norm": 0.05557673051953316,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 3.746366024017334,
+ "step": 2179
+ },
+ {
+ "epoch": 30.279232111692846,
+ "grad_norm": 0.052989061921834946,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 3.7453460693359375,
+ "step": 2180
+ },
+ {
+ "epoch": 30.293193717277486,
+ "grad_norm": 0.055707018822431564,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 3.800962209701538,
+ "step": 2181
+ },
+ {
+ "epoch": 30.30715532286213,
+ "grad_norm": 0.05560754984617233,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 3.7416906356811523,
+ "step": 2182
+ },
+ {
+ "epoch": 30.32111692844677,
+ "grad_norm": 0.055344413965940475,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 3.7672080993652344,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335078534031414,
+ "grad_norm": 0.05490704998373985,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 3.7664341926574707,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349040139616054,
+ "grad_norm": 0.0578596331179142,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 3.7632641792297363,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363001745200698,
+ "grad_norm": 0.055299609899520874,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 3.7502024173736572,
+ "step": 2186
+ },
+ {
+ "epoch": 30.376963350785342,
+ "grad_norm": 0.055016499012708664,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 3.732128143310547,
+ "step": 2187
+ },
+ {
+ "epoch": 30.390924956369982,
+ "grad_norm": 0.057352419942617416,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 3.723008871078491,
+ "step": 2188
+ },
+ {
+ "epoch": 30.404886561954626,
+ "grad_norm": 0.05596115067601204,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 3.786574363708496,
+ "step": 2189
+ },
+ {
+ "epoch": 30.418848167539267,
+ "grad_norm": 0.05596112087368965,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 3.770382881164551,
+ "step": 2190
+ },
+ {
+ "epoch": 30.43280977312391,
+ "grad_norm": 0.0562058687210083,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 3.7605714797973633,
+ "step": 2191
+ },
+ {
+ "epoch": 30.44677137870855,
+ "grad_norm": 0.056407757103443146,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 3.7485427856445312,
+ "step": 2192
+ },
+ {
+ "epoch": 30.460732984293195,
+ "grad_norm": 0.05596652254462242,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 3.7630670070648193,
+ "step": 2193
+ },
+ {
+ "epoch": 30.474694589877835,
+ "grad_norm": 0.056551866233348846,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 3.7981796264648438,
+ "step": 2194
+ },
+ {
+ "epoch": 30.48865619546248,
+ "grad_norm": 0.057354725897312164,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 3.7544898986816406,
+ "step": 2195
+ },
+ {
+ "epoch": 30.50261780104712,
+ "grad_norm": 0.05791718140244484,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 3.7241907119750977,
+ "step": 2196
+ },
+ {
+ "epoch": 30.516579406631763,
+ "grad_norm": 0.05675847455859184,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 3.794506549835205,
+ "step": 2197
+ },
+ {
+ "epoch": 30.530541012216403,
+ "grad_norm": 0.060349952429533005,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 3.75946044921875,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544502617801047,
+ "grad_norm": 0.05547260120511055,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 3.7278571128845215,
+ "step": 2199
+ },
+ {
+ "epoch": 30.55846422338569,
+ "grad_norm": 0.058400996029376984,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 3.756891965866089,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57242582897033,
+ "grad_norm": 0.057871121913194656,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 3.783863067626953,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586387434554975,
+ "grad_norm": 0.05899912118911743,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 3.7787115573883057,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600349040139616,
+ "grad_norm": 0.056164853274822235,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 3.7599334716796875,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61431064572426,
+ "grad_norm": 0.05775163322687149,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 3.7457070350646973,
+ "step": 2204
+ },
+ {
+ "epoch": 30.6282722513089,
+ "grad_norm": 0.057221125811338425,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 3.7663235664367676,
+ "step": 2205
+ },
+ {
+ "epoch": 30.642233856893544,
+ "grad_norm": 0.05788435414433479,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 3.7676830291748047,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656195462478184,
+ "grad_norm": 0.057100843638181686,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 3.7586870193481445,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670157068062828,
+ "grad_norm": 0.05703200399875641,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 3.7561769485473633,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68411867364747,
+ "grad_norm": 0.05664575845003128,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 3.761331081390381,
+ "step": 2209
+ },
+ {
+ "epoch": 30.698080279232112,
+ "grad_norm": 0.056622982025146484,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 3.737915515899658,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712041884816752,
+ "grad_norm": 0.056481845676898956,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 3.7413077354431152,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726003490401396,
+ "grad_norm": 0.057718146592378616,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 3.751267433166504,
+ "step": 2212
+ },
+ {
+ "epoch": 30.739965095986037,
+ "grad_norm": 0.05772862955927849,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 3.753678321838379,
+ "step": 2213
+ },
+ {
+ "epoch": 30.75392670157068,
+ "grad_norm": 0.056571926921606064,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 3.751033306121826,
+ "step": 2214
+ },
+ {
+ "epoch": 30.767888307155324,
+ "grad_norm": 0.05682666599750519,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 3.763153553009033,
+ "step": 2215
+ },
+ {
+ "epoch": 30.781849912739965,
+ "grad_norm": 0.05993897467851639,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 3.787890911102295,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79581151832461,
+ "grad_norm": 0.05659458413720131,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 3.7655086517333984,
+ "step": 2217
+ },
+ {
+ "epoch": 30.80977312390925,
+ "grad_norm": 0.057996224611997604,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 3.774881601333618,
+ "step": 2218
+ },
+ {
+ "epoch": 30.823734729493893,
+ "grad_norm": 0.05677317455410957,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 3.7600479125976562,
+ "step": 2219
+ },
+ {
+ "epoch": 30.837696335078533,
+ "grad_norm": 0.05664737895131111,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 3.787637233734131,
+ "step": 2220
+ },
+ {
+ "epoch": 30.851657940663177,
+ "grad_norm": 0.05808616429567337,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 3.7529220581054688,
+ "step": 2221
+ },
+ {
+ "epoch": 30.865619546247817,
+ "grad_norm": 0.05666997656226158,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 3.7790794372558594,
+ "step": 2222
+ },
+ {
+ "epoch": 30.87958115183246,
+ "grad_norm": 0.05963682755827904,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 3.762615203857422,
+ "step": 2223
+ },
+ {
+ "epoch": 30.8935427574171,
+ "grad_norm": 0.05558215081691742,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 3.7871851921081543,
+ "step": 2224
+ },
+ {
+ "epoch": 30.907504363001745,
+ "grad_norm": 0.05762261897325516,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 3.7605397701263428,
+ "step": 2225
+ },
+ {
+ "epoch": 30.921465968586386,
+ "grad_norm": 0.0565858818590641,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 3.7693653106689453,
+ "step": 2226
+ },
+ {
+ "epoch": 30.93542757417103,
+ "grad_norm": 0.05821581557393074,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 3.766758918762207,
+ "step": 2227
+ },
+ {
+ "epoch": 30.949389179755673,
+ "grad_norm": 0.055186714977025986,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 3.79498553276062,
+ "step": 2228
+ },
+ {
+ "epoch": 30.963350785340314,
+ "grad_norm": 0.05850847810506821,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 3.799769163131714,
+ "step": 2229
+ },
+ {
+ "epoch": 30.977312390924958,
+ "grad_norm": 0.05700397863984108,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 3.7555625438690186,
+ "step": 2230
+ },
+ {
+ "epoch": 30.991273996509598,
+ "grad_norm": 0.057813212275505066,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 3.777778148651123,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.041730619966983795,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 2.3564491271972656,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 0.6080329418182373,
+ "eval_runtime": 60.4607,
+ "eval_samples_per_second": 40.39,
+ "eval_steps_per_second": 0.645,
+ "step": 2232
+ },
+ {
+ "epoch": 31.013961605584644,
+ "grad_norm": 0.0563117116689682,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 3.719982624053955,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027923211169284,
+ "grad_norm": 0.05564197897911072,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 3.719735622406006,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041884816753928,
+ "grad_norm": 0.05562874674797058,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 3.739713430404663,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05584642233857,
+ "grad_norm": 0.057146716862916946,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 3.736401081085205,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069808027923212,
+ "grad_norm": 0.053163520991802216,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 3.7049455642700195,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083769633507853,
+ "grad_norm": 0.057067256420850754,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 3.723662853240967,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097731239092496,
+ "grad_norm": 0.05276816710829735,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 3.710062026977539,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111692844677137,
+ "grad_norm": 0.055126190185546875,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 3.696173667907715,
+ "step": 2240
+ },
+ {
+ "epoch": 31.12565445026178,
+ "grad_norm": 0.05533793941140175,
+ "learning_rate": 0.000201141724176723,
+ "loss": 3.710416793823242,
+ "step": 2241
+ },
+ {
+ "epoch": 31.13961605584642,
+ "grad_norm": 0.055686868727207184,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 3.7188220024108887,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153577661431065,
+ "grad_norm": 0.055188920348882675,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 3.7026562690734863,
+ "step": 2243
+ },
+ {
+ "epoch": 31.167539267015705,
+ "grad_norm": 0.053627416491508484,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 3.7224509716033936,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18150087260035,
+ "grad_norm": 0.05535290017724037,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 3.710698366165161,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195462478184993,
+ "grad_norm": 0.0528891459107399,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 3.6979012489318848,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209424083769633,
+ "grad_norm": 0.055201176553964615,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 3.7324817180633545,
+ "step": 2247
+ },
+ {
+ "epoch": 31.223385689354277,
+ "grad_norm": 0.05338778346776962,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 3.722564697265625,
+ "step": 2248
+ },
+ {
+ "epoch": 31.237347294938917,
+ "grad_norm": 0.05368843302130699,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 3.7118425369262695,
+ "step": 2249
+ },
+ {
+ "epoch": 31.25130890052356,
+ "grad_norm": 0.052788376808166504,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 3.6875901222229004,
+ "step": 2250
+ },
+ {
+ "epoch": 31.2652705061082,
+ "grad_norm": 0.05427199602127075,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 3.7042768001556396,
+ "step": 2251
+ },
+ {
+ "epoch": 31.279232111692846,
+ "grad_norm": 0.05434224382042885,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 3.7138795852661133,
+ "step": 2252
+ },
+ {
+ "epoch": 31.293193717277486,
+ "grad_norm": 0.054364923387765884,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 3.735142946243286,
+ "step": 2253
+ },
+ {
+ "epoch": 31.30715532286213,
+ "grad_norm": 0.05535430088639259,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 3.7198615074157715,
+ "step": 2254
+ },
+ {
+ "epoch": 31.32111692844677,
+ "grad_norm": 0.05463665351271629,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 3.733370542526245,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335078534031414,
+ "grad_norm": 0.0553361251950264,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 3.702805519104004,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349040139616054,
+ "grad_norm": 0.05423365905880928,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 3.7315120697021484,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363001745200698,
+ "grad_norm": 0.05499394237995148,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 3.731891632080078,
+ "step": 2258
+ },
+ {
+ "epoch": 31.376963350785342,
+ "grad_norm": 0.054968126118183136,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 3.7355899810791016,
+ "step": 2259
+ },
+ {
+ "epoch": 31.390924956369982,
+ "grad_norm": 0.05478367581963539,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 3.7143867015838623,
+ "step": 2260
+ },
+ {
+ "epoch": 31.404886561954626,
+ "grad_norm": 0.05554179847240448,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 3.7227137088775635,
+ "step": 2261
+ },
+ {
+ "epoch": 31.418848167539267,
+ "grad_norm": 0.05399509146809578,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 3.750650405883789,
+ "step": 2262
+ },
+ {
+ "epoch": 31.43280977312391,
+ "grad_norm": 0.056201111525297165,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 3.7478344440460205,
+ "step": 2263
+ },
+ {
+ "epoch": 31.44677137870855,
+ "grad_norm": 0.05437091737985611,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 3.7283718585968018,
+ "step": 2264
+ },
+ {
+ "epoch": 31.460732984293195,
+ "grad_norm": 0.057144906371831894,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 3.7482407093048096,
+ "step": 2265
+ },
+ {
+ "epoch": 31.474694589877835,
+ "grad_norm": 0.05441201478242874,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 3.726378917694092,
+ "step": 2266
+ },
+ {
+ "epoch": 31.48865619546248,
+ "grad_norm": 0.056569576263427734,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 3.7211437225341797,
+ "step": 2267
+ },
+ {
+ "epoch": 31.50261780104712,
+ "grad_norm": 0.05561831593513489,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 3.7278850078582764,
+ "step": 2268
+ },
+ {
+ "epoch": 31.516579406631763,
+ "grad_norm": 0.0569809265434742,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 3.713228702545166,
+ "step": 2269
+ },
+ {
+ "epoch": 31.530541012216403,
+ "grad_norm": 0.05654345825314522,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 3.7354423999786377,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544502617801047,
+ "grad_norm": 0.05482077598571777,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 3.7043380737304688,
+ "step": 2271
+ },
+ {
+ "epoch": 31.55846422338569,
+ "grad_norm": 0.055417660623788834,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 3.7199621200561523,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57242582897033,
+ "grad_norm": 0.054656337946653366,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 3.734508514404297,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586387434554975,
+ "grad_norm": 0.05609512701630592,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 3.7341227531433105,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600349040139616,
+ "grad_norm": 0.05602060630917549,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 3.719949960708618,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61431064572426,
+ "grad_norm": 0.05535270646214485,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 3.7124414443969727,
+ "step": 2276
+ },
+ {
+ "epoch": 31.6282722513089,
+ "grad_norm": 0.05577492713928223,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 3.706301212310791,
+ "step": 2277
+ },
+ {
+ "epoch": 31.642233856893544,
+ "grad_norm": 0.057382550090551376,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 3.7411818504333496,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656195462478184,
+ "grad_norm": 0.056383028626441956,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 3.751235008239746,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670157068062828,
+ "grad_norm": 0.05667940899729729,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 3.7379674911499023,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68411867364747,
+ "grad_norm": 0.05635855346918106,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 3.728001594543457,
+ "step": 2281
+ },
+ {
+ "epoch": 31.698080279232112,
+ "grad_norm": 0.05649873986840248,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 3.740757465362549,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712041884816752,
+ "grad_norm": 0.055871494114398956,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 3.7373616695404053,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726003490401396,
+ "grad_norm": 0.057106468826532364,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 3.7572178840637207,
+ "step": 2284
+ },
+ {
+ "epoch": 31.739965095986037,
+ "grad_norm": 0.05788225308060646,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 3.7474923133850098,
+ "step": 2285
+ },
+ {
+ "epoch": 31.75392670157068,
+ "grad_norm": 0.057152729481458664,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 3.7224936485290527,
+ "step": 2286
+ },
+ {
+ "epoch": 31.767888307155324,
+ "grad_norm": 0.0560687892138958,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 3.7336764335632324,
+ "step": 2287
+ },
+ {
+ "epoch": 31.781849912739965,
+ "grad_norm": 0.05890286713838577,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 3.754326343536377,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79581151832461,
+ "grad_norm": 0.058598410338163376,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 3.7265028953552246,
+ "step": 2289
+ },
+ {
+ "epoch": 31.80977312390925,
+ "grad_norm": 0.05728591978549957,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 3.715481758117676,
+ "step": 2290
+ },
+ {
+ "epoch": 31.823734729493893,
+ "grad_norm": 0.05874952673912048,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 3.7139101028442383,
+ "step": 2291
+ },
+ {
+ "epoch": 31.837696335078533,
+ "grad_norm": 0.056882575154304504,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 3.728553533554077,
+ "step": 2292
+ },
+ {
+ "epoch": 31.851657940663177,
+ "grad_norm": 0.057810861617326736,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 3.757908582687378,
+ "step": 2293
+ },
+ {
+ "epoch": 31.865619546247817,
+ "grad_norm": 0.057366516441106796,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 3.7265639305114746,
+ "step": 2294
+ },
+ {
+ "epoch": 31.87958115183246,
+ "grad_norm": 0.059405453503131866,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 3.740511417388916,
+ "step": 2295
+ },
+ {
+ "epoch": 31.8935427574171,
+ "grad_norm": 0.05881514027714729,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 3.751497268676758,
+ "step": 2296
+ },
+ {
+ "epoch": 31.907504363001745,
+ "grad_norm": 0.06030813232064247,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 3.75439715385437,
+ "step": 2297
+ },
+ {
+ "epoch": 31.921465968586386,
+ "grad_norm": 0.06062643602490425,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 3.7215723991394043,
+ "step": 2298
+ },
+ {
+ "epoch": 31.93542757417103,
+ "grad_norm": 0.05953266844153404,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 3.7635340690612793,
+ "step": 2299
+ },
+ {
+ "epoch": 31.949389179755673,
+ "grad_norm": 0.06019049137830734,
+ "learning_rate": 0.000186516746349841,
+ "loss": 3.7750091552734375,
+ "step": 2300
+ },
+ {
+ "epoch": 31.963350785340314,
+ "grad_norm": 0.058755744248628616,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 3.7646124362945557,
+ "step": 2301
+ },
+ {
+ "epoch": 31.977312390924958,
+ "grad_norm": 0.05986425653100014,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 3.7566630840301514,
+ "step": 2302
+ },
+ {
+ "epoch": 31.991273996509598,
+ "grad_norm": 0.057212650775909424,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 3.714637279510498,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.04401736706495285,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 2.33223032951355,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 0.6091130375862122,
+ "eval_runtime": 60.7334,
+ "eval_samples_per_second": 40.209,
+ "eval_steps_per_second": 0.642,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01396160558464,
+ "grad_norm": 0.05650010704994202,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 3.702322244644165,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02792321116929,
+ "grad_norm": 0.054223258048295975,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 3.699272632598877,
+ "step": 2306
+ },
+ {
+ "epoch": 32.04188481675393,
+ "grad_norm": 0.05616949126124382,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 3.6740047931671143,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05584642233857,
+ "grad_norm": 0.05479196831583977,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 3.6942710876464844,
+ "step": 2308
+ },
+ {
+ "epoch": 32.06980802792321,
+ "grad_norm": 0.05735877528786659,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 3.6724581718444824,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083769633507856,
+ "grad_norm": 0.056006450206041336,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 3.6736817359924316,
+ "step": 2310
+ },
+ {
+ "epoch": 32.0977312390925,
+ "grad_norm": 0.055535659193992615,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 3.710303783416748,
+ "step": 2311
+ },
+ {
+ "epoch": 32.11169284467714,
+ "grad_norm": 0.05510629341006279,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 3.669536590576172,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12565445026178,
+ "grad_norm": 0.05426035821437836,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 3.670957088470459,
+ "step": 2313
+ },
+ {
+ "epoch": 32.139616055846425,
+ "grad_norm": 0.05547798424959183,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 3.7050318717956543,
+ "step": 2314
+ },
+ {
+ "epoch": 32.153577661431065,
+ "grad_norm": 0.05412966012954712,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 3.6849310398101807,
+ "step": 2315
+ },
+ {
+ "epoch": 32.167539267015705,
+ "grad_norm": 0.0554192028939724,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 3.702563762664795,
+ "step": 2316
+ },
+ {
+ "epoch": 32.181500872600346,
+ "grad_norm": 0.054564788937568665,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 3.707547187805176,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19546247818499,
+ "grad_norm": 0.05424094945192337,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 3.650768280029297,
+ "step": 2318
+ },
+ {
+ "epoch": 32.20942408376963,
+ "grad_norm": 0.05562194436788559,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 3.6763458251953125,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223385689354274,
+ "grad_norm": 0.053762275725603104,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 3.690493583679199,
+ "step": 2320
+ },
+ {
+ "epoch": 32.23734729493892,
+ "grad_norm": 0.05513778328895569,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 3.6990034580230713,
+ "step": 2321
+ },
+ {
+ "epoch": 32.25130890052356,
+ "grad_norm": 0.05438566580414772,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 3.702357292175293,
+ "step": 2322
+ },
+ {
+ "epoch": 32.2652705061082,
+ "grad_norm": 0.054521508514881134,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 3.701066255569458,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27923211169284,
+ "grad_norm": 0.05416509136557579,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 3.673935651779175,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29319371727749,
+ "grad_norm": 0.05350432172417641,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 3.6740450859069824,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30715532286213,
+ "grad_norm": 0.054736558347940445,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 3.710815668106079,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32111692844677,
+ "grad_norm": 0.055077534168958664,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 3.6732773780822754,
+ "step": 2327
+ },
+ {
+ "epoch": 32.33507853403141,
+ "grad_norm": 0.054749298840761185,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 3.6978728771209717,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34904013961606,
+ "grad_norm": 0.05378924310207367,
+ "learning_rate": 0.000179445406945268,
+ "loss": 3.6783430576324463,
+ "step": 2329
+ },
+ {
+ "epoch": 32.3630017452007,
+ "grad_norm": 0.0553874596953392,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 3.6876044273376465,
+ "step": 2330
+ },
+ {
+ "epoch": 32.37696335078534,
+ "grad_norm": 0.053926583379507065,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 3.712519884109497,
+ "step": 2331
+ },
+ {
+ "epoch": 32.390924956369986,
+ "grad_norm": 0.055778294801712036,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 3.686643123626709,
+ "step": 2332
+ },
+ {
+ "epoch": 32.404886561954626,
+ "grad_norm": 0.05676745995879173,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 3.685737371444702,
+ "step": 2333
+ },
+ {
+ "epoch": 32.41884816753927,
+ "grad_norm": 0.054717857390642166,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 3.6684136390686035,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43280977312391,
+ "grad_norm": 0.0552208311855793,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 3.696580410003662,
+ "step": 2335
+ },
+ {
+ "epoch": 32.446771378708554,
+ "grad_norm": 0.0549275167286396,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 3.7011008262634277,
+ "step": 2336
+ },
+ {
+ "epoch": 32.460732984293195,
+ "grad_norm": 0.05518808960914612,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 3.7204928398132324,
+ "step": 2337
+ },
+ {
+ "epoch": 32.474694589877835,
+ "grad_norm": 0.05621159076690674,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 3.68092679977417,
+ "step": 2338
+ },
+ {
+ "epoch": 32.488656195462475,
+ "grad_norm": 0.054848797619342804,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 3.677091598510742,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50261780104712,
+ "grad_norm": 0.0562039278447628,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 3.7283501625061035,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51657940663176,
+ "grad_norm": 0.05434577167034149,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 3.710252285003662,
+ "step": 2341
+ },
+ {
+ "epoch": 32.5305410122164,
+ "grad_norm": 0.05737415328621864,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 3.7027816772460938,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544502617801044,
+ "grad_norm": 0.057093650102615356,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 3.7027580738067627,
+ "step": 2343
+ },
+ {
+ "epoch": 32.55846422338569,
+ "grad_norm": 0.0550924576818943,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 3.68161678314209,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57242582897033,
+ "grad_norm": 0.057579416781663895,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 3.67604923248291,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58638743455497,
+ "grad_norm": 0.05515696108341217,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 3.693655490875244,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60034904013962,
+ "grad_norm": 0.05788129195570946,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 3.725567102432251,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61431064572426,
+ "grad_norm": 0.05617021024227142,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 3.7027430534362793,
+ "step": 2348
+ },
+ {
+ "epoch": 32.6282722513089,
+ "grad_norm": 0.05649486556649208,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 3.717970371246338,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64223385689354,
+ "grad_norm": 0.05582305043935776,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 3.6795706748962402,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65619546247819,
+ "grad_norm": 0.058177400380373,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 3.6926321983337402,
+ "step": 2351
+ },
+ {
+ "epoch": 32.67015706806283,
+ "grad_norm": 0.05563648045063019,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 3.7227001190185547,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68411867364747,
+ "grad_norm": 0.058799829334020615,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 3.706979513168335,
+ "step": 2353
+ },
+ {
+ "epoch": 32.69808027923211,
+ "grad_norm": 0.05645359680056572,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 3.6872620582580566,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712041884816756,
+ "grad_norm": 0.058064281940460205,
+ "learning_rate": 0.000173176617304673,
+ "loss": 3.695530414581299,
+ "step": 2355
+ },
+ {
+ "epoch": 32.726003490401396,
+ "grad_norm": 0.05756758525967598,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 3.7229819297790527,
+ "step": 2356
+ },
+ {
+ "epoch": 32.73996509598604,
+ "grad_norm": 0.057592105120420456,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 3.7098472118377686,
+ "step": 2357
+ },
+ {
+ "epoch": 32.753926701570684,
+ "grad_norm": 0.05855923518538475,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 3.704477310180664,
+ "step": 2358
+ },
+ {
+ "epoch": 32.767888307155324,
+ "grad_norm": 0.05809807404875755,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 3.7231266498565674,
+ "step": 2359
+ },
+ {
+ "epoch": 32.781849912739965,
+ "grad_norm": 0.05663153529167175,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 3.708165168762207,
+ "step": 2360
+ },
+ {
+ "epoch": 32.795811518324605,
+ "grad_norm": 0.05867684260010719,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 3.70230770111084,
+ "step": 2361
+ },
+ {
+ "epoch": 32.80977312390925,
+ "grad_norm": 0.05850773677229881,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 3.7339577674865723,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82373472949389,
+ "grad_norm": 0.05651238560676575,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 3.7185606956481934,
+ "step": 2363
+ },
+ {
+ "epoch": 32.83769633507853,
+ "grad_norm": 0.05758710578083992,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 3.677966356277466,
+ "step": 2364
+ },
+ {
+ "epoch": 32.85165794066317,
+ "grad_norm": 0.057175327092409134,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 3.720883846282959,
+ "step": 2365
+ },
+ {
+ "epoch": 32.86561954624782,
+ "grad_norm": 0.05690391734242439,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 3.706535577774048,
+ "step": 2366
+ },
+ {
+ "epoch": 32.87958115183246,
+ "grad_norm": 0.05758208781480789,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 3.743523120880127,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8935427574171,
+ "grad_norm": 0.05831100419163704,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 3.726163387298584,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90750436300174,
+ "grad_norm": 0.05664364993572235,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 3.691837787628174,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92146596858639,
+ "grad_norm": 0.05847351998090744,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 3.7386794090270996,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93542757417103,
+ "grad_norm": 0.05673616752028465,
+ "learning_rate": 0.00016935382741164,
+ "loss": 3.72623348236084,
+ "step": 2371
+ },
+ {
+ "epoch": 32.94938917975567,
+ "grad_norm": 0.058967288583517075,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 3.733576536178589,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96335078534032,
+ "grad_norm": 0.056366659700870514,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 3.7006077766418457,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97731239092496,
+ "grad_norm": 0.05691685527563095,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 3.7104835510253906,
+ "step": 2374
+ },
+ {
+ "epoch": 32.9912739965096,
+ "grad_norm": 0.05809671804308891,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 3.7175397872924805,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.0423070453107357,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 2.3232760429382324,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 0.6105995178222656,
+ "eval_runtime": 61.6632,
+ "eval_samples_per_second": 39.602,
+ "eval_steps_per_second": 0.632,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01396160558464,
+ "grad_norm": 0.05560882017016411,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 3.6668033599853516,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02792321116929,
+ "grad_norm": 0.053790636360645294,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 3.6379659175872803,
+ "step": 2378
+ },
+ {
+ "epoch": 33.04188481675393,
+ "grad_norm": 0.056180961430072784,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 3.6418814659118652,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05584642233857,
+ "grad_norm": 0.054904621094465256,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 3.651149034500122,
+ "step": 2380
+ },
+ {
+ "epoch": 33.06980802792321,
+ "grad_norm": 0.05298716947436333,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 3.686598777770996,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083769633507856,
+ "grad_norm": 0.053739335387945175,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 3.6473655700683594,
+ "step": 2382
+ },
+ {
+ "epoch": 33.0977312390925,
+ "grad_norm": 0.05417626351118088,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 3.6908068656921387,
+ "step": 2383
+ },
+ {
+ "epoch": 33.11169284467714,
+ "grad_norm": 0.0533599928021431,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 3.6613330841064453,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12565445026178,
+ "grad_norm": 0.05309661850333214,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 3.6534783840179443,
+ "step": 2385
+ },
+ {
+ "epoch": 33.139616055846425,
+ "grad_norm": 0.05432313680648804,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 3.664217948913574,
+ "step": 2386
+ },
+ {
+ "epoch": 33.153577661431065,
+ "grad_norm": 0.054764073342084885,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 3.660922050476074,
+ "step": 2387
+ },
+ {
+ "epoch": 33.167539267015705,
+ "grad_norm": 0.05396303907036781,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 3.662687063217163,
+ "step": 2388
+ },
+ {
+ "epoch": 33.181500872600346,
+ "grad_norm": 0.05540551245212555,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 3.6671347618103027,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19546247818499,
+ "grad_norm": 0.05510672926902771,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 3.6410186290740967,
+ "step": 2390
+ },
+ {
+ "epoch": 33.20942408376963,
+ "grad_norm": 0.055180903524160385,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 3.653822422027588,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223385689354274,
+ "grad_norm": 0.054320868104696274,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 3.6835601329803467,
+ "step": 2392
+ },
+ {
+ "epoch": 33.23734729493892,
+ "grad_norm": 0.05568838492035866,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 3.671053171157837,
+ "step": 2393
+ },
+ {
+ "epoch": 33.25130890052356,
+ "grad_norm": 0.056060485541820526,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 3.682260751724243,
+ "step": 2394
+ },
+ {
+ "epoch": 33.2652705061082,
+ "grad_norm": 0.054811324924230576,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 3.6742751598358154,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27923211169284,
+ "grad_norm": 0.05570835620164871,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 3.6490235328674316,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29319371727749,
+ "grad_norm": 0.05547381564974785,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 3.6795434951782227,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30715532286213,
+ "grad_norm": 0.05463161692023277,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 3.655003547668457,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32111692844677,
+ "grad_norm": 0.05583210662007332,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 3.680175304412842,
+ "step": 2399
+ },
+ {
+ "epoch": 33.33507853403141,
+ "grad_norm": 0.056169651448726654,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 3.667175054550171,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34904013961606,
+ "grad_norm": 0.055788591504096985,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 3.6673946380615234,
+ "step": 2401
+ },
+ {
+ "epoch": 33.3630017452007,
+ "grad_norm": 0.05556197836995125,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 3.650700569152832,
+ "step": 2402
+ },
+ {
+ "epoch": 33.37696335078534,
+ "grad_norm": 0.05749715119600296,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 3.6977670192718506,
+ "step": 2403
+ },
+ {
+ "epoch": 33.390924956369986,
+ "grad_norm": 0.05672239512205124,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 3.6604905128479004,
+ "step": 2404
+ },
+ {
+ "epoch": 33.404886561954626,
+ "grad_norm": 0.05661430582404137,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 3.64540958404541,
+ "step": 2405
+ },
+ {
+ "epoch": 33.41884816753927,
+ "grad_norm": 0.05520932003855705,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 3.6648027896881104,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43280977312391,
+ "grad_norm": 0.0557708777487278,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 3.6694397926330566,
+ "step": 2407
+ },
+ {
+ "epoch": 33.446771378708554,
+ "grad_norm": 0.05510152503848076,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 3.6723294258117676,
+ "step": 2408
+ },
+ {
+ "epoch": 33.460732984293195,
+ "grad_norm": 0.05714292824268341,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 3.6610422134399414,
+ "step": 2409
+ },
+ {
+ "epoch": 33.474694589877835,
+ "grad_norm": 0.05525747686624527,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 3.655020236968994,
+ "step": 2410
+ },
+ {
+ "epoch": 33.488656195462475,
+ "grad_norm": 0.05610140040516853,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 3.6707541942596436,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50261780104712,
+ "grad_norm": 0.05564311146736145,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 3.699244976043701,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51657940663176,
+ "grad_norm": 0.05607445910573006,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 3.6682703495025635,
+ "step": 2413
+ },
+ {
+ "epoch": 33.5305410122164,
+ "grad_norm": 0.05434789881110191,
+ "learning_rate": 0.000159218843594243,
+ "loss": 3.6799025535583496,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544502617801044,
+ "grad_norm": 0.055795323103666306,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 3.6433563232421875,
+ "step": 2415
+ },
+ {
+ "epoch": 33.55846422338569,
+ "grad_norm": 0.05676727741956711,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 3.665454864501953,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57242582897033,
+ "grad_norm": 0.05657052621245384,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 3.6618518829345703,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58638743455497,
+ "grad_norm": 0.056275662034749985,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 3.6946463584899902,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60034904013962,
+ "grad_norm": 0.056533295661211014,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 3.6695780754089355,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61431064572426,
+ "grad_norm": 0.05692996829748154,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 3.6850693225860596,
+ "step": 2420
+ },
+ {
+ "epoch": 33.6282722513089,
+ "grad_norm": 0.0577891543507576,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 3.668393611907959,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64223385689354,
+ "grad_norm": 0.05549122393131256,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 3.6772003173828125,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65619546247819,
+ "grad_norm": 0.05636308714747429,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 3.6818785667419434,
+ "step": 2423
+ },
+ {
+ "epoch": 33.67015706806283,
+ "grad_norm": 0.05681503564119339,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 3.6955409049987793,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68411867364747,
+ "grad_norm": 0.05746293440461159,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 3.6917951107025146,
+ "step": 2425
+ },
+ {
+ "epoch": 33.69808027923211,
+ "grad_norm": 0.056621894240379333,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 3.635894775390625,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712041884816756,
+ "grad_norm": 0.05611368268728256,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 3.679736375808716,
+ "step": 2427
+ },
+ {
+ "epoch": 33.726003490401396,
+ "grad_norm": 0.057769183069467545,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 3.6751227378845215,
+ "step": 2428
+ },
+ {
+ "epoch": 33.73996509598604,
+ "grad_norm": 0.05695759132504463,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 3.685178756713867,
+ "step": 2429
+ },
+ {
+ "epoch": 33.753926701570684,
+ "grad_norm": 0.05776882544159889,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 3.6355032920837402,
+ "step": 2430
+ },
+ {
+ "epoch": 33.767888307155324,
+ "grad_norm": 0.056043099611997604,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 3.668668270111084,
+ "step": 2431
+ },
+ {
+ "epoch": 33.781849912739965,
+ "grad_norm": 0.05634639039635658,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 3.6788253784179688,
+ "step": 2432
+ },
+ {
+ "epoch": 33.795811518324605,
+ "grad_norm": 0.05703914538025856,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 3.676163911819458,
+ "step": 2433
+ },
+ {
+ "epoch": 33.80977312390925,
+ "grad_norm": 0.05846793204545975,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 3.657705307006836,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82373472949389,
+ "grad_norm": 0.059191398322582245,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 3.70918607711792,
+ "step": 2435
+ },
+ {
+ "epoch": 33.83769633507853,
+ "grad_norm": 0.056207913905382156,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 3.690638542175293,
+ "step": 2436
+ },
+ {
+ "epoch": 33.85165794066317,
+ "grad_norm": 0.0590900182723999,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 3.6839122772216797,
+ "step": 2437
+ },
+ {
+ "epoch": 33.86561954624782,
+ "grad_norm": 0.05649874731898308,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 3.6689414978027344,
+ "step": 2438
+ },
+ {
+ "epoch": 33.87958115183246,
+ "grad_norm": 0.05947998911142349,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 3.653146743774414,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8935427574171,
+ "grad_norm": 0.05769696831703186,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 3.685544967651367,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90750436300174,
+ "grad_norm": 0.05837882682681084,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 3.6648831367492676,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92146596858639,
+ "grad_norm": 0.05676107108592987,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 3.7186074256896973,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93542757417103,
+ "grad_norm": 0.05771078169345856,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 3.678924083709717,
+ "step": 2443
+ },
+ {
+ "epoch": 33.94938917975567,
+ "grad_norm": 0.05686977133154869,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 3.7026078701019287,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96335078534032,
+ "grad_norm": 0.0571015290915966,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 3.6880578994750977,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97731239092496,
+ "grad_norm": 0.056974101811647415,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 3.6521928310394287,
+ "step": 2446
+ },
+ {
+ "epoch": 33.9912739965096,
+ "grad_norm": 0.05962125584483147,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 3.686126708984375,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.043099287897348404,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 2.308753490447998,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 0.6120397448539734,
+ "eval_runtime": 60.5299,
+ "eval_samples_per_second": 40.344,
+ "eval_steps_per_second": 0.644,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01396160558464,
+ "grad_norm": 0.05469869077205658,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 3.6253180503845215,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02792321116929,
+ "grad_norm": 0.05455945432186127,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 3.651854991912842,
+ "step": 2450
+ },
+ {
+ "epoch": 34.04188481675393,
+ "grad_norm": 0.05369517207145691,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 3.630152702331543,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05584642233857,
+ "grad_norm": 0.05498119816184044,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 3.6293938159942627,
+ "step": 2452
+ },
+ {
+ "epoch": 34.06980802792321,
+ "grad_norm": 0.05530959740281105,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 3.6548690795898438,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083769633507856,
+ "grad_norm": 0.055240705609321594,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 3.6408982276916504,
+ "step": 2454
+ },
+ {
+ "epoch": 34.0977312390925,
+ "grad_norm": 0.05358482152223587,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 3.6274750232696533,
+ "step": 2455
+ },
+ {
+ "epoch": 34.11169284467714,
+ "grad_norm": 0.05580627918243408,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 3.6357412338256836,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12565445026178,
+ "grad_norm": 0.05458527058362961,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 3.6637980937957764,
+ "step": 2457
+ },
+ {
+ "epoch": 34.139616055846425,
+ "grad_norm": 0.05548330023884773,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 3.627671718597412,
+ "step": 2458
+ },
+ {
+ "epoch": 34.153577661431065,
+ "grad_norm": 0.05482204258441925,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 3.642627239227295,
+ "step": 2459
+ },
+ {
+ "epoch": 34.167539267015705,
+ "grad_norm": 0.05435148626565933,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 3.6398377418518066,
+ "step": 2460
+ },
+ {
+ "epoch": 34.181500872600346,
+ "grad_norm": 0.05556599423289299,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 3.62994384765625,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19546247818499,
+ "grad_norm": 0.05394179746508598,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 3.593183994293213,
+ "step": 2462
+ },
+ {
+ "epoch": 34.20942408376963,
+ "grad_norm": 0.05472215637564659,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 3.645197868347168,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223385689354274,
+ "grad_norm": 0.05505116656422615,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 3.655001163482666,
+ "step": 2464
+ },
+ {
+ "epoch": 34.23734729493892,
+ "grad_norm": 0.05473910644650459,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 3.636812210083008,
+ "step": 2465
+ },
+ {
+ "epoch": 34.25130890052356,
+ "grad_norm": 0.05537406727671623,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 3.620631217956543,
+ "step": 2466
+ },
+ {
+ "epoch": 34.2652705061082,
+ "grad_norm": 0.054202884435653687,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 3.6500911712646484,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27923211169284,
+ "grad_norm": 0.0536428838968277,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 3.6268038749694824,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29319371727749,
+ "grad_norm": 0.053046368062496185,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 3.646955966949463,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30715532286213,
+ "grad_norm": 0.05279967561364174,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 3.6142497062683105,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32111692844677,
+ "grad_norm": 0.05348867550492287,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 3.643517017364502,
+ "step": 2471
+ },
+ {
+ "epoch": 34.33507853403141,
+ "grad_norm": 0.05406436324119568,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 3.6528172492980957,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34904013961606,
+ "grad_norm": 0.05249491706490517,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 3.6664018630981445,
+ "step": 2473
+ },
+ {
+ "epoch": 34.3630017452007,
+ "grad_norm": 0.053803686052560806,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 3.609346389770508,
+ "step": 2474
+ },
+ {
+ "epoch": 34.37696335078534,
+ "grad_norm": 0.054445188492536545,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 3.6647889614105225,
+ "step": 2475
+ },
+ {
+ "epoch": 34.390924956369986,
+ "grad_norm": 0.0550406239926815,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 3.656118154525757,
+ "step": 2476
+ },
+ {
+ "epoch": 34.404886561954626,
+ "grad_norm": 0.054776549339294434,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 3.6493523120880127,
+ "step": 2477
+ },
+ {
+ "epoch": 34.41884816753927,
+ "grad_norm": 0.054829373955726624,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 3.634632110595703,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43280977312391,
+ "grad_norm": 0.054575446993112564,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 3.636216878890991,
+ "step": 2479
+ },
+ {
+ "epoch": 34.446771378708554,
+ "grad_norm": 0.056146737188100815,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 3.6584529876708984,
+ "step": 2480
+ },
+ {
+ "epoch": 34.460732984293195,
+ "grad_norm": 0.054963648319244385,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 3.6363425254821777,
+ "step": 2481
+ },
+ {
+ "epoch": 34.474694589877835,
+ "grad_norm": 0.05659656599164009,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 3.6401073932647705,
+ "step": 2482
+ },
+ {
+ "epoch": 34.488656195462475,
+ "grad_norm": 0.05448099970817566,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 3.6161837577819824,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50261780104712,
+ "grad_norm": 0.057917334139347076,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 3.6802477836608887,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51657940663176,
+ "grad_norm": 0.056171346455812454,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 3.652339220046997,
+ "step": 2485
+ },
+ {
+ "epoch": 34.5305410122164,
+ "grad_norm": 0.05800231546163559,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 3.6437389850616455,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544502617801044,
+ "grad_norm": 0.05693773180246353,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 3.6268343925476074,
+ "step": 2487
+ },
+ {
+ "epoch": 34.55846422338569,
+ "grad_norm": 0.05424632877111435,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 3.6246471405029297,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57242582897033,
+ "grad_norm": 0.05645301938056946,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 3.649033308029175,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58638743455497,
+ "grad_norm": 0.056913044303655624,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 3.658982515335083,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60034904013962,
+ "grad_norm": 0.05770258232951164,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 3.621516704559326,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61431064572426,
+ "grad_norm": 0.05483509227633476,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 3.659385919570923,
+ "step": 2492
+ },
+ {
+ "epoch": 34.6282722513089,
+ "grad_norm": 0.057591512799263,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 3.6512298583984375,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64223385689354,
+ "grad_norm": 0.055053744465112686,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 3.630310297012329,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65619546247819,
+ "grad_norm": 0.056070517748594284,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 3.6766881942749023,
+ "step": 2495
+ },
+ {
+ "epoch": 34.67015706806283,
+ "grad_norm": 0.056556008756160736,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 3.6454544067382812,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68411867364747,
+ "grad_norm": 0.05624367296695709,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 3.6482372283935547,
+ "step": 2497
+ },
+ {
+ "epoch": 34.69808027923211,
+ "grad_norm": 0.056266799569129944,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 3.6528916358947754,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712041884816756,
+ "grad_norm": 0.05540672317147255,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 3.636371374130249,
+ "step": 2499
+ },
+ {
+ "epoch": 34.726003490401396,
+ "grad_norm": 0.05527781695127487,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 3.6440672874450684,
+ "step": 2500
+ },
+ {
+ "epoch": 34.73996509598604,
+ "grad_norm": 0.05645914003252983,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 3.6747167110443115,
+ "step": 2501
+ },
+ {
+ "epoch": 34.753926701570684,
+ "grad_norm": 0.055584684014320374,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 3.637603282928467,
+ "step": 2502
+ },
+ {
+ "epoch": 34.767888307155324,
+ "grad_norm": 0.0554841123521328,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 3.6546664237976074,
+ "step": 2503
+ },
+ {
+ "epoch": 34.781849912739965,
+ "grad_norm": 0.0557134784758091,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 3.6366615295410156,
+ "step": 2504
+ },
+ {
+ "epoch": 34.795811518324605,
+ "grad_norm": 0.056569263339042664,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 3.657125949859619,
+ "step": 2505
+ },
+ {
+ "epoch": 34.80977312390925,
+ "grad_norm": 0.05737468600273132,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 3.6409783363342285,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82373472949389,
+ "grad_norm": 0.05602499470114708,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 3.6670620441436768,
+ "step": 2507
+ },
+ {
+ "epoch": 34.83769633507853,
+ "grad_norm": 0.059535399079322815,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 3.6585612297058105,
+ "step": 2508
+ },
+ {
+ "epoch": 34.85165794066317,
+ "grad_norm": 0.05610577017068863,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 3.6544179916381836,
+ "step": 2509
+ },
+ {
+ "epoch": 34.86561954624782,
+ "grad_norm": 0.057512786239385605,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 3.658529281616211,
+ "step": 2510
+ },
+ {
+ "epoch": 34.87958115183246,
+ "grad_norm": 0.05605233088135719,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 3.622344493865967,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8935427574171,
+ "grad_norm": 0.055069468915462494,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 3.6812562942504883,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90750436300174,
+ "grad_norm": 0.05756533145904541,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 3.651491165161133,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92146596858639,
+ "grad_norm": 0.05596763640642166,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 3.633408784866333,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93542757417103,
+ "grad_norm": 0.057164326310157776,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 3.632655620574951,
+ "step": 2515
+ },
+ {
+ "epoch": 34.94938917975567,
+ "grad_norm": 0.059363316744565964,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 3.637542486190796,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96335078534032,
+ "grad_norm": 0.05572650581598282,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 3.666663885116577,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97731239092496,
+ "grad_norm": 0.05825378745794296,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 3.654174327850342,
+ "step": 2518
+ },
+ {
+ "epoch": 34.9912739965096,
+ "grad_norm": 0.058226607739925385,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 3.630509376525879,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.041674062609672546,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 2.2687463760375977,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 0.6132010817527771,
+ "eval_runtime": 58.9507,
+ "eval_samples_per_second": 41.424,
+ "eval_steps_per_second": 0.662,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01396160558464,
+ "grad_norm": 0.053901202976703644,
+ "learning_rate": 0.0001349604788260823,
+ "loss": 3.6262593269348145,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02792321116929,
+ "grad_norm": 0.051058750599622726,
+ "learning_rate": 0.00013474073268599152,
+ "loss": 3.596529483795166,
+ "step": 2522
+ },
+ {
+ "epoch": 35.04188481675393,
+ "grad_norm": 0.05482964590191841,
+ "learning_rate": 0.00013452112210935339,
+ "loss": 3.6197128295898438,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05584642233857,
+ "grad_norm": 0.05166293680667877,
+ "learning_rate": 0.00013430164727310399,
+ "loss": 3.6150989532470703,
+ "step": 2524
+ },
+ {
+ "epoch": 35.06980802792321,
+ "grad_norm": 0.051854558289051056,
+ "learning_rate": 0.00013408230835407027,
+ "loss": 3.594954013824463,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083769633507856,
+ "grad_norm": 0.05247301980853081,
+ "learning_rate": 0.00013386310552896977,
+ "loss": 3.6349754333496094,
+ "step": 2526
+ },
+ {
+ "epoch": 35.0977312390925,
+ "grad_norm": 0.05379469692707062,
+ "learning_rate": 0.00013364403897441016,
+ "loss": 3.611581802368164,
+ "step": 2527
+ },
+ {
+ "epoch": 35.11169284467714,
+ "grad_norm": 0.053294215351343155,
+ "learning_rate": 0.00013342510886688944,
+ "loss": 3.5933470726013184,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12565445026178,
+ "grad_norm": 0.05274510383605957,
+ "learning_rate": 0.00013320631538279548,
+ "loss": 3.6055941581726074,
+ "step": 2529
+ },
+ {
+ "epoch": 35.139616055846425,
+ "grad_norm": 0.05449166148900986,
+ "learning_rate": 0.00013298765869840658,
+ "loss": 3.5891900062561035,
+ "step": 2530
+ },
+ {
+ "epoch": 35.153577661431065,
+ "grad_norm": 0.05245492234826088,
+ "learning_rate": 0.00013276913898989013,
+ "loss": 3.614109992980957,
+ "step": 2531
+ },
+ {
+ "epoch": 35.167539267015705,
+ "grad_norm": 0.05323590710759163,
+ "learning_rate": 0.00013255075643330369,
+ "loss": 3.6126725673675537,
+ "step": 2532
+ },
+ {
+ "epoch": 35.181500872600346,
+ "grad_norm": 0.05381206050515175,
+ "learning_rate": 0.00013233251120459407,
+ "loss": 3.5820865631103516,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19546247818499,
+ "grad_norm": 0.054380469024181366,
+ "learning_rate": 0.0001321144034795976,
+ "loss": 3.6311044692993164,
+ "step": 2534
+ },
+ {
+ "epoch": 35.20942408376963,
+ "grad_norm": 0.053128909319639206,
+ "learning_rate": 0.00013189643343403977,
+ "loss": 3.5665154457092285,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223385689354274,
+ "grad_norm": 0.052764274179935455,
+ "learning_rate": 0.00013167860124353476,
+ "loss": 3.609287738800049,
+ "step": 2536
+ },
+ {
+ "epoch": 35.23734729493892,
+ "grad_norm": 0.05363981053233147,
+ "learning_rate": 0.00013146090708358657,
+ "loss": 3.6260581016540527,
+ "step": 2537
+ },
+ {
+ "epoch": 35.25130890052356,
+ "grad_norm": 0.05316983535885811,
+ "learning_rate": 0.00013124335112958704,
+ "loss": 3.589477300643921,
+ "step": 2538
+ },
+ {
+ "epoch": 35.2652705061082,
+ "grad_norm": 0.05451323837041855,
+ "learning_rate": 0.0001310259335568172,
+ "loss": 3.621201515197754,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27923211169284,
+ "grad_norm": 0.053846996277570724,
+ "learning_rate": 0.00013080865454044647,
+ "loss": 3.610521078109741,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29319371727749,
+ "grad_norm": 0.05316411331295967,
+ "learning_rate": 0.00013059151425553263,
+ "loss": 3.6240267753601074,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30715532286213,
+ "grad_norm": 0.053650014102458954,
+ "learning_rate": 0.00013037451287702183,
+ "loss": 3.6098155975341797,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32111692844677,
+ "grad_norm": 0.054066307842731476,
+ "learning_rate": 0.0001301576505797479,
+ "loss": 3.619396686553955,
+ "step": 2543
+ },
+ {
+ "epoch": 35.33507853403141,
+ "grad_norm": 0.05257592722773552,
+ "learning_rate": 0.00012994092753843293,
+ "loss": 3.655679225921631,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34904013961606,
+ "grad_norm": 0.05437186732888222,
+ "learning_rate": 0.00012972434392768687,
+ "loss": 3.6167373657226562,
+ "step": 2545
+ },
+ {
+ "epoch": 35.3630017452007,
+ "grad_norm": 0.05342245474457741,
+ "learning_rate": 0.00012950789992200714,
+ "loss": 3.5970587730407715,
+ "step": 2546
+ },
+ {
+ "epoch": 35.37696335078534,
+ "grad_norm": 0.054295122623443604,
+ "learning_rate": 0.00012929159569577886,
+ "loss": 3.5975115299224854,
+ "step": 2547
+ },
+ {
+ "epoch": 35.390924956369986,
+ "grad_norm": 0.05349617451429367,
+ "learning_rate": 0.0001290754314232743,
+ "loss": 3.603315591812134,
+ "step": 2548
+ },
+ {
+ "epoch": 35.404886561954626,
+ "grad_norm": 0.05614254251122475,
+ "learning_rate": 0.00012885940727865334,
+ "loss": 3.6229186058044434,
+ "step": 2549
+ },
+ {
+ "epoch": 35.41884816753927,
+ "grad_norm": 0.054123859852552414,
+ "learning_rate": 0.00012864352343596247,
+ "loss": 3.641162395477295,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43280977312391,
+ "grad_norm": 0.05414072424173355,
+ "learning_rate": 0.0001284277800691355,
+ "loss": 3.614049196243286,
+ "step": 2551
+ },
+ {
+ "epoch": 35.446771378708554,
+ "grad_norm": 0.05422188341617584,
+ "learning_rate": 0.00012821217735199298,
+ "loss": 3.5817785263061523,
+ "step": 2552
+ },
+ {
+ "epoch": 35.460732984293195,
+ "grad_norm": 0.05528900399804115,
+ "learning_rate": 0.00012799671545824212,
+ "loss": 3.618821859359741,
+ "step": 2553
+ },
+ {
+ "epoch": 35.474694589877835,
+ "grad_norm": 0.0527317114174366,
+ "learning_rate": 0.0001277813945614768,
+ "loss": 3.6165270805358887,
+ "step": 2554
+ },
+ {
+ "epoch": 35.488656195462475,
+ "grad_norm": 0.05585351586341858,
+ "learning_rate": 0.00012756621483517682,
+ "loss": 3.6158246994018555,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50261780104712,
+ "grad_norm": 0.05375446751713753,
+ "learning_rate": 0.00012735117645270905,
+ "loss": 3.610105037689209,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51657940663176,
+ "grad_norm": 0.055324699729681015,
+ "learning_rate": 0.00012713627958732567,
+ "loss": 3.6106624603271484,
+ "step": 2557
+ },
+ {
+ "epoch": 35.5305410122164,
+ "grad_norm": 0.0544339083135128,
+ "learning_rate": 0.00012692152441216539,
+ "loss": 3.605295181274414,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544502617801044,
+ "grad_norm": 0.05522064492106438,
+ "learning_rate": 0.0001267069111002525,
+ "loss": 3.607207775115967,
+ "step": 2559
+ },
+ {
+ "epoch": 35.55846422338569,
+ "grad_norm": 0.055271610617637634,
+ "learning_rate": 0.00012649243982449718,
+ "loss": 3.610219955444336,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57242582897033,
+ "grad_norm": 0.05532364174723625,
+ "learning_rate": 0.00012627811075769502,
+ "loss": 3.6226587295532227,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58638743455497,
+ "grad_norm": 0.05356820300221443,
+ "learning_rate": 0.00012606392407252687,
+ "loss": 3.628182888031006,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60034904013962,
+ "grad_norm": 0.054853104054927826,
+ "learning_rate": 0.00012584987994155943,
+ "loss": 3.596072196960449,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61431064572426,
+ "grad_norm": 0.05462052673101425,
+ "learning_rate": 0.00012563597853724388,
+ "loss": 3.6201977729797363,
+ "step": 2564
+ },
+ {
+ "epoch": 35.6282722513089,
+ "grad_norm": 0.05532895773649216,
+ "learning_rate": 0.000125422220031917,
+ "loss": 3.642716407775879,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64223385689354,
+ "grad_norm": 0.055425263941287994,
+ "learning_rate": 0.00012520860459779972,
+ "loss": 3.5959243774414062,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65619546247819,
+ "grad_norm": 0.056545358151197433,
+ "learning_rate": 0.0001249951324069986,
+ "loss": 3.5796356201171875,
+ "step": 2567
+ },
+ {
+ "epoch": 35.67015706806283,
+ "grad_norm": 0.055357806384563446,
+ "learning_rate": 0.0001247818036315042,
+ "loss": 3.6106321811676025,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68411867364747,
+ "grad_norm": 0.05461106821894646,
+ "learning_rate": 0.00012456861844319155,
+ "loss": 3.591688632965088,
+ "step": 2569
+ },
+ {
+ "epoch": 35.69808027923211,
+ "grad_norm": 0.055133529007434845,
+ "learning_rate": 0.00012435557701382012,
+ "loss": 3.6243996620178223,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712041884816756,
+ "grad_norm": 0.055249396711587906,
+ "learning_rate": 0.0001241426795150336,
+ "loss": 3.608017683029175,
+ "step": 2571
+ },
+ {
+ "epoch": 35.726003490401396,
+ "grad_norm": 0.056398529559373856,
+ "learning_rate": 0.00012392992611835973,
+ "loss": 3.6282615661621094,
+ "step": 2572
+ },
+ {
+ "epoch": 35.73996509598604,
+ "grad_norm": 0.0552498884499073,
+ "learning_rate": 0.0001237173169952098,
+ "loss": 3.6407179832458496,
+ "step": 2573
+ },
+ {
+ "epoch": 35.753926701570684,
+ "grad_norm": 0.056712064892053604,
+ "learning_rate": 0.00012350485231687954,
+ "loss": 3.616793632507324,
+ "step": 2574
+ },
+ {
+ "epoch": 35.767888307155324,
+ "grad_norm": 0.055920615792274475,
+ "learning_rate": 0.0001232925322545476,
+ "loss": 3.6330060958862305,
+ "step": 2575
+ },
+ {
+ "epoch": 35.781849912739965,
+ "grad_norm": 0.05519915744662285,
+ "learning_rate": 0.0001230803569792765,
+ "loss": 3.61584210395813,
+ "step": 2576
+ },
+ {
+ "epoch": 35.795811518324605,
+ "grad_norm": 0.05472308024764061,
+ "learning_rate": 0.0001228683266620121,
+ "loss": 3.6116104125976562,
+ "step": 2577
+ },
+ {
+ "epoch": 35.80977312390925,
+ "grad_norm": 0.0556313656270504,
+ "learning_rate": 0.00012265644147358326,
+ "loss": 3.634107828140259,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82373472949389,
+ "grad_norm": 0.057370636612176895,
+ "learning_rate": 0.00012244470158470226,
+ "loss": 3.6127190589904785,
+ "step": 2579
+ },
+ {
+ "epoch": 35.83769633507853,
+ "grad_norm": 0.05609549954533577,
+ "learning_rate": 0.00012223310716596387,
+ "loss": 3.6323940753936768,
+ "step": 2580
+ },
+ {
+ "epoch": 35.85165794066317,
+ "grad_norm": 0.054692406207323074,
+ "learning_rate": 0.00012202165838784601,
+ "loss": 3.6339163780212402,
+ "step": 2581
+ },
+ {
+ "epoch": 35.86561954624782,
+ "grad_norm": 0.05703910440206528,
+ "learning_rate": 0.00012181035542070913,
+ "loss": 3.6473350524902344,
+ "step": 2582
+ },
+ {
+ "epoch": 35.87958115183246,
+ "grad_norm": 0.055098868906497955,
+ "learning_rate": 0.00012159919843479617,
+ "loss": 3.6110334396362305,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8935427574171,
+ "grad_norm": 0.0570925697684288,
+ "learning_rate": 0.00012138818760023257,
+ "loss": 3.630516767501831,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90750436300174,
+ "grad_norm": 0.05486847832798958,
+ "learning_rate": 0.00012117732308702592,
+ "loss": 3.622447967529297,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92146596858639,
+ "grad_norm": 0.057239681482315063,
+ "learning_rate": 0.00012096660506506605,
+ "loss": 3.6436092853546143,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93542757417103,
+ "grad_norm": 0.05756063386797905,
+ "learning_rate": 0.00012075603370412443,
+ "loss": 3.6313462257385254,
+ "step": 2587
+ },
+ {
+ "epoch": 35.94938917975567,
+ "grad_norm": 0.05791015550494194,
+ "learning_rate": 0.00012054560917385476,
+ "loss": 3.617950677871704,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96335078534032,
+ "grad_norm": 0.05538728088140488,
+ "learning_rate": 0.00012033533164379225,
+ "loss": 3.665337085723877,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97731239092496,
+ "grad_norm": 0.05843896046280861,
+ "learning_rate": 0.0001201252012833537,
+ "loss": 3.6643996238708496,
+ "step": 2590
+ },
+ {
+ "epoch": 35.9912739965096,
+ "grad_norm": 0.05609741806983948,
+ "learning_rate": 0.00011991521826183747,
+ "loss": 3.6212263107299805,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.04305510222911835,
+ "learning_rate": 0.00011970538274842282,
+ "loss": 2.270846128463745,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 0.6145148873329163,
+ "eval_runtime": 60.714,
+ "eval_samples_per_second": 40.221,
+ "eval_steps_per_second": 0.642,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01396160558464,
+ "grad_norm": 0.054062873125076294,
+ "learning_rate": 0.00011949569491217073,
+ "loss": 3.581265926361084,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02792321116929,
+ "grad_norm": 0.051643744111061096,
+ "learning_rate": 0.00011928615492202269,
+ "loss": 3.586836814880371,
+ "step": 2594
+ },
+ {
+ "epoch": 36.04188481675393,
+ "grad_norm": 0.054067034274339676,
+ "learning_rate": 0.0001190767629468014,
+ "loss": 3.6008591651916504,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05584642233857,
+ "grad_norm": 0.05220582336187363,
+ "learning_rate": 0.00011886751915521007,
+ "loss": 3.5975303649902344,
+ "step": 2596
+ },
+ {
+ "epoch": 36.06980802792321,
+ "grad_norm": 0.053417500108480453,
+ "learning_rate": 0.00011865842371583278,
+ "loss": 3.5652084350585938,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083769633507856,
+ "grad_norm": 0.05344022065401077,
+ "learning_rate": 0.00011844947679713396,
+ "loss": 3.561967372894287,
+ "step": 2598
+ },
+ {
+ "epoch": 36.0977312390925,
+ "grad_norm": 0.05316327139735222,
+ "learning_rate": 0.00011824067856745812,
+ "loss": 3.590935707092285,
+ "step": 2599
+ },
+ {
+ "epoch": 36.11169284467714,
+ "grad_norm": 0.054267000406980515,
+ "learning_rate": 0.00011803202919503064,
+ "loss": 3.5788426399230957,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12565445026178,
+ "grad_norm": 0.05376334860920906,
+ "learning_rate": 0.00011782352884795615,
+ "loss": 3.589162826538086,
+ "step": 2601
+ },
+ {
+ "epoch": 36.139616055846425,
+ "grad_norm": 0.05369187891483307,
+ "learning_rate": 0.00011761517769421983,
+ "loss": 3.558194160461426,
+ "step": 2602
+ },
+ {
+ "epoch": 36.153577661431065,
+ "grad_norm": 0.05407107621431351,
+ "learning_rate": 0.00011740697590168635,
+ "loss": 3.571460247039795,
+ "step": 2603
+ },
+ {
+ "epoch": 36.167539267015705,
+ "grad_norm": 0.05279739201068878,
+ "learning_rate": 0.00011719892363810018,
+ "loss": 3.588043689727783,
+ "step": 2604
+ },
+ {
+ "epoch": 36.181500872600346,
+ "grad_norm": 0.05608730763196945,
+ "learning_rate": 0.00011699102107108539,
+ "loss": 3.5704660415649414,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19546247818499,
+ "grad_norm": 0.05290020629763603,
+ "learning_rate": 0.00011678326836814507,
+ "loss": 3.5612759590148926,
+ "step": 2606
+ },
+ {
+ "epoch": 36.20942408376963,
+ "grad_norm": 0.05398614704608917,
+ "learning_rate": 0.00011657566569666198,
+ "loss": 3.579409122467041,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223385689354274,
+ "grad_norm": 0.051791902631521225,
+ "learning_rate": 0.00011636821322389777,
+ "loss": 3.5976991653442383,
+ "step": 2608
+ },
+ {
+ "epoch": 36.23734729493892,
+ "grad_norm": 0.05329076945781708,
+ "learning_rate": 0.00011616091111699333,
+ "loss": 3.6209723949432373,
+ "step": 2609
+ },
+ {
+ "epoch": 36.25130890052356,
+ "grad_norm": 0.052824556827545166,
+ "learning_rate": 0.00011595375954296786,
+ "loss": 3.5926363468170166,
+ "step": 2610
+ },
+ {
+ "epoch": 36.2652705061082,
+ "grad_norm": 0.05411917343735695,
+ "learning_rate": 0.00011574675866871997,
+ "loss": 3.607084274291992,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27923211169284,
+ "grad_norm": 0.0527656190097332,
+ "learning_rate": 0.00011553990866102651,
+ "loss": 3.579601526260376,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29319371727749,
+ "grad_norm": 0.05329589545726776,
+ "learning_rate": 0.00011533320968654265,
+ "loss": 3.5800671577453613,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30715532286213,
+ "grad_norm": 0.05175339803099632,
+ "learning_rate": 0.00011512666191180204,
+ "loss": 3.594729423522949,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32111692844677,
+ "grad_norm": 0.05224638432264328,
+ "learning_rate": 0.00011492026550321653,
+ "loss": 3.567016839981079,
+ "step": 2615
+ },
+ {
+ "epoch": 36.33507853403141,
+ "grad_norm": 0.053442589938640594,
+ "learning_rate": 0.00011471402062707607,
+ "loss": 3.5964465141296387,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34904013961606,
+ "grad_norm": 0.05282003805041313,
+ "learning_rate": 0.00011450792744954827,
+ "loss": 3.5932140350341797,
+ "step": 2617
+ },
+ {
+ "epoch": 36.3630017452007,
+ "grad_norm": 0.054600607603788376,
+ "learning_rate": 0.00011430198613667861,
+ "loss": 3.5916247367858887,
+ "step": 2618
+ },
+ {
+ "epoch": 36.37696335078534,
+ "grad_norm": 0.053483184427022934,
+ "learning_rate": 0.00011409619685439064,
+ "loss": 3.588195323944092,
+ "step": 2619
+ },
+ {
+ "epoch": 36.390924956369986,
+ "grad_norm": 0.052354127168655396,
+ "learning_rate": 0.00011389055976848477,
+ "loss": 3.6031956672668457,
+ "step": 2620
+ },
+ {
+ "epoch": 36.404886561954626,
+ "grad_norm": 0.05411195009946823,
+ "learning_rate": 0.00011368507504463914,
+ "loss": 3.61525559425354,
+ "step": 2621
+ },
+ {
+ "epoch": 36.41884816753927,
+ "grad_norm": 0.055229391902685165,
+ "learning_rate": 0.00011347974284840914,
+ "loss": 3.58837890625,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43280977312391,
+ "grad_norm": 0.05329221487045288,
+ "learning_rate": 0.00011327456334522725,
+ "loss": 3.565725803375244,
+ "step": 2623
+ },
+ {
+ "epoch": 36.446771378708554,
+ "grad_norm": 0.05464653670787811,
+ "learning_rate": 0.00011306953670040275,
+ "loss": 3.6044445037841797,
+ "step": 2624
+ },
+ {
+ "epoch": 36.460732984293195,
+ "grad_norm": 0.05323687940835953,
+ "learning_rate": 0.00011286466307912184,
+ "loss": 3.5857181549072266,
+ "step": 2625
+ },
+ {
+ "epoch": 36.474694589877835,
+ "grad_norm": 0.054866403341293335,
+ "learning_rate": 0.00011265994264644783,
+ "loss": 3.592473030090332,
+ "step": 2626
+ },
+ {
+ "epoch": 36.488656195462475,
+ "grad_norm": 0.05508954077959061,
+ "learning_rate": 0.00011245537556732001,
+ "loss": 3.5929830074310303,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50261780104712,
+ "grad_norm": 0.057712286710739136,
+ "learning_rate": 0.00011225096200655442,
+ "loss": 3.613492488861084,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51657940663176,
+ "grad_norm": 0.05367693305015564,
+ "learning_rate": 0.00011204670212884342,
+ "loss": 3.5849246978759766,
+ "step": 2629
+ },
+ {
+ "epoch": 36.5305410122164,
+ "grad_norm": 0.055529702454805374,
+ "learning_rate": 0.00011184259609875557,
+ "loss": 3.60062313079834,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544502617801044,
+ "grad_norm": 0.05650486797094345,
+ "learning_rate": 0.00011163864408073522,
+ "loss": 3.570746898651123,
+ "step": 2631
+ },
+ {
+ "epoch": 36.55846422338569,
+ "grad_norm": 0.055417656898498535,
+ "learning_rate": 0.00011143484623910293,
+ "loss": 3.5839993953704834,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57242582897033,
+ "grad_norm": 0.055162668228149414,
+ "learning_rate": 0.00011123120273805496,
+ "loss": 3.5858545303344727,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58638743455497,
+ "grad_norm": 0.054991088807582855,
+ "learning_rate": 0.0001110277137416632,
+ "loss": 3.5967633724212646,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60034904013962,
+ "grad_norm": 0.055452100932598114,
+ "learning_rate": 0.00011082437941387512,
+ "loss": 3.6110477447509766,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61431064572426,
+ "grad_norm": 0.05554793030023575,
+ "learning_rate": 0.00011062119991851322,
+ "loss": 3.5871753692626953,
+ "step": 2636
+ },
+ {
+ "epoch": 36.6282722513089,
+ "grad_norm": 0.054689038544893265,
+ "learning_rate": 0.000110418175419276,
+ "loss": 3.590219020843506,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64223385689354,
+ "grad_norm": 0.05510031804442406,
+ "learning_rate": 0.00011021530607973626,
+ "loss": 3.5875062942504883,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65619546247819,
+ "grad_norm": 0.05421333760023117,
+ "learning_rate": 0.00011001259206334235,
+ "loss": 3.602912187576294,
+ "step": 2639
+ },
+ {
+ "epoch": 36.67015706806283,
+ "grad_norm": 0.05532202124595642,
+ "learning_rate": 0.00010981003353341721,
+ "loss": 3.5950570106506348,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68411867364747,
+ "grad_norm": 0.05355961248278618,
+ "learning_rate": 0.00010960763065315864,
+ "loss": 3.5744097232818604,
+ "step": 2641
+ },
+ {
+ "epoch": 36.69808027923211,
+ "grad_norm": 0.05564635619521141,
+ "learning_rate": 0.00010940538358563906,
+ "loss": 3.5986971855163574,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712041884816756,
+ "grad_norm": 0.054115377366542816,
+ "learning_rate": 0.00010920329249380509,
+ "loss": 3.596283197402954,
+ "step": 2643
+ },
+ {
+ "epoch": 36.726003490401396,
+ "grad_norm": 0.05481656640768051,
+ "learning_rate": 0.00010900135754047799,
+ "loss": 3.5836915969848633,
+ "step": 2644
+ },
+ {
+ "epoch": 36.73996509598604,
+ "grad_norm": 0.05519521236419678,
+ "learning_rate": 0.00010879957888835307,
+ "loss": 3.6295974254608154,
+ "step": 2645
+ },
+ {
+ "epoch": 36.753926701570684,
+ "grad_norm": 0.05460044741630554,
+ "learning_rate": 0.00010859795669999981,
+ "loss": 3.5932111740112305,
+ "step": 2646
+ },
+ {
+ "epoch": 36.767888307155324,
+ "grad_norm": 0.054085709154605865,
+ "learning_rate": 0.00010839649113786137,
+ "loss": 3.584069013595581,
+ "step": 2647
+ },
+ {
+ "epoch": 36.781849912739965,
+ "grad_norm": 0.05407894030213356,
+ "learning_rate": 0.00010819518236425514,
+ "loss": 3.5967845916748047,
+ "step": 2648
+ },
+ {
+ "epoch": 36.795811518324605,
+ "grad_norm": 0.05513349175453186,
+ "learning_rate": 0.00010799403054137197,
+ "loss": 3.598283290863037,
+ "step": 2649
+ },
+ {
+ "epoch": 36.80977312390925,
+ "grad_norm": 0.0558769516646862,
+ "learning_rate": 0.00010779303583127609,
+ "loss": 3.60675048828125,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82373472949389,
+ "grad_norm": 0.05440812557935715,
+ "learning_rate": 0.00010759219839590535,
+ "loss": 3.5989885330200195,
+ "step": 2651
+ },
+ {
+ "epoch": 36.83769633507853,
+ "grad_norm": 0.05533481389284134,
+ "learning_rate": 0.00010739151839707089,
+ "loss": 3.6296310424804688,
+ "step": 2652
+ },
+ {
+ "epoch": 36.85165794066317,
+ "grad_norm": 0.055396221578121185,
+ "learning_rate": 0.00010719099599645706,
+ "loss": 3.6243319511413574,
+ "step": 2653
+ },
+ {
+ "epoch": 36.86561954624782,
+ "grad_norm": 0.05608600378036499,
+ "learning_rate": 0.0001069906313556208,
+ "loss": 3.6124401092529297,
+ "step": 2654
+ },
+ {
+ "epoch": 36.87958115183246,
+ "grad_norm": 0.05493833124637604,
+ "learning_rate": 0.00010679042463599263,
+ "loss": 3.6008474826812744,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8935427574171,
+ "grad_norm": 0.055183470249176025,
+ "learning_rate": 0.00010659037599887545,
+ "loss": 3.5892064571380615,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90750436300174,
+ "grad_norm": 0.05589289590716362,
+ "learning_rate": 0.0001063904856054446,
+ "loss": 3.594189405441284,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92146596858639,
+ "grad_norm": 0.05483611673116684,
+ "learning_rate": 0.00010619075361674836,
+ "loss": 3.586012601852417,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93542757417103,
+ "grad_norm": 0.055471133440732956,
+ "learning_rate": 0.0001059911801937071,
+ "loss": 3.602313995361328,
+ "step": 2659
+ },
+ {
+ "epoch": 36.94938917975567,
+ "grad_norm": 0.0559491291642189,
+ "learning_rate": 0.00010579176549711365,
+ "loss": 3.6010072231292725,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96335078534032,
+ "grad_norm": 0.055661238729953766,
+ "learning_rate": 0.00010559250968763265,
+ "loss": 3.6093027591705322,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97731239092496,
+ "grad_norm": 0.05550997704267502,
+ "learning_rate": 0.00010539341292580086,
+ "loss": 3.614647150039673,
+ "step": 2662
+ },
+ {
+ "epoch": 36.9912739965096,
+ "grad_norm": 0.056113287806510925,
+ "learning_rate": 0.00010519447537202729,
+ "loss": 3.5756256580352783,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.04090592637658119,
+ "learning_rate": 0.00010499569718659201,
+ "loss": 2.237126588821411,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 0.6159548759460449,
+ "eval_runtime": 60.4892,
+ "eval_samples_per_second": 40.371,
+ "eval_steps_per_second": 0.645,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01396160558464,
+ "grad_norm": 0.05119108036160469,
+ "learning_rate": 0.00010479707852964713,
+ "loss": 3.5844168663024902,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02792321116929,
+ "grad_norm": 0.050482187420129776,
+ "learning_rate": 0.00010459861956121611,
+ "loss": 3.5596923828125,
+ "step": 2666
+ },
+ {
+ "epoch": 37.04188481675393,
+ "grad_norm": 0.05054338648915291,
+ "learning_rate": 0.00010440032044119383,
+ "loss": 3.533599376678467,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05584642233857,
+ "grad_norm": 0.050228655338287354,
+ "learning_rate": 0.00010420218132934614,
+ "loss": 3.557422637939453,
+ "step": 2668
+ },
+ {
+ "epoch": 37.06980802792321,
+ "grad_norm": 0.04899684339761734,
+ "learning_rate": 0.00010400420238531023,
+ "loss": 3.5452675819396973,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083769633507856,
+ "grad_norm": 0.049607474356889725,
+ "learning_rate": 0.00010380638376859416,
+ "loss": 3.571603298187256,
+ "step": 2670
+ },
+ {
+ "epoch": 37.0977312390925,
+ "grad_norm": 0.04938557744026184,
+ "learning_rate": 0.00010360872563857682,
+ "loss": 3.5608606338500977,
+ "step": 2671
+ },
+ {
+ "epoch": 37.11169284467714,
+ "grad_norm": 0.05051542446017265,
+ "learning_rate": 0.0001034112281545079,
+ "loss": 3.5389974117279053,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12565445026178,
+ "grad_norm": 0.04959443211555481,
+ "learning_rate": 0.00010321389147550725,
+ "loss": 3.5671193599700928,
+ "step": 2673
+ },
+ {
+ "epoch": 37.139616055846425,
+ "grad_norm": 0.0502617284655571,
+ "learning_rate": 0.00010301671576056588,
+ "loss": 3.5699734687805176,
+ "step": 2674
+ },
+ {
+ "epoch": 37.153577661431065,
+ "grad_norm": 0.051576778292655945,
+ "learning_rate": 0.00010281970116854436,
+ "loss": 3.562732219696045,
+ "step": 2675
+ },
+ {
+ "epoch": 37.167539267015705,
+ "grad_norm": 0.05000219866633415,
+ "learning_rate": 0.00010262284785817392,
+ "loss": 3.545559883117676,
+ "step": 2676
+ },
+ {
+ "epoch": 37.181500872600346,
+ "grad_norm": 0.049805279821157455,
+ "learning_rate": 0.00010242615598805574,
+ "loss": 3.5819478034973145,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19546247818499,
+ "grad_norm": 0.05044523999094963,
+ "learning_rate": 0.00010222962571666088,
+ "loss": 3.57983136177063,
+ "step": 2678
+ },
+ {
+ "epoch": 37.20942408376963,
+ "grad_norm": 0.05141332373023033,
+ "learning_rate": 0.00010203325720233032,
+ "loss": 3.5591063499450684,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223385689354274,
+ "grad_norm": 0.05136275291442871,
+ "learning_rate": 0.00010183705060327433,
+ "loss": 3.5776171684265137,
+ "step": 2680
+ },
+ {
+ "epoch": 37.23734729493892,
+ "grad_norm": 0.05119960755109787,
+ "learning_rate": 0.00010164100607757346,
+ "loss": 3.5721652507781982,
+ "step": 2681
+ },
+ {
+ "epoch": 37.25130890052356,
+ "grad_norm": 0.05113738030195236,
+ "learning_rate": 0.00010144512378317687,
+ "loss": 3.5741496086120605,
+ "step": 2682
+ },
+ {
+ "epoch": 37.2652705061082,
+ "grad_norm": 0.05131804198026657,
+ "learning_rate": 0.00010124940387790354,
+ "loss": 3.577519416809082,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27923211169284,
+ "grad_norm": 0.05085485801100731,
+ "learning_rate": 0.00010105384651944148,
+ "loss": 3.5454201698303223,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29319371727749,
+ "grad_norm": 0.05160171166062355,
+ "learning_rate": 0.00010085845186534769,
+ "loss": 3.574143886566162,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30715532286213,
+ "grad_norm": 0.05132637917995453,
+ "learning_rate": 0.00010066322007304819,
+ "loss": 3.5619382858276367,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32111692844677,
+ "grad_norm": 0.05075863003730774,
+ "learning_rate": 0.00010046815129983757,
+ "loss": 3.5536012649536133,
+ "step": 2687
+ },
+ {
+ "epoch": 37.33507853403141,
+ "grad_norm": 0.052456241101026535,
+ "learning_rate": 0.00010027324570287925,
+ "loss": 3.586211681365967,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34904013961606,
+ "grad_norm": 0.051105350255966187,
+ "learning_rate": 0.00010007850343920519,
+ "loss": 3.5617852210998535,
+ "step": 2689
+ },
+ {
+ "epoch": 37.3630017452007,
+ "grad_norm": 0.052098311483860016,
+ "learning_rate": 9.988392466571572e-05,
+ "loss": 3.5693583488464355,
+ "step": 2690
+ },
+ {
+ "epoch": 37.37696335078534,
+ "grad_norm": 0.05069877579808235,
+ "learning_rate": 9.968950953917945e-05,
+ "loss": 3.547999620437622,
+ "step": 2691
+ },
+ {
+ "epoch": 37.390924956369986,
+ "grad_norm": 0.05264188349246979,
+ "learning_rate": 9.949525821623309e-05,
+ "loss": 3.554316520690918,
+ "step": 2692
+ },
+ {
+ "epoch": 37.404886561954626,
+ "grad_norm": 0.05224690958857536,
+ "learning_rate": 9.930117085338156e-05,
+ "loss": 3.531437397003174,
+ "step": 2693
+ },
+ {
+ "epoch": 37.41884816753927,
+ "grad_norm": 0.05134997144341469,
+ "learning_rate": 9.910724760699731e-05,
+ "loss": 3.5924055576324463,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43280977312391,
+ "grad_norm": 0.05177623778581619,
+ "learning_rate": 9.891348863332093e-05,
+ "loss": 3.580904006958008,
+ "step": 2695
+ },
+ {
+ "epoch": 37.446771378708554,
+ "grad_norm": 0.052150946110486984,
+ "learning_rate": 9.871989408846051e-05,
+ "loss": 3.568195343017578,
+ "step": 2696
+ },
+ {
+ "epoch": 37.460732984293195,
+ "grad_norm": 0.05220150575041771,
+ "learning_rate": 9.852646412839178e-05,
+ "loss": 3.5756545066833496,
+ "step": 2697
+ },
+ {
+ "epoch": 37.474694589877835,
+ "grad_norm": 0.053251419216394424,
+ "learning_rate": 9.833319890895756e-05,
+ "loss": 3.5686798095703125,
+ "step": 2698
+ },
+ {
+ "epoch": 37.488656195462475,
+ "grad_norm": 0.05219034478068352,
+ "learning_rate": 9.814009858586817e-05,
+ "loss": 3.5607855319976807,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50261780104712,
+ "grad_norm": 0.05180756375193596,
+ "learning_rate": 9.794716331470131e-05,
+ "loss": 3.567056179046631,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51657940663176,
+ "grad_norm": 0.0524761863052845,
+ "learning_rate": 9.77543932509012e-05,
+ "loss": 3.5638341903686523,
+ "step": 2701
+ },
+ {
+ "epoch": 37.5305410122164,
+ "grad_norm": 0.05237184837460518,
+ "learning_rate": 9.756178854977925e-05,
+ "loss": 3.575751781463623,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544502617801044,
+ "grad_norm": 0.05274934321641922,
+ "learning_rate": 9.736934936651362e-05,
+ "loss": 3.5704338550567627,
+ "step": 2703
+ },
+ {
+ "epoch": 37.55846422338569,
+ "grad_norm": 0.053565170615911484,
+ "learning_rate": 9.717707585614916e-05,
+ "loss": 3.566765785217285,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57242582897033,
+ "grad_norm": 0.05329638719558716,
+ "learning_rate": 9.698496817359698e-05,
+ "loss": 3.588357448577881,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58638743455497,
+ "grad_norm": 0.05382629856467247,
+ "learning_rate": 9.679302647363476e-05,
+ "loss": 3.578054428100586,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60034904013962,
+ "grad_norm": 0.05376645550131798,
+ "learning_rate": 9.660125091090675e-05,
+ "loss": 3.5657711029052734,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61431064572426,
+ "grad_norm": 0.05371774733066559,
+ "learning_rate": 9.64096416399228e-05,
+ "loss": 3.591853141784668,
+ "step": 2708
+ },
+ {
+ "epoch": 37.6282722513089,
+ "grad_norm": 0.055276691913604736,
+ "learning_rate": 9.621819881505918e-05,
+ "loss": 3.5756590366363525,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64223385689354,
+ "grad_norm": 0.05425848439335823,
+ "learning_rate": 9.602692259055767e-05,
+ "loss": 3.5689821243286133,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65619546247819,
+ "grad_norm": 0.0549544058740139,
+ "learning_rate": 9.583581312052646e-05,
+ "loss": 3.5869078636169434,
+ "step": 2711
+ },
+ {
+ "epoch": 37.67015706806283,
+ "grad_norm": 0.05412036553025246,
+ "learning_rate": 9.564487055893867e-05,
+ "loss": 3.57331919670105,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68411867364747,
+ "grad_norm": 0.05490295961499214,
+ "learning_rate": 9.545409505963338e-05,
+ "loss": 3.581331729888916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.69808027923211,
+ "grad_norm": 0.05448489263653755,
+ "learning_rate": 9.526348677631499e-05,
+ "loss": 3.605595588684082,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712041884816756,
+ "grad_norm": 0.05397019162774086,
+ "learning_rate": 9.507304586255311e-05,
+ "loss": 3.5643043518066406,
+ "step": 2715
+ },
+ {
+ "epoch": 37.726003490401396,
+ "grad_norm": 0.05399048700928688,
+ "learning_rate": 9.488277247178267e-05,
+ "loss": 3.5841126441955566,
+ "step": 2716
+ },
+ {
+ "epoch": 37.73996509598604,
+ "grad_norm": 0.055010151118040085,
+ "learning_rate": 9.469266675730319e-05,
+ "loss": 3.5840625762939453,
+ "step": 2717
+ },
+ {
+ "epoch": 37.753926701570684,
+ "grad_norm": 0.05478813126683235,
+ "learning_rate": 9.450272887227983e-05,
+ "loss": 3.56626033782959,
+ "step": 2718
+ },
+ {
+ "epoch": 37.767888307155324,
+ "grad_norm": 0.05396709218621254,
+ "learning_rate": 9.431295896974182e-05,
+ "loss": 3.5726799964904785,
+ "step": 2719
+ },
+ {
+ "epoch": 37.781849912739965,
+ "grad_norm": 0.05348159372806549,
+ "learning_rate": 9.412335720258341e-05,
+ "loss": 3.572139263153076,
+ "step": 2720
+ },
+ {
+ "epoch": 37.795811518324605,
+ "grad_norm": 0.055067192763090134,
+ "learning_rate": 9.393392372356335e-05,
+ "loss": 3.5794320106506348,
+ "step": 2721
+ },
+ {
+ "epoch": 37.80977312390925,
+ "grad_norm": 0.054990511387586594,
+ "learning_rate": 9.374465868530477e-05,
+ "loss": 3.573702335357666,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82373472949389,
+ "grad_norm": 0.05417406186461449,
+ "learning_rate": 9.355556224029515e-05,
+ "loss": 3.574542999267578,
+ "step": 2723
+ },
+ {
+ "epoch": 37.83769633507853,
+ "grad_norm": 0.05412792041897774,
+ "learning_rate": 9.336663454088593e-05,
+ "loss": 3.5837464332580566,
+ "step": 2724
+ },
+ {
+ "epoch": 37.85165794066317,
+ "grad_norm": 0.055187419056892395,
+ "learning_rate": 9.317787573929282e-05,
+ "loss": 3.5841941833496094,
+ "step": 2725
+ },
+ {
+ "epoch": 37.86561954624782,
+ "grad_norm": 0.05523138865828514,
+ "learning_rate": 9.298928598759541e-05,
+ "loss": 3.582552909851074,
+ "step": 2726
+ },
+ {
+ "epoch": 37.87958115183246,
+ "grad_norm": 0.05365348979830742,
+ "learning_rate": 9.280086543773698e-05,
+ "loss": 3.5693717002868652,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8935427574171,
+ "grad_norm": 0.0547465942800045,
+ "learning_rate": 9.26126142415247e-05,
+ "loss": 3.551072597503662,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90750436300174,
+ "grad_norm": 0.05491182208061218,
+ "learning_rate": 9.2424532550629e-05,
+ "loss": 3.573578357696533,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92146596858639,
+ "grad_norm": 0.055363260209560394,
+ "learning_rate": 9.223662051658408e-05,
+ "loss": 3.5630593299865723,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93542757417103,
+ "grad_norm": 0.05482657626271248,
+ "learning_rate": 9.204887829078709e-05,
+ "loss": 3.5579466819763184,
+ "step": 2731
+ },
+ {
+ "epoch": 37.94938917975567,
+ "grad_norm": 0.05555591359734535,
+ "learning_rate": 9.186130602449861e-05,
+ "loss": 3.572491407394409,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96335078534032,
+ "grad_norm": 0.0549536868929863,
+ "learning_rate": 9.167390386884224e-05,
+ "loss": 3.576451301574707,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97731239092496,
+ "grad_norm": 0.0547570176422596,
+ "learning_rate": 9.148667197480455e-05,
+ "loss": 3.5588786602020264,
+ "step": 2734
+ },
+ {
+ "epoch": 37.9912739965096,
+ "grad_norm": 0.055114034563302994,
+ "learning_rate": 9.129961049323494e-05,
+ "loss": 3.578408718109131,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.04000422731041908,
+ "learning_rate": 9.111271957484519e-05,
+ "loss": 2.212750196456909,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 0.6171724796295166,
+ "eval_runtime": 60.0184,
+ "eval_samples_per_second": 40.687,
+ "eval_steps_per_second": 0.65,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01396160558464,
+ "grad_norm": 0.05178803205490112,
+ "learning_rate": 9.092599937021031e-05,
+ "loss": 3.5207724571228027,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02792321116929,
+ "grad_norm": 0.048786960542201996,
+ "learning_rate": 9.073945002976715e-05,
+ "loss": 3.524625778198242,
+ "step": 2738
+ },
+ {
+ "epoch": 38.04188481675393,
+ "grad_norm": 0.05165952071547508,
+ "learning_rate": 9.055307170381522e-05,
+ "loss": 3.5370655059814453,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05584642233857,
+ "grad_norm": 0.05022848770022392,
+ "learning_rate": 9.036686454251613e-05,
+ "loss": 3.52911376953125,
+ "step": 2740
+ },
+ {
+ "epoch": 38.06980802792321,
+ "grad_norm": 0.04977148398756981,
+ "learning_rate": 9.018082869589369e-05,
+ "loss": 3.548173666000366,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083769633507856,
+ "grad_norm": 0.04955532029271126,
+ "learning_rate": 8.999496431383368e-05,
+ "loss": 3.5462260246276855,
+ "step": 2742
+ },
+ {
+ "epoch": 38.0977312390925,
+ "grad_norm": 0.051178816705942154,
+ "learning_rate": 8.980927154608341e-05,
+ "loss": 3.539468288421631,
+ "step": 2743
+ },
+ {
+ "epoch": 38.11169284467714,
+ "grad_norm": 0.05035502463579178,
+ "learning_rate": 8.962375054225255e-05,
+ "loss": 3.5545785427093506,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12565445026178,
+ "grad_norm": 0.04895094409584999,
+ "learning_rate": 8.943840145181177e-05,
+ "loss": 3.535142421722412,
+ "step": 2745
+ },
+ {
+ "epoch": 38.139616055846425,
+ "grad_norm": 0.050836317241191864,
+ "learning_rate": 8.925322442409361e-05,
+ "loss": 3.5562942028045654,
+ "step": 2746
+ },
+ {
+ "epoch": 38.153577661431065,
+ "grad_norm": 0.04898427426815033,
+ "learning_rate": 8.906821960829163e-05,
+ "loss": 3.533984661102295,
+ "step": 2747
+ },
+ {
+ "epoch": 38.167539267015705,
+ "grad_norm": 0.05146025866270065,
+ "learning_rate": 8.888338715346113e-05,
+ "loss": 3.5528769493103027,
+ "step": 2748
+ },
+ {
+ "epoch": 38.181500872600346,
+ "grad_norm": 0.049640901386737823,
+ "learning_rate": 8.869872720851831e-05,
+ "loss": 3.5450353622436523,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19546247818499,
+ "grad_norm": 0.04954954609274864,
+ "learning_rate": 8.851423992224012e-05,
+ "loss": 3.566117763519287,
+ "step": 2750
+ },
+ {
+ "epoch": 38.20942408376963,
+ "grad_norm": 0.04950375854969025,
+ "learning_rate": 8.832992544326473e-05,
+ "loss": 3.549238443374634,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223385689354274,
+ "grad_norm": 0.05026113614439964,
+ "learning_rate": 8.814578392009104e-05,
+ "loss": 3.5252106189727783,
+ "step": 2752
+ },
+ {
+ "epoch": 38.23734729493892,
+ "grad_norm": 0.050842445343732834,
+ "learning_rate": 8.796181550107857e-05,
+ "loss": 3.5351614952087402,
+ "step": 2753
+ },
+ {
+ "epoch": 38.25130890052356,
+ "grad_norm": 0.050328366458415985,
+ "learning_rate": 8.777802033444712e-05,
+ "loss": 3.5556483268737793,
+ "step": 2754
+ },
+ {
+ "epoch": 38.2652705061082,
+ "grad_norm": 0.05114470049738884,
+ "learning_rate": 8.75943985682774e-05,
+ "loss": 3.5427448749542236,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27923211169284,
+ "grad_norm": 0.05121666193008423,
+ "learning_rate": 8.741095035051017e-05,
+ "loss": 3.5361504554748535,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29319371727749,
+ "grad_norm": 0.0501880943775177,
+ "learning_rate": 8.722767582894613e-05,
+ "loss": 3.5314698219299316,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30715532286213,
+ "grad_norm": 0.05276653543114662,
+ "learning_rate": 8.704457515124636e-05,
+ "loss": 3.5526633262634277,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32111692844677,
+ "grad_norm": 0.05037686228752136,
+ "learning_rate": 8.686164846493176e-05,
+ "loss": 3.537220001220703,
+ "step": 2759
+ },
+ {
+ "epoch": 38.33507853403141,
+ "grad_norm": 0.051579032093286514,
+ "learning_rate": 8.667889591738317e-05,
+ "loss": 3.5302658081054688,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34904013961606,
+ "grad_norm": 0.05040337145328522,
+ "learning_rate": 8.649631765584083e-05,
+ "loss": 3.528028964996338,
+ "step": 2761
+ },
+ {
+ "epoch": 38.3630017452007,
+ "grad_norm": 0.05162221938371658,
+ "learning_rate": 8.631391382740482e-05,
+ "loss": 3.5771780014038086,
+ "step": 2762
+ },
+ {
+ "epoch": 38.37696335078534,
+ "grad_norm": 0.05134383216500282,
+ "learning_rate": 8.613168457903459e-05,
+ "loss": 3.532721996307373,
+ "step": 2763
+ },
+ {
+ "epoch": 38.390924956369986,
+ "grad_norm": 0.05137534812092781,
+ "learning_rate": 8.594963005754901e-05,
+ "loss": 3.550847291946411,
+ "step": 2764
+ },
+ {
+ "epoch": 38.404886561954626,
+ "grad_norm": 0.05198913812637329,
+ "learning_rate": 8.57677504096261e-05,
+ "loss": 3.556427478790283,
+ "step": 2765
+ },
+ {
+ "epoch": 38.41884816753927,
+ "grad_norm": 0.05211595818400383,
+ "learning_rate": 8.558604578180301e-05,
+ "loss": 3.579143524169922,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43280977312391,
+ "grad_norm": 0.05178852006793022,
+ "learning_rate": 8.540451632047593e-05,
+ "loss": 3.548065423965454,
+ "step": 2767
+ },
+ {
+ "epoch": 38.446771378708554,
+ "grad_norm": 0.05172461271286011,
+ "learning_rate": 8.522316217189972e-05,
+ "loss": 3.523592948913574,
+ "step": 2768
+ },
+ {
+ "epoch": 38.460732984293195,
+ "grad_norm": 0.05164046585559845,
+ "learning_rate": 8.504198348218821e-05,
+ "loss": 3.5595390796661377,
+ "step": 2769
+ },
+ {
+ "epoch": 38.474694589877835,
+ "grad_norm": 0.05212831869721413,
+ "learning_rate": 8.486098039731384e-05,
+ "loss": 3.517983913421631,
+ "step": 2770
+ },
+ {
+ "epoch": 38.488656195462475,
+ "grad_norm": 0.05294293537735939,
+ "learning_rate": 8.46801530631075e-05,
+ "loss": 3.5454447269439697,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50261780104712,
+ "grad_norm": 0.05115564912557602,
+ "learning_rate": 8.449950162525859e-05,
+ "loss": 3.562196731567383,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51657940663176,
+ "grad_norm": 0.05171671882271767,
+ "learning_rate": 8.431902622931443e-05,
+ "loss": 3.526003360748291,
+ "step": 2773
+ },
+ {
+ "epoch": 38.5305410122164,
+ "grad_norm": 0.052197013050317764,
+ "learning_rate": 8.413872702068119e-05,
+ "loss": 3.5459299087524414,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544502617801044,
+ "grad_norm": 0.05189663916826248,
+ "learning_rate": 8.395860414462238e-05,
+ "loss": 3.572622299194336,
+ "step": 2775
+ },
+ {
+ "epoch": 38.55846422338569,
+ "grad_norm": 0.05171465501189232,
+ "learning_rate": 8.377865774625985e-05,
+ "loss": 3.535243034362793,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57242582897033,
+ "grad_norm": 0.052626628428697586,
+ "learning_rate": 8.359888797057316e-05,
+ "loss": 3.555469036102295,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58638743455497,
+ "grad_norm": 0.05242423340678215,
+ "learning_rate": 8.341929496239958e-05,
+ "loss": 3.5629148483276367,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60034904013962,
+ "grad_norm": 0.051777057349681854,
+ "learning_rate": 8.323987886643402e-05,
+ "loss": 3.5607125759124756,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61431064572426,
+ "grad_norm": 0.05259648710489273,
+ "learning_rate": 8.306063982722855e-05,
+ "loss": 3.5630576610565186,
+ "step": 2780
+ },
+ {
+ "epoch": 38.6282722513089,
+ "grad_norm": 0.05220061168074608,
+ "learning_rate": 8.288157798919318e-05,
+ "loss": 3.545198917388916,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64223385689354,
+ "grad_norm": 0.05149109661579132,
+ "learning_rate": 8.27026934965945e-05,
+ "loss": 3.5277156829833984,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65619546247819,
+ "grad_norm": 0.05252330005168915,
+ "learning_rate": 8.25239864935566e-05,
+ "loss": 3.541299343109131,
+ "step": 2783
+ },
+ {
+ "epoch": 38.67015706806283,
+ "grad_norm": 0.05302104353904724,
+ "learning_rate": 8.23454571240605e-05,
+ "loss": 3.5433006286621094,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68411867364747,
+ "grad_norm": 0.053501565009355545,
+ "learning_rate": 8.216710553194416e-05,
+ "loss": 3.5382585525512695,
+ "step": 2785
+ },
+ {
+ "epoch": 38.69808027923211,
+ "grad_norm": 0.053820304572582245,
+ "learning_rate": 8.198893186090222e-05,
+ "loss": 3.5610904693603516,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712041884816756,
+ "grad_norm": 0.05330271273851395,
+ "learning_rate": 8.181093625448585e-05,
+ "loss": 3.5504231452941895,
+ "step": 2787
+ },
+ {
+ "epoch": 38.726003490401396,
+ "grad_norm": 0.05421965569257736,
+ "learning_rate": 8.163311885610299e-05,
+ "loss": 3.550283432006836,
+ "step": 2788
+ },
+ {
+ "epoch": 38.73996509598604,
+ "grad_norm": 0.053781405091285706,
+ "learning_rate": 8.145547980901791e-05,
+ "loss": 3.5624003410339355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.753926701570684,
+ "grad_norm": 0.05342188850045204,
+ "learning_rate": 8.127801925635126e-05,
+ "loss": 3.524596691131592,
+ "step": 2790
+ },
+ {
+ "epoch": 38.767888307155324,
+ "grad_norm": 0.05213013291358948,
+ "learning_rate": 8.110073734107954e-05,
+ "loss": 3.5316734313964844,
+ "step": 2791
+ },
+ {
+ "epoch": 38.781849912739965,
+ "grad_norm": 0.052924077957868576,
+ "learning_rate": 8.092363420603584e-05,
+ "loss": 3.5611703395843506,
+ "step": 2792
+ },
+ {
+ "epoch": 38.795811518324605,
+ "grad_norm": 0.05358070507645607,
+ "learning_rate": 8.074670999390896e-05,
+ "loss": 3.545983076095581,
+ "step": 2793
+ },
+ {
+ "epoch": 38.80977312390925,
+ "grad_norm": 0.053544074296951294,
+ "learning_rate": 8.056996484724342e-05,
+ "loss": 3.5701904296875,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82373472949389,
+ "grad_norm": 0.054532721638679504,
+ "learning_rate": 8.039339890843958e-05,
+ "loss": 3.543095111846924,
+ "step": 2795
+ },
+ {
+ "epoch": 38.83769633507853,
+ "grad_norm": 0.0532798208296299,
+ "learning_rate": 8.02170123197535e-05,
+ "loss": 3.5167088508605957,
+ "step": 2796
+ },
+ {
+ "epoch": 38.85165794066317,
+ "grad_norm": 0.05314210429787636,
+ "learning_rate": 8.004080522329674e-05,
+ "loss": 3.5611109733581543,
+ "step": 2797
+ },
+ {
+ "epoch": 38.86561954624782,
+ "grad_norm": 0.053358301520347595,
+ "learning_rate": 7.986477776103601e-05,
+ "loss": 3.550701141357422,
+ "step": 2798
+ },
+ {
+ "epoch": 38.87958115183246,
+ "grad_norm": 0.0538727343082428,
+ "learning_rate": 7.968893007479343e-05,
+ "loss": 3.562330722808838,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8935427574171,
+ "grad_norm": 0.05346875637769699,
+ "learning_rate": 7.951326230624658e-05,
+ "loss": 3.6011950969696045,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90750436300174,
+ "grad_norm": 0.05399498715996742,
+ "learning_rate": 7.933777459692756e-05,
+ "loss": 3.552870750427246,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92146596858639,
+ "grad_norm": 0.053817737847566605,
+ "learning_rate": 7.916246708822373e-05,
+ "loss": 3.5696794986724854,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93542757417103,
+ "grad_norm": 0.05272316932678223,
+ "learning_rate": 7.898733992137715e-05,
+ "loss": 3.55513334274292,
+ "step": 2803
+ },
+ {
+ "epoch": 38.94938917975567,
+ "grad_norm": 0.053616899996995926,
+ "learning_rate": 7.881239323748475e-05,
+ "loss": 3.5549991130828857,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96335078534032,
+ "grad_norm": 0.053872063755989075,
+ "learning_rate": 7.863762717749771e-05,
+ "loss": 3.537346839904785,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97731239092496,
+ "grad_norm": 0.054122194647789,
+ "learning_rate": 7.846304188222189e-05,
+ "loss": 3.552178144454956,
+ "step": 2806
+ },
+ {
+ "epoch": 38.9912739965096,
+ "grad_norm": 0.053457897156476974,
+ "learning_rate": 7.828863749231777e-05,
+ "loss": 3.574368715286255,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.04022035375237465,
+ "learning_rate": 7.811441414829958e-05,
+ "loss": 2.2162294387817383,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 0.6184564828872681,
+ "eval_runtime": 60.5596,
+ "eval_samples_per_second": 40.324,
+ "eval_steps_per_second": 0.644,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01396160558464,
+ "grad_norm": 0.04818584397435188,
+ "learning_rate": 7.794037199053598e-05,
+ "loss": 3.5161890983581543,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02792321116929,
+ "grad_norm": 0.04768325760960579,
+ "learning_rate": 7.776651115924959e-05,
+ "loss": 3.5117123126983643,
+ "step": 2810
+ },
+ {
+ "epoch": 39.04188481675393,
+ "grad_norm": 0.049668096005916595,
+ "learning_rate": 7.759283179451704e-05,
+ "loss": 3.5076825618743896,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05584642233857,
+ "grad_norm": 0.04928773269057274,
+ "learning_rate": 7.741933403626848e-05,
+ "loss": 3.522817611694336,
+ "step": 2812
+ },
+ {
+ "epoch": 39.06980802792321,
+ "grad_norm": 0.049023669213056564,
+ "learning_rate": 7.724601802428805e-05,
+ "loss": 3.4985058307647705,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083769633507856,
+ "grad_norm": 0.04910000413656235,
+ "learning_rate": 7.707288389821329e-05,
+ "loss": 3.5706303119659424,
+ "step": 2814
+ },
+ {
+ "epoch": 39.0977312390925,
+ "grad_norm": 0.04971414431929588,
+ "learning_rate": 7.689993179753519e-05,
+ "loss": 3.5187277793884277,
+ "step": 2815
+ },
+ {
+ "epoch": 39.11169284467714,
+ "grad_norm": 0.049427829682826996,
+ "learning_rate": 7.672716186159829e-05,
+ "loss": 3.5230023860931396,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12565445026178,
+ "grad_norm": 0.04888913407921791,
+ "learning_rate": 7.655457422959993e-05,
+ "loss": 3.525777816772461,
+ "step": 2817
+ },
+ {
+ "epoch": 39.139616055846425,
+ "grad_norm": 0.04879172518849373,
+ "learning_rate": 7.638216904059122e-05,
+ "loss": 3.5179781913757324,
+ "step": 2818
+ },
+ {
+ "epoch": 39.153577661431065,
+ "grad_norm": 0.04999098926782608,
+ "learning_rate": 7.620994643347559e-05,
+ "loss": 3.53208327293396,
+ "step": 2819
+ },
+ {
+ "epoch": 39.167539267015705,
+ "grad_norm": 0.051719557493925095,
+ "learning_rate": 7.60379065470098e-05,
+ "loss": 3.500316619873047,
+ "step": 2820
+ },
+ {
+ "epoch": 39.181500872600346,
+ "grad_norm": 0.05016033351421356,
+ "learning_rate": 7.586604951980326e-05,
+ "loss": 3.4964962005615234,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19546247818499,
+ "grad_norm": 0.048761315643787384,
+ "learning_rate": 7.569437549031813e-05,
+ "loss": 3.5171115398406982,
+ "step": 2822
+ },
+ {
+ "epoch": 39.20942408376963,
+ "grad_norm": 0.05029286816716194,
+ "learning_rate": 7.55228845968691e-05,
+ "loss": 3.5414791107177734,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223385689354274,
+ "grad_norm": 0.0502147302031517,
+ "learning_rate": 7.535157697762313e-05,
+ "loss": 3.519467830657959,
+ "step": 2824
+ },
+ {
+ "epoch": 39.23734729493892,
+ "grad_norm": 0.05098641291260719,
+ "learning_rate": 7.518045277059979e-05,
+ "loss": 3.539252519607544,
+ "step": 2825
+ },
+ {
+ "epoch": 39.25130890052356,
+ "grad_norm": 0.04990671947598457,
+ "learning_rate": 7.500951211367068e-05,
+ "loss": 3.5308918952941895,
+ "step": 2826
+ },
+ {
+ "epoch": 39.2652705061082,
+ "grad_norm": 0.04992193728685379,
+ "learning_rate": 7.483875514455979e-05,
+ "loss": 3.521590232849121,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27923211169284,
+ "grad_norm": 0.05016932636499405,
+ "learning_rate": 7.466818200084264e-05,
+ "loss": 3.5208449363708496,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29319371727749,
+ "grad_norm": 0.050798505544662476,
+ "learning_rate": 7.449779281994712e-05,
+ "loss": 3.5616073608398438,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30715532286213,
+ "grad_norm": 0.05047072842717171,
+ "learning_rate": 7.43275877391528e-05,
+ "loss": 3.540660858154297,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32111692844677,
+ "grad_norm": 0.051101770251989365,
+ "learning_rate": 7.415756689559061e-05,
+ "loss": 3.532931327819824,
+ "step": 2831
+ },
+ {
+ "epoch": 39.33507853403141,
+ "grad_norm": 0.049927644431591034,
+ "learning_rate": 7.398773042624341e-05,
+ "loss": 3.5045418739318848,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34904013961606,
+ "grad_norm": 0.05085239186882973,
+ "learning_rate": 7.381807846794532e-05,
+ "loss": 3.4864392280578613,
+ "step": 2833
+ },
+ {
+ "epoch": 39.3630017452007,
+ "grad_norm": 0.05125649273395538,
+ "learning_rate": 7.3648611157382e-05,
+ "loss": 3.51094913482666,
+ "step": 2834
+ },
+ {
+ "epoch": 39.37696335078534,
+ "grad_norm": 0.05072709918022156,
+ "learning_rate": 7.34793286310899e-05,
+ "loss": 3.531437873840332,
+ "step": 2835
+ },
+ {
+ "epoch": 39.390924956369986,
+ "grad_norm": 0.050076793879270554,
+ "learning_rate": 7.331023102545716e-05,
+ "loss": 3.5222954750061035,
+ "step": 2836
+ },
+ {
+ "epoch": 39.404886561954626,
+ "grad_norm": 0.050684161484241486,
+ "learning_rate": 7.314131847672268e-05,
+ "loss": 3.525660991668701,
+ "step": 2837
+ },
+ {
+ "epoch": 39.41884816753927,
+ "grad_norm": 0.05221536383032799,
+ "learning_rate": 7.297259112097608e-05,
+ "loss": 3.547537326812744,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43280977312391,
+ "grad_norm": 0.050070326775312424,
+ "learning_rate": 7.280404909415801e-05,
+ "loss": 3.524541139602661,
+ "step": 2839
+ },
+ {
+ "epoch": 39.446771378708554,
+ "grad_norm": 0.05107896402478218,
+ "learning_rate": 7.263569253205973e-05,
+ "loss": 3.4862544536590576,
+ "step": 2840
+ },
+ {
+ "epoch": 39.460732984293195,
+ "grad_norm": 0.05145040526986122,
+ "learning_rate": 7.24675215703232e-05,
+ "loss": 3.5242197513580322,
+ "step": 2841
+ },
+ {
+ "epoch": 39.474694589877835,
+ "grad_norm": 0.05019835755228996,
+ "learning_rate": 7.229953634444055e-05,
+ "loss": 3.5409276485443115,
+ "step": 2842
+ },
+ {
+ "epoch": 39.488656195462475,
+ "grad_norm": 0.051458705216646194,
+ "learning_rate": 7.213173698975444e-05,
+ "loss": 3.5637054443359375,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50261780104712,
+ "grad_norm": 0.05186862125992775,
+ "learning_rate": 7.1964123641458e-05,
+ "loss": 3.5018935203552246,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51657940663176,
+ "grad_norm": 0.05114978924393654,
+ "learning_rate": 7.179669643459405e-05,
+ "loss": 3.5413007736206055,
+ "step": 2845
+ },
+ {
+ "epoch": 39.5305410122164,
+ "grad_norm": 0.051313482224941254,
+ "learning_rate": 7.162945550405575e-05,
+ "loss": 3.5065975189208984,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544502617801044,
+ "grad_norm": 0.0509142242372036,
+ "learning_rate": 7.146240098458613e-05,
+ "loss": 3.521883964538574,
+ "step": 2847
+ },
+ {
+ "epoch": 39.55846422338569,
+ "grad_norm": 0.05225839465856552,
+ "learning_rate": 7.129553301077801e-05,
+ "loss": 3.534417152404785,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57242582897033,
+ "grad_norm": 0.05145575851202011,
+ "learning_rate": 7.112885171707378e-05,
+ "loss": 3.5341527462005615,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58638743455497,
+ "grad_norm": 0.0504966638982296,
+ "learning_rate": 7.096235723776562e-05,
+ "loss": 3.529374599456787,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60034904013962,
+ "grad_norm": 0.05249404534697533,
+ "learning_rate": 7.079604970699514e-05,
+ "loss": 3.537140130996704,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61431064572426,
+ "grad_norm": 0.052380647510290146,
+ "learning_rate": 7.062992925875332e-05,
+ "loss": 3.529085159301758,
+ "step": 2852
+ },
+ {
+ "epoch": 39.6282722513089,
+ "grad_norm": 0.05162700638175011,
+ "learning_rate": 7.046399602688045e-05,
+ "loss": 3.549781084060669,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64223385689354,
+ "grad_norm": 0.05245045945048332,
+ "learning_rate": 7.029825014506572e-05,
+ "loss": 3.5424249172210693,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65619546247819,
+ "grad_norm": 0.05116629973053932,
+ "learning_rate": 7.013269174684795e-05,
+ "loss": 3.5407791137695312,
+ "step": 2855
+ },
+ {
+ "epoch": 39.67015706806283,
+ "grad_norm": 0.0507299080491066,
+ "learning_rate": 6.996732096561422e-05,
+ "loss": 3.550326108932495,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68411867364747,
+ "grad_norm": 0.05050551891326904,
+ "learning_rate": 6.980213793460092e-05,
+ "loss": 3.50667667388916,
+ "step": 2857
+ },
+ {
+ "epoch": 39.69808027923211,
+ "grad_norm": 0.052720312029123306,
+ "learning_rate": 6.963714278689304e-05,
+ "loss": 3.5426530838012695,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712041884816756,
+ "grad_norm": 0.05040572211146355,
+ "learning_rate": 6.947233565542417e-05,
+ "loss": 3.50482177734375,
+ "step": 2859
+ },
+ {
+ "epoch": 39.726003490401396,
+ "grad_norm": 0.05204465985298157,
+ "learning_rate": 6.930771667297653e-05,
+ "loss": 3.5070583820343018,
+ "step": 2860
+ },
+ {
+ "epoch": 39.73996509598604,
+ "grad_norm": 0.05265089496970177,
+ "learning_rate": 6.914328597218042e-05,
+ "loss": 3.53663969039917,
+ "step": 2861
+ },
+ {
+ "epoch": 39.753926701570684,
+ "grad_norm": 0.05254863202571869,
+ "learning_rate": 6.8979043685515e-05,
+ "loss": 3.5625903606414795,
+ "step": 2862
+ },
+ {
+ "epoch": 39.767888307155324,
+ "grad_norm": 0.05163634940981865,
+ "learning_rate": 6.881498994530708e-05,
+ "loss": 3.541079044342041,
+ "step": 2863
+ },
+ {
+ "epoch": 39.781849912739965,
+ "grad_norm": 0.05227374657988548,
+ "learning_rate": 6.865112488373186e-05,
+ "loss": 3.52247953414917,
+ "step": 2864
+ },
+ {
+ "epoch": 39.795811518324605,
+ "grad_norm": 0.051065072417259216,
+ "learning_rate": 6.84874486328125e-05,
+ "loss": 3.5435099601745605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.80977312390925,
+ "grad_norm": 0.05301819369196892,
+ "learning_rate": 6.832396132441993e-05,
+ "loss": 3.551717758178711,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82373472949389,
+ "grad_norm": 0.052438024431467056,
+ "learning_rate": 6.816066309027309e-05,
+ "loss": 3.545943260192871,
+ "step": 2867
+ },
+ {
+ "epoch": 39.83769633507853,
+ "grad_norm": 0.05212542042136192,
+ "learning_rate": 6.799755406193815e-05,
+ "loss": 3.5160136222839355,
+ "step": 2868
+ },
+ {
+ "epoch": 39.85165794066317,
+ "grad_norm": 0.05442605912685394,
+ "learning_rate": 6.783463437082929e-05,
+ "loss": 3.5255932807922363,
+ "step": 2869
+ },
+ {
+ "epoch": 39.86561954624782,
+ "grad_norm": 0.05317089334130287,
+ "learning_rate": 6.767190414820792e-05,
+ "loss": 3.5328879356384277,
+ "step": 2870
+ },
+ {
+ "epoch": 39.87958115183246,
+ "grad_norm": 0.05377183482050896,
+ "learning_rate": 6.750936352518284e-05,
+ "loss": 3.512451171875,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8935427574171,
+ "grad_norm": 0.05238938704133034,
+ "learning_rate": 6.734701263271011e-05,
+ "loss": 3.5253477096557617,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90750436300174,
+ "grad_norm": 0.0536922849714756,
+ "learning_rate": 6.718485160159289e-05,
+ "loss": 3.532573938369751,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92146596858639,
+ "grad_norm": 0.05387766659259796,
+ "learning_rate": 6.702288056248145e-05,
+ "loss": 3.535872459411621,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93542757417103,
+ "grad_norm": 0.05426036939024925,
+ "learning_rate": 6.686109964587285e-05,
+ "loss": 3.534602165222168,
+ "step": 2875
+ },
+ {
+ "epoch": 39.94938917975567,
+ "grad_norm": 0.05351133272051811,
+ "learning_rate": 6.669950898211106e-05,
+ "loss": 3.507835865020752,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96335078534032,
+ "grad_norm": 0.05391782522201538,
+ "learning_rate": 6.653810870138676e-05,
+ "loss": 3.543053150177002,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97731239092496,
+ "grad_norm": 0.05348298326134682,
+ "learning_rate": 6.637689893373729e-05,
+ "loss": 3.528005361557007,
+ "step": 2878
+ },
+ {
+ "epoch": 39.9912739965096,
+ "grad_norm": 0.05260506272315979,
+ "learning_rate": 6.621587980904651e-05,
+ "loss": 3.542234420776367,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.040236134082078934,
+ "learning_rate": 6.605505145704437e-05,
+ "loss": 2.2050814628601074,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 0.6198474764823914,
+ "eval_runtime": 59.2234,
+ "eval_samples_per_second": 41.234,
+ "eval_steps_per_second": 0.659,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01396160558464,
+ "grad_norm": 0.048478540033102036,
+ "learning_rate": 6.58944140073077e-05,
+ "loss": 3.524498224258423,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02792321116929,
+ "grad_norm": 0.04602562263607979,
+ "learning_rate": 6.573396758925891e-05,
+ "loss": 3.5231332778930664,
+ "step": 2882
+ },
+ {
+ "epoch": 40.04188481675393,
+ "grad_norm": 0.04725645110011101,
+ "learning_rate": 6.557371233216692e-05,
+ "loss": 3.4903368949890137,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05584642233857,
+ "grad_norm": 0.04776613414287567,
+ "learning_rate": 6.541364836514646e-05,
+ "loss": 3.4806885719299316,
+ "step": 2884
+ },
+ {
+ "epoch": 40.06980802792321,
+ "grad_norm": 0.04646219685673714,
+ "learning_rate": 6.525377581715829e-05,
+ "loss": 3.5238184928894043,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083769633507856,
+ "grad_norm": 0.04782753065228462,
+ "learning_rate": 6.509409481700864e-05,
+ "loss": 3.4997220039367676,
+ "step": 2886
+ },
+ {
+ "epoch": 40.0977312390925,
+ "grad_norm": 0.04805218428373337,
+ "learning_rate": 6.49346054933496e-05,
+ "loss": 3.493037700653076,
+ "step": 2887
+ },
+ {
+ "epoch": 40.11169284467714,
+ "grad_norm": 0.047421909868717194,
+ "learning_rate": 6.477530797467911e-05,
+ "loss": 3.510756492614746,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12565445026178,
+ "grad_norm": 0.04791928082704544,
+ "learning_rate": 6.461620238934006e-05,
+ "loss": 3.503185510635376,
+ "step": 2889
+ },
+ {
+ "epoch": 40.139616055846425,
+ "grad_norm": 0.047295596450567245,
+ "learning_rate": 6.445728886552109e-05,
+ "loss": 3.5190024375915527,
+ "step": 2890
+ },
+ {
+ "epoch": 40.153577661431065,
+ "grad_norm": 0.047632548958063126,
+ "learning_rate": 6.429856753125573e-05,
+ "loss": 3.5584514141082764,
+ "step": 2891
+ },
+ {
+ "epoch": 40.167539267015705,
+ "grad_norm": 0.04803183302283287,
+ "learning_rate": 6.414003851442318e-05,
+ "loss": 3.5070176124572754,
+ "step": 2892
+ },
+ {
+ "epoch": 40.181500872600346,
+ "grad_norm": 0.047644875943660736,
+ "learning_rate": 6.398170194274722e-05,
+ "loss": 3.489863395690918,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19546247818499,
+ "grad_norm": 0.04730125144124031,
+ "learning_rate": 6.38235579437968e-05,
+ "loss": 3.5007224082946777,
+ "step": 2894
+ },
+ {
+ "epoch": 40.20942408376963,
+ "grad_norm": 0.04779860004782677,
+ "learning_rate": 6.366560664498572e-05,
+ "loss": 3.5284690856933594,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223385689354274,
+ "grad_norm": 0.04776514694094658,
+ "learning_rate": 6.350784817357242e-05,
+ "loss": 3.462017774581909,
+ "step": 2896
+ },
+ {
+ "epoch": 40.23734729493892,
+ "grad_norm": 0.04804418608546257,
+ "learning_rate": 6.335028265666023e-05,
+ "loss": 3.5062084197998047,
+ "step": 2897
+ },
+ {
+ "epoch": 40.25130890052356,
+ "grad_norm": 0.047954075038433075,
+ "learning_rate": 6.319291022119653e-05,
+ "loss": 3.4866459369659424,
+ "step": 2898
+ },
+ {
+ "epoch": 40.2652705061082,
+ "grad_norm": 0.0490020215511322,
+ "learning_rate": 6.303573099397378e-05,
+ "loss": 3.4901952743530273,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27923211169284,
+ "grad_norm": 0.04876933991909027,
+ "learning_rate": 6.287874510162821e-05,
+ "loss": 3.5119152069091797,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29319371727749,
+ "grad_norm": 0.0485818088054657,
+ "learning_rate": 6.272195267064063e-05,
+ "loss": 3.4974474906921387,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30715532286213,
+ "grad_norm": 0.04771488532423973,
+ "learning_rate": 6.25653538273358e-05,
+ "loss": 3.523752212524414,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32111692844677,
+ "grad_norm": 0.048272743821144104,
+ "learning_rate": 6.240894869788267e-05,
+ "loss": 3.4994726181030273,
+ "step": 2903
+ },
+ {
+ "epoch": 40.33507853403141,
+ "grad_norm": 0.04918459802865982,
+ "learning_rate": 6.225273740829404e-05,
+ "loss": 3.5262579917907715,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34904013961606,
+ "grad_norm": 0.04880300536751747,
+ "learning_rate": 6.209672008442635e-05,
+ "loss": 3.4975790977478027,
+ "step": 2905
+ },
+ {
+ "epoch": 40.3630017452007,
+ "grad_norm": 0.049021199345588684,
+ "learning_rate": 6.19408968519801e-05,
+ "loss": 3.5101144313812256,
+ "step": 2906
+ },
+ {
+ "epoch": 40.37696335078534,
+ "grad_norm": 0.04934848099946976,
+ "learning_rate": 6.178526783649916e-05,
+ "loss": 3.4800543785095215,
+ "step": 2907
+ },
+ {
+ "epoch": 40.390924956369986,
+ "grad_norm": 0.048395898193120956,
+ "learning_rate": 6.162983316337109e-05,
+ "loss": 3.5071773529052734,
+ "step": 2908
+ },
+ {
+ "epoch": 40.404886561954626,
+ "grad_norm": 0.049257755279541016,
+ "learning_rate": 6.147459295782676e-05,
+ "loss": 3.509805202484131,
+ "step": 2909
+ },
+ {
+ "epoch": 40.41884816753927,
+ "grad_norm": 0.051024191081523895,
+ "learning_rate": 6.131954734494045e-05,
+ "loss": 3.5087671279907227,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43280977312391,
+ "grad_norm": 0.04894109442830086,
+ "learning_rate": 6.11646964496296e-05,
+ "loss": 3.5116591453552246,
+ "step": 2911
+ },
+ {
+ "epoch": 40.446771378708554,
+ "grad_norm": 0.04954129084944725,
+ "learning_rate": 6.101004039665471e-05,
+ "loss": 3.4806766510009766,
+ "step": 2912
+ },
+ {
+ "epoch": 40.460732984293195,
+ "grad_norm": 0.04909701645374298,
+ "learning_rate": 6.085557931061937e-05,
+ "loss": 3.5208868980407715,
+ "step": 2913
+ },
+ {
+ "epoch": 40.474694589877835,
+ "grad_norm": 0.04900852590799332,
+ "learning_rate": 6.070131331597015e-05,
+ "loss": 3.5316522121429443,
+ "step": 2914
+ },
+ {
+ "epoch": 40.488656195462475,
+ "grad_norm": 0.04930015280842781,
+ "learning_rate": 6.054724253699636e-05,
+ "loss": 3.4901177883148193,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50261780104712,
+ "grad_norm": 0.04921771585941315,
+ "learning_rate": 6.03933670978301e-05,
+ "loss": 3.500121593475342,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51657940663176,
+ "grad_norm": 0.05016683414578438,
+ "learning_rate": 6.0239687122445826e-05,
+ "loss": 3.5284152030944824,
+ "step": 2917
+ },
+ {
+ "epoch": 40.5305410122164,
+ "grad_norm": 0.049245234578847885,
+ "learning_rate": 6.0086202734661e-05,
+ "loss": 3.5167784690856934,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544502617801044,
+ "grad_norm": 0.049730222672224045,
+ "learning_rate": 5.9932914058135004e-05,
+ "loss": 3.5100388526916504,
+ "step": 2919
+ },
+ {
+ "epoch": 40.55846422338569,
+ "grad_norm": 0.05063595995306969,
+ "learning_rate": 5.977982121636984e-05,
+ "loss": 3.5240533351898193,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57242582897033,
+ "grad_norm": 0.049625277519226074,
+ "learning_rate": 5.962692433270962e-05,
+ "loss": 3.4961838722229004,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58638743455497,
+ "grad_norm": 0.048995789140462875,
+ "learning_rate": 5.9474223530340634e-05,
+ "loss": 3.519697427749634,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60034904013962,
+ "grad_norm": 0.05007295310497284,
+ "learning_rate": 5.932171893229124e-05,
+ "loss": 3.4762754440307617,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61431064572426,
+ "grad_norm": 0.05034065619111061,
+ "learning_rate": 5.916941066143137e-05,
+ "loss": 3.510772705078125,
+ "step": 2924
+ },
+ {
+ "epoch": 40.6282722513089,
+ "grad_norm": 0.05077948421239853,
+ "learning_rate": 5.9017298840473364e-05,
+ "loss": 3.50581693649292,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64223385689354,
+ "grad_norm": 0.05035000666975975,
+ "learning_rate": 5.8865383591970775e-05,
+ "loss": 3.509791374206543,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65619546247819,
+ "grad_norm": 0.04955745488405228,
+ "learning_rate": 5.8713665038319e-05,
+ "loss": 3.5151772499084473,
+ "step": 2927
+ },
+ {
+ "epoch": 40.67015706806283,
+ "grad_norm": 0.05059986561536789,
+ "learning_rate": 5.856214330175498e-05,
+ "loss": 3.5117061138153076,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68411867364747,
+ "grad_norm": 0.050653424113988876,
+ "learning_rate": 5.841081850435704e-05,
+ "loss": 3.49762225151062,
+ "step": 2929
+ },
+ {
+ "epoch": 40.69808027923211,
+ "grad_norm": 0.049989622086286545,
+ "learning_rate": 5.825969076804488e-05,
+ "loss": 3.5447728633880615,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712041884816756,
+ "grad_norm": 0.05025019496679306,
+ "learning_rate": 5.810876021457925e-05,
+ "loss": 3.500591993331909,
+ "step": 2931
+ },
+ {
+ "epoch": 40.726003490401396,
+ "grad_norm": 0.04964078962802887,
+ "learning_rate": 5.795802696556222e-05,
+ "loss": 3.536098003387451,
+ "step": 2932
+ },
+ {
+ "epoch": 40.73996509598604,
+ "grad_norm": 0.05067792907357216,
+ "learning_rate": 5.78074911424369e-05,
+ "loss": 3.4908149242401123,
+ "step": 2933
+ },
+ {
+ "epoch": 40.753926701570684,
+ "grad_norm": 0.05067761242389679,
+ "learning_rate": 5.765715286648734e-05,
+ "loss": 3.5012640953063965,
+ "step": 2934
+ },
+ {
+ "epoch": 40.767888307155324,
+ "grad_norm": 0.05116443336009979,
+ "learning_rate": 5.7507012258838154e-05,
+ "loss": 3.5251986980438232,
+ "step": 2935
+ },
+ {
+ "epoch": 40.781849912739965,
+ "grad_norm": 0.05042244866490364,
+ "learning_rate": 5.7357069440455095e-05,
+ "loss": 3.493983030319214,
+ "step": 2936
+ },
+ {
+ "epoch": 40.795811518324605,
+ "grad_norm": 0.050487346947193146,
+ "learning_rate": 5.7207324532144454e-05,
+ "loss": 3.51411771774292,
+ "step": 2937
+ },
+ {
+ "epoch": 40.80977312390925,
+ "grad_norm": 0.050840575248003006,
+ "learning_rate": 5.7057777654552785e-05,
+ "loss": 3.528165340423584,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82373472949389,
+ "grad_norm": 0.052268560975790024,
+ "learning_rate": 5.690842892816741e-05,
+ "loss": 3.513277292251587,
+ "step": 2939
+ },
+ {
+ "epoch": 40.83769633507853,
+ "grad_norm": 0.05126330256462097,
+ "learning_rate": 5.675927847331593e-05,
+ "loss": 3.4808921813964844,
+ "step": 2940
+ },
+ {
+ "epoch": 40.85165794066317,
+ "grad_norm": 0.05069379135966301,
+ "learning_rate": 5.661032641016619e-05,
+ "loss": 3.517787456512451,
+ "step": 2941
+ },
+ {
+ "epoch": 40.86561954624782,
+ "grad_norm": 0.051330823451280594,
+ "learning_rate": 5.6461572858726086e-05,
+ "loss": 3.5521302223205566,
+ "step": 2942
+ },
+ {
+ "epoch": 40.87958115183246,
+ "grad_norm": 0.05037380009889603,
+ "learning_rate": 5.6313017938843595e-05,
+ "loss": 3.505967617034912,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8935427574171,
+ "grad_norm": 0.05150613188743591,
+ "learning_rate": 5.6164661770207026e-05,
+ "loss": 3.518454074859619,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90750436300174,
+ "grad_norm": 0.05246536061167717,
+ "learning_rate": 5.6016504472344e-05,
+ "loss": 3.518848180770874,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92146596858639,
+ "grad_norm": 0.052759598940610886,
+ "learning_rate": 5.5868546164622284e-05,
+ "loss": 3.511413097381592,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93542757417103,
+ "grad_norm": 0.05075768381357193,
+ "learning_rate": 5.572078696624917e-05,
+ "loss": 3.4923768043518066,
+ "step": 2947
+ },
+ {
+ "epoch": 40.94938917975567,
+ "grad_norm": 0.051101621240377426,
+ "learning_rate": 5.557322699627168e-05,
+ "loss": 3.545301675796509,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96335078534032,
+ "grad_norm": 0.05053454637527466,
+ "learning_rate": 5.542586637357607e-05,
+ "loss": 3.4888358116149902,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97731239092496,
+ "grad_norm": 0.05186663195490837,
+ "learning_rate": 5.527870521688817e-05,
+ "loss": 3.5280933380126953,
+ "step": 2950
+ },
+ {
+ "epoch": 40.9912739965096,
+ "grad_norm": 0.05131891742348671,
+ "learning_rate": 5.513174364477311e-05,
+ "loss": 3.5003533363342285,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.03856140375137329,
+ "learning_rate": 5.498498177563518e-05,
+ "loss": 2.1855430603027344,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 0.621021568775177,
+ "eval_runtime": 60.3507,
+ "eval_samples_per_second": 40.463,
+ "eval_steps_per_second": 0.646,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01396160558464,
+ "grad_norm": 0.04582727700471878,
+ "learning_rate": 5.4838419727717716e-05,
+ "loss": 3.492494821548462,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02792321116929,
+ "grad_norm": 0.04572693631052971,
+ "learning_rate": 5.4692057619103144e-05,
+ "loss": 3.4678001403808594,
+ "step": 2954
+ },
+ {
+ "epoch": 41.04188481675393,
+ "grad_norm": 0.04531959071755409,
+ "learning_rate": 5.454589556771286e-05,
+ "loss": 3.5139541625976562,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05584642233857,
+ "grad_norm": 0.046629469841718674,
+ "learning_rate": 5.439993369130681e-05,
+ "loss": 3.483097553253174,
+ "step": 2956
+ },
+ {
+ "epoch": 41.06980802792321,
+ "grad_norm": 0.04704796150326729,
+ "learning_rate": 5.4254172107483924e-05,
+ "loss": 3.484652042388916,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083769633507856,
+ "grad_norm": 0.04632614925503731,
+ "learning_rate": 5.410861093368169e-05,
+ "loss": 3.4961838722229004,
+ "step": 2958
+ },
+ {
+ "epoch": 41.0977312390925,
+ "grad_norm": 0.046637099236249924,
+ "learning_rate": 5.3963250287176094e-05,
+ "loss": 3.5225467681884766,
+ "step": 2959
+ },
+ {
+ "epoch": 41.11169284467714,
+ "grad_norm": 0.04615108296275139,
+ "learning_rate": 5.381809028508168e-05,
+ "loss": 3.478269577026367,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12565445026178,
+ "grad_norm": 0.04565376043319702,
+ "learning_rate": 5.367313104435104e-05,
+ "loss": 3.4609298706054688,
+ "step": 2961
+ },
+ {
+ "epoch": 41.139616055846425,
+ "grad_norm": 0.04716965928673744,
+ "learning_rate": 5.352837268177544e-05,
+ "loss": 3.498220443725586,
+ "step": 2962
+ },
+ {
+ "epoch": 41.153577661431065,
+ "grad_norm": 0.04637673497200012,
+ "learning_rate": 5.3383815313983916e-05,
+ "loss": 3.4881229400634766,
+ "step": 2963
+ },
+ {
+ "epoch": 41.167539267015705,
+ "grad_norm": 0.047194816172122955,
+ "learning_rate": 5.323945905744379e-05,
+ "loss": 3.4718289375305176,
+ "step": 2964
+ },
+ {
+ "epoch": 41.181500872600346,
+ "grad_norm": 0.04609799385070801,
+ "learning_rate": 5.3095304028460284e-05,
+ "loss": 3.460082530975342,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19546247818499,
+ "grad_norm": 0.04652848839759827,
+ "learning_rate": 5.2951350343176495e-05,
+ "loss": 3.4704749584198,
+ "step": 2966
+ },
+ {
+ "epoch": 41.20942408376963,
+ "grad_norm": 0.046700265258550644,
+ "learning_rate": 5.2807598117573384e-05,
+ "loss": 3.4888997077941895,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223385689354274,
+ "grad_norm": 0.046414151787757874,
+ "learning_rate": 5.266404746746939e-05,
+ "loss": 3.506500244140625,
+ "step": 2968
+ },
+ {
+ "epoch": 41.23734729493892,
+ "grad_norm": 0.04814400523900986,
+ "learning_rate": 5.252069850852072e-05,
+ "loss": 3.515608787536621,
+ "step": 2969
+ },
+ {
+ "epoch": 41.25130890052356,
+ "grad_norm": 0.046656325459480286,
+ "learning_rate": 5.237755135622107e-05,
+ "loss": 3.4709742069244385,
+ "step": 2970
+ },
+ {
+ "epoch": 41.2652705061082,
+ "grad_norm": 0.04764151945710182,
+ "learning_rate": 5.223460612590159e-05,
+ "loss": 3.5001049041748047,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27923211169284,
+ "grad_norm": 0.04743436351418495,
+ "learning_rate": 5.209186293273044e-05,
+ "loss": 3.4994988441467285,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29319371727749,
+ "grad_norm": 0.04705984145402908,
+ "learning_rate": 5.1949321891713446e-05,
+ "loss": 3.469075918197632,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30715532286213,
+ "grad_norm": 0.04596104845404625,
+ "learning_rate": 5.180698311769338e-05,
+ "loss": 3.5107405185699463,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32111692844677,
+ "grad_norm": 0.04772106930613518,
+ "learning_rate": 5.166484672534983e-05,
+ "loss": 3.4777815341949463,
+ "step": 2975
+ },
+ {
+ "epoch": 41.33507853403141,
+ "grad_norm": 0.04846428334712982,
+ "learning_rate": 5.152291282919961e-05,
+ "loss": 3.5055108070373535,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34904013961606,
+ "grad_norm": 0.04617132991552353,
+ "learning_rate": 5.138118154359632e-05,
+ "loss": 3.5047807693481445,
+ "step": 2977
+ },
+ {
+ "epoch": 41.3630017452007,
+ "grad_norm": 0.04738635942339897,
+ "learning_rate": 5.1239652982730335e-05,
+ "loss": 3.458731174468994,
+ "step": 2978
+ },
+ {
+ "epoch": 41.37696335078534,
+ "grad_norm": 0.046908650547266006,
+ "learning_rate": 5.1098327260628466e-05,
+ "loss": 3.4803903102874756,
+ "step": 2979
+ },
+ {
+ "epoch": 41.390924956369986,
+ "grad_norm": 0.04845135658979416,
+ "learning_rate": 5.09572044911545e-05,
+ "loss": 3.4671194553375244,
+ "step": 2980
+ },
+ {
+ "epoch": 41.404886561954626,
+ "grad_norm": 0.0482504665851593,
+ "learning_rate": 5.0816284788008527e-05,
+ "loss": 3.4812138080596924,
+ "step": 2981
+ },
+ {
+ "epoch": 41.41884816753927,
+ "grad_norm": 0.04697469249367714,
+ "learning_rate": 5.067556826472684e-05,
+ "loss": 3.4910900592803955,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43280977312391,
+ "grad_norm": 0.04753987118601799,
+ "learning_rate": 5.053505503468228e-05,
+ "loss": 3.504899024963379,
+ "step": 2983
+ },
+ {
+ "epoch": 41.446771378708554,
+ "grad_norm": 0.04830222949385643,
+ "learning_rate": 5.0394745211083785e-05,
+ "loss": 3.487231492996216,
+ "step": 2984
+ },
+ {
+ "epoch": 41.460732984293195,
+ "grad_norm": 0.04917479306459427,
+ "learning_rate": 5.025463890697655e-05,
+ "loss": 3.491119146347046,
+ "step": 2985
+ },
+ {
+ "epoch": 41.474694589877835,
+ "grad_norm": 0.047515422105789185,
+ "learning_rate": 5.011473623524159e-05,
+ "loss": 3.4952025413513184,
+ "step": 2986
+ },
+ {
+ "epoch": 41.488656195462475,
+ "grad_norm": 0.04785816743969917,
+ "learning_rate": 4.9975037308595864e-05,
+ "loss": 3.5042450428009033,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50261780104712,
+ "grad_norm": 0.04761168360710144,
+ "learning_rate": 4.983554223959257e-05,
+ "loss": 3.4989638328552246,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51657940663176,
+ "grad_norm": 0.04817018285393715,
+ "learning_rate": 4.9696251140620134e-05,
+ "loss": 3.473546028137207,
+ "step": 2989
+ },
+ {
+ "epoch": 41.5305410122164,
+ "grad_norm": 0.047623272985219955,
+ "learning_rate": 4.9557164123902924e-05,
+ "loss": 3.5037989616394043,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544502617801044,
+ "grad_norm": 0.046707481145858765,
+ "learning_rate": 4.9418281301500844e-05,
+ "loss": 3.521557569503784,
+ "step": 2991
+ },
+ {
+ "epoch": 41.55846422338569,
+ "grad_norm": 0.048793643712997437,
+ "learning_rate": 4.9279602785309365e-05,
+ "loss": 3.4975666999816895,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57242582897033,
+ "grad_norm": 0.048892680555582047,
+ "learning_rate": 4.914112868705911e-05,
+ "loss": 3.499896764755249,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58638743455497,
+ "grad_norm": 0.050392571836709976,
+ "learning_rate": 4.900285911831624e-05,
+ "loss": 3.494752883911133,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60034904013962,
+ "grad_norm": 0.049144912511110306,
+ "learning_rate": 4.886479419048201e-05,
+ "loss": 3.4834141731262207,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61431064572426,
+ "grad_norm": 0.04847243055701256,
+ "learning_rate": 4.872693401479292e-05,
+ "loss": 3.452786922454834,
+ "step": 2996
+ },
+ {
+ "epoch": 41.6282722513089,
+ "grad_norm": 0.04898214712738991,
+ "learning_rate": 4.858927870232039e-05,
+ "loss": 3.495100975036621,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64223385689354,
+ "grad_norm": 0.048891909420490265,
+ "learning_rate": 4.845182836397068e-05,
+ "loss": 3.499159574508667,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65619546247819,
+ "grad_norm": 0.049408622086048126,
+ "learning_rate": 4.831458311048532e-05,
+ "loss": 3.4985241889953613,
+ "step": 2999
+ },
+ {
+ "epoch": 41.67015706806283,
+ "grad_norm": 0.04940970614552498,
+ "learning_rate": 4.817754305244008e-05,
+ "loss": 3.5193381309509277,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68411867364747,
+ "grad_norm": 0.04961863160133362,
+ "learning_rate": 4.804070830024577e-05,
+ "loss": 3.5064754486083984,
+ "step": 3001
+ },
+ {
+ "epoch": 41.69808027923211,
+ "grad_norm": 0.049569353461265564,
+ "learning_rate": 4.7904078964147654e-05,
+ "loss": 3.4965577125549316,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712041884816756,
+ "grad_norm": 0.04980974644422531,
+ "learning_rate": 4.776765515422557e-05,
+ "loss": 3.5100278854370117,
+ "step": 3003
+ },
+ {
+ "epoch": 41.726003490401396,
+ "grad_norm": 0.0492531843483448,
+ "learning_rate": 4.763143698039371e-05,
+ "loss": 3.4880599975585938,
+ "step": 3004
+ },
+ {
+ "epoch": 41.73996509598604,
+ "grad_norm": 0.05068065971136093,
+ "learning_rate": 4.749542455240043e-05,
+ "loss": 3.5037388801574707,
+ "step": 3005
+ },
+ {
+ "epoch": 41.753926701570684,
+ "grad_norm": 0.04909749701619148,
+ "learning_rate": 4.735961797982876e-05,
+ "loss": 3.4796335697174072,
+ "step": 3006
+ },
+ {
+ "epoch": 41.767888307155324,
+ "grad_norm": 0.05050486698746681,
+ "learning_rate": 4.7224017372095414e-05,
+ "loss": 3.4703454971313477,
+ "step": 3007
+ },
+ {
+ "epoch": 41.781849912739965,
+ "grad_norm": 0.0496763177216053,
+ "learning_rate": 4.708862283845143e-05,
+ "loss": 3.494141101837158,
+ "step": 3008
+ },
+ {
+ "epoch": 41.795811518324605,
+ "grad_norm": 0.0503719225525856,
+ "learning_rate": 4.695343448798163e-05,
+ "loss": 3.5245471000671387,
+ "step": 3009
+ },
+ {
+ "epoch": 41.80977312390925,
+ "grad_norm": 0.04876178130507469,
+ "learning_rate": 4.681845242960492e-05,
+ "loss": 3.485657215118408,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82373472949389,
+ "grad_norm": 0.05160895735025406,
+ "learning_rate": 4.668367677207398e-05,
+ "loss": 3.5181169509887695,
+ "step": 3011
+ },
+ {
+ "epoch": 41.83769633507853,
+ "grad_norm": 0.04962378367781639,
+ "learning_rate": 4.654910762397499e-05,
+ "loss": 3.4959146976470947,
+ "step": 3012
+ },
+ {
+ "epoch": 41.85165794066317,
+ "grad_norm": 0.04914521425962448,
+ "learning_rate": 4.6414745093727913e-05,
+ "loss": 3.4929981231689453,
+ "step": 3013
+ },
+ {
+ "epoch": 41.86561954624782,
+ "grad_norm": 0.049995288252830505,
+ "learning_rate": 4.6280589289586255e-05,
+ "loss": 3.4831490516662598,
+ "step": 3014
+ },
+ {
+ "epoch": 41.87958115183246,
+ "grad_norm": 0.04932008311152458,
+ "learning_rate": 4.614664031963692e-05,
+ "loss": 3.4770588874816895,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8935427574171,
+ "grad_norm": 0.050988808274269104,
+ "learning_rate": 4.601289829180004e-05,
+ "loss": 3.5015549659729004,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90750436300174,
+ "grad_norm": 0.049767497926950455,
+ "learning_rate": 4.587936331382934e-05,
+ "loss": 3.480461597442627,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92146596858639,
+ "grad_norm": 0.050207316875457764,
+ "learning_rate": 4.574603549331151e-05,
+ "loss": 3.522735118865967,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93542757417103,
+ "grad_norm": 0.04978947341442108,
+ "learning_rate": 4.561291493766625e-05,
+ "loss": 3.5139577388763428,
+ "step": 3019
+ },
+ {
+ "epoch": 41.94938917975567,
+ "grad_norm": 0.04990898072719574,
+ "learning_rate": 4.5480001754146455e-05,
+ "loss": 3.512300491333008,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96335078534032,
+ "grad_norm": 0.050290387123823166,
+ "learning_rate": 4.5347296049837875e-05,
+ "loss": 3.4925060272216797,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97731239092496,
+ "grad_norm": 0.049074798822402954,
+ "learning_rate": 4.52147979316592e-05,
+ "loss": 3.4772558212280273,
+ "step": 3022
+ },
+ {
+ "epoch": 41.9912739965096,
+ "grad_norm": 0.04929618537425995,
+ "learning_rate": 4.5082507506361574e-05,
+ "loss": 3.4907312393188477,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.03784853219985962,
+ "learning_rate": 4.495042488052904e-05,
+ "loss": 2.182621955871582,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 0.6221460700035095,
+ "eval_runtime": 60.4686,
+ "eval_samples_per_second": 40.385,
+ "eval_steps_per_second": 0.645,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01396160558464,
+ "grad_norm": 0.0446116179227829,
+ "learning_rate": 4.481855016057839e-05,
+ "loss": 3.5182933807373047,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02792321116929,
+ "grad_norm": 0.04415718466043472,
+ "learning_rate": 4.468688345275848e-05,
+ "loss": 3.4862308502197266,
+ "step": 3026
+ },
+ {
+ "epoch": 42.04188481675393,
+ "grad_norm": 0.043365973979234695,
+ "learning_rate": 4.455542486315086e-05,
+ "loss": 3.4973888397216797,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05584642233857,
+ "grad_norm": 0.045916847884655,
+ "learning_rate": 4.442417449766938e-05,
+ "loss": 3.4444098472595215,
+ "step": 3028
+ },
+ {
+ "epoch": 42.06980802792321,
+ "grad_norm": 0.04477037489414215,
+ "learning_rate": 4.429313246206014e-05,
+ "loss": 3.4581756591796875,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083769633507856,
+ "grad_norm": 0.04387771338224411,
+ "learning_rate": 4.416229886190119e-05,
+ "loss": 3.4740853309631348,
+ "step": 3030
+ },
+ {
+ "epoch": 42.0977312390925,
+ "grad_norm": 0.043969713151454926,
+ "learning_rate": 4.403167380260281e-05,
+ "loss": 3.4641780853271484,
+ "step": 3031
+ },
+ {
+ "epoch": 42.11169284467714,
+ "grad_norm": 0.045231763273477554,
+ "learning_rate": 4.3901257389407475e-05,
+ "loss": 3.481898069381714,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12565445026178,
+ "grad_norm": 0.043941181153059006,
+ "learning_rate": 4.3771049727389075e-05,
+ "loss": 3.479301929473877,
+ "step": 3033
+ },
+ {
+ "epoch": 42.139616055846425,
+ "grad_norm": 0.04559679329395294,
+ "learning_rate": 4.364105092145377e-05,
+ "loss": 3.474754810333252,
+ "step": 3034
+ },
+ {
+ "epoch": 42.153577661431065,
+ "grad_norm": 0.044603366404771805,
+ "learning_rate": 4.3511261076339e-05,
+ "loss": 3.469996452331543,
+ "step": 3035
+ },
+ {
+ "epoch": 42.167539267015705,
+ "grad_norm": 0.04484036564826965,
+ "learning_rate": 4.3381680296614394e-05,
+ "loss": 3.474886417388916,
+ "step": 3036
+ },
+ {
+ "epoch": 42.181500872600346,
+ "grad_norm": 0.04458961635828018,
+ "learning_rate": 4.3252308686680626e-05,
+ "loss": 3.4669456481933594,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19546247818499,
+ "grad_norm": 0.04605228826403618,
+ "learning_rate": 4.3123146350770146e-05,
+ "loss": 3.4829368591308594,
+ "step": 3038
+ },
+ {
+ "epoch": 42.20942408376963,
+ "grad_norm": 0.04483795538544655,
+ "learning_rate": 4.2994193392946726e-05,
+ "loss": 3.4611988067626953,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223385689354274,
+ "grad_norm": 0.04524310305714607,
+ "learning_rate": 4.286544991710541e-05,
+ "loss": 3.4750466346740723,
+ "step": 3040
+ },
+ {
+ "epoch": 42.23734729493892,
+ "grad_norm": 0.04514177143573761,
+ "learning_rate": 4.2736916026972576e-05,
+ "loss": 3.456035614013672,
+ "step": 3041
+ },
+ {
+ "epoch": 42.25130890052356,
+ "grad_norm": 0.04549764469265938,
+ "learning_rate": 4.260859182610542e-05,
+ "loss": 3.450291633605957,
+ "step": 3042
+ },
+ {
+ "epoch": 42.2652705061082,
+ "grad_norm": 0.04712701961398125,
+ "learning_rate": 4.2480477417892776e-05,
+ "loss": 3.4654693603515625,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27923211169284,
+ "grad_norm": 0.045718107372522354,
+ "learning_rate": 4.235257290555381e-05,
+ "loss": 3.491769790649414,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29319371727749,
+ "grad_norm": 0.045064907521009445,
+ "learning_rate": 4.222487839213903e-05,
+ "loss": 3.49348783493042,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30715532286213,
+ "grad_norm": 0.04526374116539955,
+ "learning_rate": 4.209739398052956e-05,
+ "loss": 3.4869022369384766,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32111692844677,
+ "grad_norm": 0.045719344168901443,
+ "learning_rate": 4.1970119773437316e-05,
+ "loss": 3.484297752380371,
+ "step": 3047
+ },
+ {
+ "epoch": 42.33507853403141,
+ "grad_norm": 0.047497447580099106,
+ "learning_rate": 4.184305587340483e-05,
+ "loss": 3.4783167839050293,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34904013961606,
+ "grad_norm": 0.044837966561317444,
+ "learning_rate": 4.171620238280511e-05,
+ "loss": 3.460331678390503,
+ "step": 3049
+ },
+ {
+ "epoch": 42.3630017452007,
+ "grad_norm": 0.0457470640540123,
+ "learning_rate": 4.158955940384179e-05,
+ "loss": 3.4783971309661865,
+ "step": 3050
+ },
+ {
+ "epoch": 42.37696335078534,
+ "grad_norm": 0.04576839506626129,
+ "learning_rate": 4.146312703854881e-05,
+ "loss": 3.4972984790802,
+ "step": 3051
+ },
+ {
+ "epoch": 42.390924956369986,
+ "grad_norm": 0.04589150846004486,
+ "learning_rate": 4.133690538879046e-05,
+ "loss": 3.4743404388427734,
+ "step": 3052
+ },
+ {
+ "epoch": 42.404886561954626,
+ "grad_norm": 0.04680817574262619,
+ "learning_rate": 4.1210894556261226e-05,
+ "loss": 3.4869651794433594,
+ "step": 3053
+ },
+ {
+ "epoch": 42.41884816753927,
+ "grad_norm": 0.04661082476377487,
+ "learning_rate": 4.108509464248578e-05,
+ "loss": 3.47898530960083,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43280977312391,
+ "grad_norm": 0.04658796265721321,
+ "learning_rate": 4.095950574881891e-05,
+ "loss": 3.446974277496338,
+ "step": 3055
+ },
+ {
+ "epoch": 42.446771378708554,
+ "grad_norm": 0.0446607768535614,
+ "learning_rate": 4.083412797644514e-05,
+ "loss": 3.481013298034668,
+ "step": 3056
+ },
+ {
+ "epoch": 42.460732984293195,
+ "grad_norm": 0.04672912880778313,
+ "learning_rate": 4.0708961426379214e-05,
+ "loss": 3.493116855621338,
+ "step": 3057
+ },
+ {
+ "epoch": 42.474694589877835,
+ "grad_norm": 0.04644620046019554,
+ "learning_rate": 4.0584006199465484e-05,
+ "loss": 3.476749897003174,
+ "step": 3058
+ },
+ {
+ "epoch": 42.488656195462475,
+ "grad_norm": 0.0467224083840847,
+ "learning_rate": 4.0459262396378165e-05,
+ "loss": 3.440511703491211,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50261780104712,
+ "grad_norm": 0.045566949993371964,
+ "learning_rate": 4.033473011762116e-05,
+ "loss": 3.4723963737487793,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51657940663176,
+ "grad_norm": 0.04642454907298088,
+ "learning_rate": 4.0210409463527656e-05,
+ "loss": 3.4766931533813477,
+ "step": 3061
+ },
+ {
+ "epoch": 42.5305410122164,
+ "grad_norm": 0.046170495450496674,
+ "learning_rate": 4.008630053426082e-05,
+ "loss": 3.4428517818450928,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544502617801044,
+ "grad_norm": 0.046656228601932526,
+ "learning_rate": 3.996240342981279e-05,
+ "loss": 3.490520477294922,
+ "step": 3063
+ },
+ {
+ "epoch": 42.55846422338569,
+ "grad_norm": 0.04764910787343979,
+ "learning_rate": 3.983871825000529e-05,
+ "loss": 3.478440761566162,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57242582897033,
+ "grad_norm": 0.04596332833170891,
+ "learning_rate": 3.971524509448925e-05,
+ "loss": 3.4757537841796875,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58638743455497,
+ "grad_norm": 0.047473009675741196,
+ "learning_rate": 3.959198406274472e-05,
+ "loss": 3.4758293628692627,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60034904013962,
+ "grad_norm": 0.046200625598430634,
+ "learning_rate": 3.946893525408097e-05,
+ "loss": 3.476270914077759,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61431064572426,
+ "grad_norm": 0.0474882535636425,
+ "learning_rate": 3.934609876763604e-05,
+ "loss": 3.4846224784851074,
+ "step": 3068
+ },
+ {
+ "epoch": 42.6282722513089,
+ "grad_norm": 0.046172887086868286,
+ "learning_rate": 3.922347470237728e-05,
+ "loss": 3.4917209148406982,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64223385689354,
+ "grad_norm": 0.047964416444301605,
+ "learning_rate": 3.910106315710053e-05,
+ "loss": 3.493190050125122,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65619546247819,
+ "grad_norm": 0.048526931554079056,
+ "learning_rate": 3.897886423043061e-05,
+ "loss": 3.471339225769043,
+ "step": 3071
+ },
+ {
+ "epoch": 42.67015706806283,
+ "grad_norm": 0.04729713499546051,
+ "learning_rate": 3.8856878020820975e-05,
+ "loss": 3.4923644065856934,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68411867364747,
+ "grad_norm": 0.0482463575899601,
+ "learning_rate": 3.873510462655375e-05,
+ "loss": 3.496248722076416,
+ "step": 3073
+ },
+ {
+ "epoch": 42.69808027923211,
+ "grad_norm": 0.04724344238638878,
+ "learning_rate": 3.861354414573954e-05,
+ "loss": 3.4864397048950195,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712041884816756,
+ "grad_norm": 0.04727556183934212,
+ "learning_rate": 3.849219667631735e-05,
+ "loss": 3.483755111694336,
+ "step": 3075
+ },
+ {
+ "epoch": 42.726003490401396,
+ "grad_norm": 0.04790109395980835,
+ "learning_rate": 3.8371062316054764e-05,
+ "loss": 3.4795761108398438,
+ "step": 3076
+ },
+ {
+ "epoch": 42.73996509598604,
+ "grad_norm": 0.04786139354109764,
+ "learning_rate": 3.8250141162547466e-05,
+ "loss": 3.502222776412964,
+ "step": 3077
+ },
+ {
+ "epoch": 42.753926701570684,
+ "grad_norm": 0.048031169921159744,
+ "learning_rate": 3.812943331321956e-05,
+ "loss": 3.456325054168701,
+ "step": 3078
+ },
+ {
+ "epoch": 42.767888307155324,
+ "grad_norm": 0.04753265529870987,
+ "learning_rate": 3.8008938865322955e-05,
+ "loss": 3.482483386993408,
+ "step": 3079
+ },
+ {
+ "epoch": 42.781849912739965,
+ "grad_norm": 0.04813816025853157,
+ "learning_rate": 3.788865791593813e-05,
+ "loss": 3.4626238346099854,
+ "step": 3080
+ },
+ {
+ "epoch": 42.795811518324605,
+ "grad_norm": 0.04639475792646408,
+ "learning_rate": 3.7768590561973064e-05,
+ "loss": 3.484825611114502,
+ "step": 3081
+ },
+ {
+ "epoch": 42.80977312390925,
+ "grad_norm": 0.048058658838272095,
+ "learning_rate": 3.764873690016394e-05,
+ "loss": 3.479367733001709,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82373472949389,
+ "grad_norm": 0.04669356718659401,
+ "learning_rate": 3.7529097027074654e-05,
+ "loss": 3.4717469215393066,
+ "step": 3083
+ },
+ {
+ "epoch": 42.83769633507853,
+ "grad_norm": 0.04729066789150238,
+ "learning_rate": 3.7409671039096936e-05,
+ "loss": 3.448122978210449,
+ "step": 3084
+ },
+ {
+ "epoch": 42.85165794066317,
+ "grad_norm": 0.04941864311695099,
+ "learning_rate": 3.7290459032450184e-05,
+ "loss": 3.4794251918792725,
+ "step": 3085
+ },
+ {
+ "epoch": 42.86561954624782,
+ "grad_norm": 0.047708433121442795,
+ "learning_rate": 3.717146110318123e-05,
+ "loss": 3.4708971977233887,
+ "step": 3086
+ },
+ {
+ "epoch": 42.87958115183246,
+ "grad_norm": 0.04743608459830284,
+ "learning_rate": 3.7052677347164614e-05,
+ "loss": 3.487725019454956,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8935427574171,
+ "grad_norm": 0.047558411955833435,
+ "learning_rate": 3.693410786010231e-05,
+ "loss": 3.488107204437256,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90750436300174,
+ "grad_norm": 0.0468779057264328,
+ "learning_rate": 3.6815752737523537e-05,
+ "loss": 3.4700937271118164,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92146596858639,
+ "grad_norm": 0.04860694333910942,
+ "learning_rate": 3.6697612074784944e-05,
+ "loss": 3.4898247718811035,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93542757417103,
+ "grad_norm": 0.0477910116314888,
+ "learning_rate": 3.657968596707029e-05,
+ "loss": 3.497626781463623,
+ "step": 3091
+ },
+ {
+ "epoch": 42.94938917975567,
+ "grad_norm": 0.04593159258365631,
+ "learning_rate": 3.64619745093906e-05,
+ "loss": 3.4628067016601562,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96335078534032,
+ "grad_norm": 0.048044558614492416,
+ "learning_rate": 3.6344477796583714e-05,
+ "loss": 3.4711837768554688,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97731239092496,
+ "grad_norm": 0.04747995361685753,
+ "learning_rate": 3.622719592331472e-05,
+ "loss": 3.483624219894409,
+ "step": 3094
+ },
+ {
+ "epoch": 42.9912739965096,
+ "grad_norm": 0.04785338044166565,
+ "learning_rate": 3.611012898407544e-05,
+ "loss": 3.4870688915252686,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.036172445863485336,
+ "learning_rate": 3.599327707318464e-05,
+ "loss": 2.202502965927124,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 0.6233968734741211,
+ "eval_runtime": 59.5357,
+ "eval_samples_per_second": 41.017,
+ "eval_steps_per_second": 0.655,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01396160558464,
+ "grad_norm": 0.041919440031051636,
+ "learning_rate": 3.587664028478779e-05,
+ "loss": 3.4489996433258057,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02792321116929,
+ "grad_norm": 0.041333068162202835,
+ "learning_rate": 3.5760218712857046e-05,
+ "loss": 3.459347724914551,
+ "step": 3098
+ },
+ {
+ "epoch": 43.04188481675393,
+ "grad_norm": 0.04244478419423103,
+ "learning_rate": 3.564401245119119e-05,
+ "loss": 3.469055414199829,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05584642233857,
+ "grad_norm": 0.04218090698122978,
+ "learning_rate": 3.552802159341543e-05,
+ "loss": 3.4696848392486572,
+ "step": 3100
+ },
+ {
+ "epoch": 43.06980802792321,
+ "grad_norm": 0.04310792684555054,
+ "learning_rate": 3.541224623298155e-05,
+ "loss": 3.445007801055908,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083769633507856,
+ "grad_norm": 0.043118201196193695,
+ "learning_rate": 3.529668646316765e-05,
+ "loss": 3.4675815105438232,
+ "step": 3102
+ },
+ {
+ "epoch": 43.0977312390925,
+ "grad_norm": 0.04221854358911514,
+ "learning_rate": 3.5181342377078166e-05,
+ "loss": 3.4767966270446777,
+ "step": 3103
+ },
+ {
+ "epoch": 43.11169284467714,
+ "grad_norm": 0.043059397488832474,
+ "learning_rate": 3.5066214067643785e-05,
+ "loss": 3.4651613235473633,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12565445026178,
+ "grad_norm": 0.04187485948204994,
+ "learning_rate": 3.495130162762119e-05,
+ "loss": 3.446305751800537,
+ "step": 3105
+ },
+ {
+ "epoch": 43.139616055846425,
+ "grad_norm": 0.04304322227835655,
+ "learning_rate": 3.483660514959343e-05,
+ "loss": 3.4663028717041016,
+ "step": 3106
+ },
+ {
+ "epoch": 43.153577661431065,
+ "grad_norm": 0.04343954101204872,
+ "learning_rate": 3.4722124725969235e-05,
+ "loss": 3.4554028511047363,
+ "step": 3107
+ },
+ {
+ "epoch": 43.167539267015705,
+ "grad_norm": 0.04271049425005913,
+ "learning_rate": 3.460786044898346e-05,
+ "loss": 3.4608445167541504,
+ "step": 3108
+ },
+ {
+ "epoch": 43.181500872600346,
+ "grad_norm": 0.042723026126623154,
+ "learning_rate": 3.4493812410696784e-05,
+ "loss": 3.4895715713500977,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19546247818499,
+ "grad_norm": 0.044030871242284775,
+ "learning_rate": 3.437998070299564e-05,
+ "loss": 3.4766366481781006,
+ "step": 3110
+ },
+ {
+ "epoch": 43.20942408376963,
+ "grad_norm": 0.04285202547907829,
+ "learning_rate": 3.426636541759225e-05,
+ "loss": 3.458911418914795,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223385689354274,
+ "grad_norm": 0.04367215558886528,
+ "learning_rate": 3.415296664602425e-05,
+ "loss": 3.4974024295806885,
+ "step": 3112
+ },
+ {
+ "epoch": 43.23734729493892,
+ "grad_norm": 0.04419645667076111,
+ "learning_rate": 3.403978447965507e-05,
+ "loss": 3.485166549682617,
+ "step": 3113
+ },
+ {
+ "epoch": 43.25130890052356,
+ "grad_norm": 0.04392880201339722,
+ "learning_rate": 3.392681900967355e-05,
+ "loss": 3.459162473678589,
+ "step": 3114
+ },
+ {
+ "epoch": 43.2652705061082,
+ "grad_norm": 0.04220719635486603,
+ "learning_rate": 3.381407032709395e-05,
+ "loss": 3.4838223457336426,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27923211169284,
+ "grad_norm": 0.04468657821416855,
+ "learning_rate": 3.3701538522755714e-05,
+ "loss": 3.4502077102661133,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29319371727749,
+ "grad_norm": 0.04435672610998154,
+ "learning_rate": 3.35892236873238e-05,
+ "loss": 3.4528560638427734,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30715532286213,
+ "grad_norm": 0.044321656227111816,
+ "learning_rate": 3.3477125911288284e-05,
+ "loss": 3.463057041168213,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32111692844677,
+ "grad_norm": 0.044570039957761765,
+ "learning_rate": 3.3365245284964225e-05,
+ "loss": 3.42390775680542,
+ "step": 3119
+ },
+ {
+ "epoch": 43.33507853403141,
+ "grad_norm": 0.04383990913629532,
+ "learning_rate": 3.3253581898491836e-05,
+ "loss": 3.468905210494995,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34904013961606,
+ "grad_norm": 0.044629745185375214,
+ "learning_rate": 3.3142135841836276e-05,
+ "loss": 3.470247983932495,
+ "step": 3121
+ },
+ {
+ "epoch": 43.3630017452007,
+ "grad_norm": 0.04445461556315422,
+ "learning_rate": 3.303090720478773e-05,
+ "loss": 3.4802684783935547,
+ "step": 3122
+ },
+ {
+ "epoch": 43.37696335078534,
+ "grad_norm": 0.04420622065663338,
+ "learning_rate": 3.291989607696092e-05,
+ "loss": 3.470055103302002,
+ "step": 3123
+ },
+ {
+ "epoch": 43.390924956369986,
+ "grad_norm": 0.04278041422367096,
+ "learning_rate": 3.280910254779565e-05,
+ "loss": 3.4415011405944824,
+ "step": 3124
+ },
+ {
+ "epoch": 43.404886561954626,
+ "grad_norm": 0.04414992034435272,
+ "learning_rate": 3.269852670655629e-05,
+ "loss": 3.4492762088775635,
+ "step": 3125
+ },
+ {
+ "epoch": 43.41884816753927,
+ "grad_norm": 0.044683173298835754,
+ "learning_rate": 3.258816864233169e-05,
+ "loss": 3.473803997039795,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43280977312391,
+ "grad_norm": 0.043872565031051636,
+ "learning_rate": 3.247802844403536e-05,
+ "loss": 3.4689323902130127,
+ "step": 3127
+ },
+ {
+ "epoch": 43.446771378708554,
+ "grad_norm": 0.044481247663497925,
+ "learning_rate": 3.236810620040532e-05,
+ "loss": 3.474726438522339,
+ "step": 3128
+ },
+ {
+ "epoch": 43.460732984293195,
+ "grad_norm": 0.044009409844875336,
+ "learning_rate": 3.225840200000396e-05,
+ "loss": 3.4411134719848633,
+ "step": 3129
+ },
+ {
+ "epoch": 43.474694589877835,
+ "grad_norm": 0.04399845376610756,
+ "learning_rate": 3.214891593121788e-05,
+ "loss": 3.4436731338500977,
+ "step": 3130
+ },
+ {
+ "epoch": 43.488656195462475,
+ "grad_norm": 0.04343521595001221,
+ "learning_rate": 3.203964808225801e-05,
+ "loss": 3.452680826187134,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50261780104712,
+ "grad_norm": 0.04366961866617203,
+ "learning_rate": 3.193059854115968e-05,
+ "loss": 3.468557357788086,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51657940663176,
+ "grad_norm": 0.04462398961186409,
+ "learning_rate": 3.182176739578195e-05,
+ "loss": 3.4636874198913574,
+ "step": 3133
+ },
+ {
+ "epoch": 43.5305410122164,
+ "grad_norm": 0.04596228525042534,
+ "learning_rate": 3.171315473380815e-05,
+ "loss": 3.483391284942627,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544502617801044,
+ "grad_norm": 0.04510573297739029,
+ "learning_rate": 3.160476064274555e-05,
+ "loss": 3.4479193687438965,
+ "step": 3135
+ },
+ {
+ "epoch": 43.55846422338569,
+ "grad_norm": 0.044678136706352234,
+ "learning_rate": 3.149658520992537e-05,
+ "loss": 3.471526622772217,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57242582897033,
+ "grad_norm": 0.043749045580625534,
+ "learning_rate": 3.138862852250249e-05,
+ "loss": 3.4809961318969727,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58638743455497,
+ "grad_norm": 0.044585809111595154,
+ "learning_rate": 3.1280890667455694e-05,
+ "loss": 3.4486989974975586,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60034904013962,
+ "grad_norm": 0.04449521377682686,
+ "learning_rate": 3.1173371731587496e-05,
+ "loss": 3.4768877029418945,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61431064572426,
+ "grad_norm": 0.04507734626531601,
+ "learning_rate": 3.1066071801523905e-05,
+ "loss": 3.4656381607055664,
+ "step": 3140
+ },
+ {
+ "epoch": 43.6282722513089,
+ "grad_norm": 0.04444066062569618,
+ "learning_rate": 3.095899096371456e-05,
+ "loss": 3.449824333190918,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64223385689354,
+ "grad_norm": 0.04459245502948761,
+ "learning_rate": 3.0852129304432506e-05,
+ "loss": 3.4757838249206543,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65619546247819,
+ "grad_norm": 0.04539329186081886,
+ "learning_rate": 3.074548690977434e-05,
+ "loss": 3.467472791671753,
+ "step": 3143
+ },
+ {
+ "epoch": 43.67015706806283,
+ "grad_norm": 0.04432376101613045,
+ "learning_rate": 3.0639063865659834e-05,
+ "loss": 3.4804115295410156,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68411867364747,
+ "grad_norm": 0.044664494693279266,
+ "learning_rate": 3.0532860257832144e-05,
+ "loss": 3.458524703979492,
+ "step": 3145
+ },
+ {
+ "epoch": 43.69808027923211,
+ "grad_norm": 0.04534987732768059,
+ "learning_rate": 3.04268761718576e-05,
+ "loss": 3.472733497619629,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712041884816756,
+ "grad_norm": 0.04504983872175217,
+ "learning_rate": 3.0321111693125648e-05,
+ "loss": 3.447598934173584,
+ "step": 3147
+ },
+ {
+ "epoch": 43.726003490401396,
+ "grad_norm": 0.04506697878241539,
+ "learning_rate": 3.0215566906848906e-05,
+ "loss": 3.479231595993042,
+ "step": 3148
+ },
+ {
+ "epoch": 43.73996509598604,
+ "grad_norm": 0.04495571181178093,
+ "learning_rate": 3.0110241898062726e-05,
+ "loss": 3.4603395462036133,
+ "step": 3149
+ },
+ {
+ "epoch": 43.753926701570684,
+ "grad_norm": 0.04532467573881149,
+ "learning_rate": 3.0005136751625748e-05,
+ "loss": 3.4643521308898926,
+ "step": 3150
+ },
+ {
+ "epoch": 43.767888307155324,
+ "grad_norm": 0.046149928122758865,
+ "learning_rate": 2.9900251552219178e-05,
+ "loss": 3.466956615447998,
+ "step": 3151
+ },
+ {
+ "epoch": 43.781849912739965,
+ "grad_norm": 0.045516159385442734,
+ "learning_rate": 2.9795586384347202e-05,
+ "loss": 3.4780664443969727,
+ "step": 3152
+ },
+ {
+ "epoch": 43.795811518324605,
+ "grad_norm": 0.04423840343952179,
+ "learning_rate": 2.9691141332336494e-05,
+ "loss": 3.4812045097351074,
+ "step": 3153
+ },
+ {
+ "epoch": 43.80977312390925,
+ "grad_norm": 0.047003038227558136,
+ "learning_rate": 2.9586916480336735e-05,
+ "loss": 3.4673991203308105,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82373472949389,
+ "grad_norm": 0.045799724757671356,
+ "learning_rate": 2.9482911912320006e-05,
+ "loss": 3.4602036476135254,
+ "step": 3155
+ },
+ {
+ "epoch": 43.83769633507853,
+ "grad_norm": 0.04607130587100983,
+ "learning_rate": 2.9379127712080756e-05,
+ "loss": 3.4666333198547363,
+ "step": 3156
+ },
+ {
+ "epoch": 43.85165794066317,
+ "grad_norm": 0.04715318605303764,
+ "learning_rate": 2.9275563963236107e-05,
+ "loss": 3.4574172496795654,
+ "step": 3157
+ },
+ {
+ "epoch": 43.86561954624782,
+ "grad_norm": 0.044526107609272,
+ "learning_rate": 2.9172220749225548e-05,
+ "loss": 3.4708940982818604,
+ "step": 3158
+ },
+ {
+ "epoch": 43.87958115183246,
+ "grad_norm": 0.04484846815466881,
+ "learning_rate": 2.906909815331085e-05,
+ "loss": 3.466261386871338,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8935427574171,
+ "grad_norm": 0.04613117873668671,
+ "learning_rate": 2.896619625857588e-05,
+ "loss": 3.4540066719055176,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90750436300174,
+ "grad_norm": 0.0452883206307888,
+ "learning_rate": 2.8863515147927003e-05,
+ "loss": 3.4572741985321045,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92146596858639,
+ "grad_norm": 0.045408450067043304,
+ "learning_rate": 2.8761054904092514e-05,
+ "loss": 3.43621563911438,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93542757417103,
+ "grad_norm": 0.045606113970279694,
+ "learning_rate": 2.8658815609622665e-05,
+ "loss": 3.4194536209106445,
+ "step": 3163
+ },
+ {
+ "epoch": 43.94938917975567,
+ "grad_norm": 0.04547363147139549,
+ "learning_rate": 2.855679734688993e-05,
+ "loss": 3.4742746353149414,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96335078534032,
+ "grad_norm": 0.046189647167921066,
+ "learning_rate": 2.8455000198088502e-05,
+ "loss": 3.459886074066162,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97731239092496,
+ "grad_norm": 0.045051224529743195,
+ "learning_rate": 2.8353424245234582e-05,
+ "loss": 3.460083484649658,
+ "step": 3166
+ },
+ {
+ "epoch": 43.9912739965096,
+ "grad_norm": 0.04479827359318733,
+ "learning_rate": 2.825206957016599e-05,
+ "loss": 3.4630260467529297,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.03426015004515648,
+ "learning_rate": 2.8150936254542376e-05,
+ "loss": 2.1829705238342285,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 0.6243340373039246,
+ "eval_runtime": 59.6021,
+ "eval_samples_per_second": 40.972,
+ "eval_steps_per_second": 0.654,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01396160558464,
+ "grad_norm": 0.039502255618572235,
+ "learning_rate": 2.8050024379845118e-05,
+ "loss": 3.453655481338501,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02792321116929,
+ "grad_norm": 0.039988402277231216,
+ "learning_rate": 2.7949334027376986e-05,
+ "loss": 3.478839874267578,
+ "step": 3170
+ },
+ {
+ "epoch": 44.04188481675393,
+ "grad_norm": 0.04035341739654541,
+ "learning_rate": 2.7848865278262427e-05,
+ "loss": 3.432953357696533,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05584642233857,
+ "grad_norm": 0.03990163281559944,
+ "learning_rate": 2.7748618213447268e-05,
+ "loss": 3.463597297668457,
+ "step": 3172
+ },
+ {
+ "epoch": 44.06980802792321,
+ "grad_norm": 0.04220886528491974,
+ "learning_rate": 2.764859291369885e-05,
+ "loss": 3.427955150604248,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083769633507856,
+ "grad_norm": 0.041814520955085754,
+ "learning_rate": 2.7548789459605603e-05,
+ "loss": 3.471323013305664,
+ "step": 3174
+ },
+ {
+ "epoch": 44.0977312390925,
+ "grad_norm": 0.040427837520837784,
+ "learning_rate": 2.744920793157743e-05,
+ "loss": 3.4614925384521484,
+ "step": 3175
+ },
+ {
+ "epoch": 44.11169284467714,
+ "grad_norm": 0.04101467877626419,
+ "learning_rate": 2.7349848409845462e-05,
+ "loss": 3.4553816318511963,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12565445026178,
+ "grad_norm": 0.04197322204709053,
+ "learning_rate": 2.725071097446174e-05,
+ "loss": 3.4764347076416016,
+ "step": 3177
+ },
+ {
+ "epoch": 44.139616055846425,
+ "grad_norm": 0.042121630162000656,
+ "learning_rate": 2.715179570529959e-05,
+ "loss": 3.4732165336608887,
+ "step": 3178
+ },
+ {
+ "epoch": 44.153577661431065,
+ "grad_norm": 0.041242565959692,
+ "learning_rate": 2.7053102682053134e-05,
+ "loss": 3.4677939414978027,
+ "step": 3179
+ },
+ {
+ "epoch": 44.167539267015705,
+ "grad_norm": 0.040816668421030045,
+ "learning_rate": 2.6954631984237766e-05,
+ "loss": 3.4444072246551514,
+ "step": 3180
+ },
+ {
+ "epoch": 44.181500872600346,
+ "grad_norm": 0.04216597601771355,
+ "learning_rate": 2.6856383691189347e-05,
+ "loss": 3.450758934020996,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19546247818499,
+ "grad_norm": 0.04147190600633621,
+ "learning_rate": 2.675835788206485e-05,
+ "loss": 3.44083571434021,
+ "step": 3182
+ },
+ {
+ "epoch": 44.20942408376963,
+ "grad_norm": 0.04059838876128197,
+ "learning_rate": 2.6660554635841912e-05,
+ "loss": 3.446469783782959,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223385689354274,
+ "grad_norm": 0.042010124772787094,
+ "learning_rate": 2.6562974031318788e-05,
+ "loss": 3.419384002685547,
+ "step": 3184
+ },
+ {
+ "epoch": 44.23734729493892,
+ "grad_norm": 0.042179301381111145,
+ "learning_rate": 2.6465616147114485e-05,
+ "loss": 3.4438977241516113,
+ "step": 3185
+ },
+ {
+ "epoch": 44.25130890052356,
+ "grad_norm": 0.04143347218632698,
+ "learning_rate": 2.6368481061668344e-05,
+ "loss": 3.478097438812256,
+ "step": 3186
+ },
+ {
+ "epoch": 44.2652705061082,
+ "grad_norm": 0.041317183524370193,
+ "learning_rate": 2.6271568853240557e-05,
+ "loss": 3.4608001708984375,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27923211169284,
+ "grad_norm": 0.041217777878046036,
+ "learning_rate": 2.6174879599911347e-05,
+ "loss": 3.4443745613098145,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29319371727749,
+ "grad_norm": 0.04093816131353378,
+ "learning_rate": 2.6078413379581627e-05,
+ "loss": 3.443121910095215,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30715532286213,
+ "grad_norm": 0.040228866040706635,
+ "learning_rate": 2.5982170269972397e-05,
+ "loss": 3.441892623901367,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32111692844677,
+ "grad_norm": 0.042651765048503876,
+ "learning_rate": 2.588615034862507e-05,
+ "loss": 3.4565064907073975,
+ "step": 3191
+ },
+ {
+ "epoch": 44.33507853403141,
+ "grad_norm": 0.042239658534526825,
+ "learning_rate": 2.5790353692901158e-05,
+ "loss": 3.451909065246582,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34904013961606,
+ "grad_norm": 0.04031132906675339,
+ "learning_rate": 2.5694780379982238e-05,
+ "loss": 3.4511477947235107,
+ "step": 3193
+ },
+ {
+ "epoch": 44.3630017452007,
+ "grad_norm": 0.04258335009217262,
+ "learning_rate": 2.5599430486870055e-05,
+ "loss": 3.484065294265747,
+ "step": 3194
+ },
+ {
+ "epoch": 44.37696335078534,
+ "grad_norm": 0.041619960218667984,
+ "learning_rate": 2.5504304090386225e-05,
+ "loss": 3.4708070755004883,
+ "step": 3195
+ },
+ {
+ "epoch": 44.390924956369986,
+ "grad_norm": 0.041788406670093536,
+ "learning_rate": 2.5409401267172474e-05,
+ "loss": 3.4092917442321777,
+ "step": 3196
+ },
+ {
+ "epoch": 44.404886561954626,
+ "grad_norm": 0.0416782908141613,
+ "learning_rate": 2.5314722093690176e-05,
+ "loss": 3.4138784408569336,
+ "step": 3197
+ },
+ {
+ "epoch": 44.41884816753927,
+ "grad_norm": 0.042210739105939865,
+ "learning_rate": 2.522026664622071e-05,
+ "loss": 3.453545093536377,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43280977312391,
+ "grad_norm": 0.04266810044646263,
+ "learning_rate": 2.5126035000865124e-05,
+ "loss": 3.424774169921875,
+ "step": 3199
+ },
+ {
+ "epoch": 44.446771378708554,
+ "grad_norm": 0.04081276059150696,
+ "learning_rate": 2.503202723354413e-05,
+ "loss": 3.4538846015930176,
+ "step": 3200
+ },
+ {
+ "epoch": 44.460732984293195,
+ "grad_norm": 0.0413636788725853,
+ "learning_rate": 2.4938243419998047e-05,
+ "loss": 3.4658732414245605,
+ "step": 3201
+ },
+ {
+ "epoch": 44.474694589877835,
+ "grad_norm": 0.04199403524398804,
+ "learning_rate": 2.484468363578689e-05,
+ "loss": 3.4627909660339355,
+ "step": 3202
+ },
+ {
+ "epoch": 44.488656195462475,
+ "grad_norm": 0.04227396100759506,
+ "learning_rate": 2.475134795629004e-05,
+ "loss": 3.4517412185668945,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50261780104712,
+ "grad_norm": 0.04252436012029648,
+ "learning_rate": 2.4658236456706357e-05,
+ "loss": 3.4830689430236816,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51657940663176,
+ "grad_norm": 0.04169167950749397,
+ "learning_rate": 2.456534921205403e-05,
+ "loss": 3.450204372406006,
+ "step": 3205
+ },
+ {
+ "epoch": 44.5305410122164,
+ "grad_norm": 0.041747767478227615,
+ "learning_rate": 2.4472686297170822e-05,
+ "loss": 3.472952365875244,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544502617801044,
+ "grad_norm": 0.0426458977162838,
+ "learning_rate": 2.4380247786713403e-05,
+ "loss": 3.4559519290924072,
+ "step": 3207
+ },
+ {
+ "epoch": 44.55846422338569,
+ "grad_norm": 0.04284683242440224,
+ "learning_rate": 2.428803375515788e-05,
+ "loss": 3.4351186752319336,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57242582897033,
+ "grad_norm": 0.041979581117630005,
+ "learning_rate": 2.4196044276799397e-05,
+ "loss": 3.4520907402038574,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58638743455497,
+ "grad_norm": 0.042444370687007904,
+ "learning_rate": 2.410427942575228e-05,
+ "loss": 3.4589779376983643,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60034904013962,
+ "grad_norm": 0.04146238788962364,
+ "learning_rate": 2.4012739275949746e-05,
+ "loss": 3.468356132507324,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61431064572426,
+ "grad_norm": 0.043098706752061844,
+ "learning_rate": 2.3921423901144008e-05,
+ "loss": 3.422999620437622,
+ "step": 3212
+ },
+ {
+ "epoch": 44.6282722513089,
+ "grad_norm": 0.042554691433906555,
+ "learning_rate": 2.3830333374906342e-05,
+ "loss": 3.441878080368042,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64223385689354,
+ "grad_norm": 0.04226307198405266,
+ "learning_rate": 2.3739467770626637e-05,
+ "loss": 3.4962363243103027,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65619546247819,
+ "grad_norm": 0.041877422481775284,
+ "learning_rate": 2.3648827161513686e-05,
+ "loss": 3.471935510635376,
+ "step": 3215
+ },
+ {
+ "epoch": 44.67015706806283,
+ "grad_norm": 0.04196307063102722,
+ "learning_rate": 2.3558411620595038e-05,
+ "loss": 3.468240737915039,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68411867364747,
+ "grad_norm": 0.04305512458086014,
+ "learning_rate": 2.3468221220716844e-05,
+ "loss": 3.457876205444336,
+ "step": 3217
+ },
+ {
+ "epoch": 44.69808027923211,
+ "grad_norm": 0.04215867817401886,
+ "learning_rate": 2.33782560345438e-05,
+ "loss": 3.4575209617614746,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712041884816756,
+ "grad_norm": 0.041869666427373886,
+ "learning_rate": 2.328851613455932e-05,
+ "loss": 3.442275285720825,
+ "step": 3219
+ },
+ {
+ "epoch": 44.726003490401396,
+ "grad_norm": 0.04279697686433792,
+ "learning_rate": 2.3199001593065204e-05,
+ "loss": 3.4583919048309326,
+ "step": 3220
+ },
+ {
+ "epoch": 44.73996509598604,
+ "grad_norm": 0.04206348955631256,
+ "learning_rate": 2.3109712482181684e-05,
+ "loss": 3.4497156143188477,
+ "step": 3221
+ },
+ {
+ "epoch": 44.753926701570684,
+ "grad_norm": 0.0420316718518734,
+ "learning_rate": 2.3020648873847414e-05,
+ "loss": 3.4462804794311523,
+ "step": 3222
+ },
+ {
+ "epoch": 44.767888307155324,
+ "grad_norm": 0.0421106293797493,
+ "learning_rate": 2.2931810839819235e-05,
+ "loss": 3.461982250213623,
+ "step": 3223
+ },
+ {
+ "epoch": 44.781849912739965,
+ "grad_norm": 0.04225604981184006,
+ "learning_rate": 2.284319845167253e-05,
+ "loss": 3.440438747406006,
+ "step": 3224
+ },
+ {
+ "epoch": 44.795811518324605,
+ "grad_norm": 0.04363993555307388,
+ "learning_rate": 2.2754811780800532e-05,
+ "loss": 3.4595792293548584,
+ "step": 3225
+ },
+ {
+ "epoch": 44.80977312390925,
+ "grad_norm": 0.04277531057596207,
+ "learning_rate": 2.2666650898414827e-05,
+ "loss": 3.444552421569824,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82373472949389,
+ "grad_norm": 0.042936112731695175,
+ "learning_rate": 2.2578715875545095e-05,
+ "loss": 3.4602608680725098,
+ "step": 3227
+ },
+ {
+ "epoch": 44.83769633507853,
+ "grad_norm": 0.041564665734767914,
+ "learning_rate": 2.2491006783038976e-05,
+ "loss": 3.4477994441986084,
+ "step": 3228
+ },
+ {
+ "epoch": 44.85165794066317,
+ "grad_norm": 0.043317802250385284,
+ "learning_rate": 2.2403523691562092e-05,
+ "loss": 3.481790542602539,
+ "step": 3229
+ },
+ {
+ "epoch": 44.86561954624782,
+ "grad_norm": 0.042523398995399475,
+ "learning_rate": 2.2316266671598003e-05,
+ "loss": 3.432340621948242,
+ "step": 3230
+ },
+ {
+ "epoch": 44.87958115183246,
+ "grad_norm": 0.04249822348356247,
+ "learning_rate": 2.222923579344805e-05,
+ "loss": 3.4209306240081787,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8935427574171,
+ "grad_norm": 0.04233477637171745,
+ "learning_rate": 2.2142431127231502e-05,
+ "loss": 3.454221248626709,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90750436300174,
+ "grad_norm": 0.04137618467211723,
+ "learning_rate": 2.2055852742885322e-05,
+ "loss": 3.463860511779785,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92146596858639,
+ "grad_norm": 0.04398854076862335,
+ "learning_rate": 2.196950071016409e-05,
+ "loss": 3.4566519260406494,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93542757417103,
+ "grad_norm": 0.04244855418801308,
+ "learning_rate": 2.1883375098640134e-05,
+ "loss": 3.4292609691619873,
+ "step": 3235
+ },
+ {
+ "epoch": 44.94938917975567,
+ "grad_norm": 0.04317071661353111,
+ "learning_rate": 2.1797475977703352e-05,
+ "loss": 3.4438283443450928,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96335078534032,
+ "grad_norm": 0.04333546757698059,
+ "learning_rate": 2.1711803416560964e-05,
+ "loss": 3.420079231262207,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97731239092496,
+ "grad_norm": 0.04234454035758972,
+ "learning_rate": 2.16263574842379e-05,
+ "loss": 3.4233388900756836,
+ "step": 3238
+ },
+ {
+ "epoch": 44.9912739965096,
+ "grad_norm": 0.04304066300392151,
+ "learning_rate": 2.15411382495764e-05,
+ "loss": 3.4663498401641846,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.03293231502175331,
+ "learning_rate": 2.1456145781236057e-05,
+ "loss": 2.176724433898926,
+ "step": 3240
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.3825324108637798e+18,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3240/training_args.bin b/runs/l2r90-wd0033-ssl/checkpoint-3240/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..313c41733ff55284c06034a59bb885b30e40ffc8
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3240/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7ddd0bdddba68d3e19ab33477000f3fc684fe4ee8e3580eeb8c02646e5fc27ac
+size 4856
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-360/chat_template.jinja b/runs/l2r90-wd0033-ssl/checkpoint-360/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-360/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-360/config.json b/runs/l2r90-wd0033-ssl/checkpoint-360/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-360/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-360/generation_config.json b/runs/l2r90-wd0033-ssl/checkpoint-360/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-360/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-360/model.safetensors b/runs/l2r90-wd0033-ssl/checkpoint-360/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..ef339a755b28aec27ed1934a50c6ff45a2455033
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-360/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ad082ba115939acac18cf1abaaf531f8fec93fc01a9000a8b0dd8ef21acc295e
+size 583356232
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-360/optimizer.pt b/runs/l2r90-wd0033-ssl/checkpoint-360/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..e75396fe4b50c9e2043acde4a7b19b1426649e65
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-360/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7fab7c4ad57a97ac44904f3b09efa413fa5b4e6d4fb0ab1673ae699c5e4f180e
+size 1166825338
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-360/rng_state_0.pth b/runs/l2r90-wd0033-ssl/checkpoint-360/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..4e78f383cb1e7d6a2d8a1b079d14a1f3b3221638
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-360/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6e2010a01db67d9a5a7d71f395859654334844f890d83931a09da17ec1cd1ab1
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-360/rng_state_1.pth b/runs/l2r90-wd0033-ssl/checkpoint-360/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..839c184c57f95a9dcf1dd0a689ea22b2436936d6
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-360/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:305dc5a67afe9b6699a059d8c540b05e90e64fdceb539aa95425b71c657404a2
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-360/scheduler.pt b/runs/l2r90-wd0033-ssl/checkpoint-360/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..d63b22e047ab13a6c29adc08f78969e1d6e569ff
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-360/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2802ceffe363a15cca2d2192020ec6c680a5afa8cc5d08bb488000d79d1a8eb0
+size 1064
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-360/tokenizer.json b/runs/l2r90-wd0033-ssl/checkpoint-360/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-360/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-360/tokenizer_config.json b/runs/l2r90-wd0033-ssl/checkpoint-360/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-360/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-360/trainer_state.json b/runs/l2r90-wd0033-ssl/checkpoint-360/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..ad965c8b8ff9b55daf736ba29e56d3b6bfb0d177
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-360/trainer_state.json
@@ -0,0 +1,2586 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 5.0,
+ "eval_steps": 500,
+ "global_step": 360,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013961605584642234,
+ "grad_norm": 0.26733073592185974,
+ "learning_rate": 0.0,
+ "loss": 7.990724086761475,
+ "step": 1
+ },
+ {
+ "epoch": 0.027923211169284468,
+ "grad_norm": 0.26712340116500854,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 7.990110874176025,
+ "step": 2
+ },
+ {
+ "epoch": 0.041884816753926704,
+ "grad_norm": 0.2646932601928711,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 7.956935405731201,
+ "step": 3
+ },
+ {
+ "epoch": 0.055846422338568937,
+ "grad_norm": 0.24684545397758484,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 7.890539646148682,
+ "step": 4
+ },
+ {
+ "epoch": 0.06980802792321117,
+ "grad_norm": 0.21016643941402435,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 7.8146138191223145,
+ "step": 5
+ },
+ {
+ "epoch": 0.08376963350785341,
+ "grad_norm": 0.17992742359638214,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 7.724307060241699,
+ "step": 6
+ },
+ {
+ "epoch": 0.09773123909249563,
+ "grad_norm": 0.14356893301010132,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 7.645662307739258,
+ "step": 7
+ },
+ {
+ "epoch": 0.11169284467713787,
+ "grad_norm": 0.10830964893102646,
+ "learning_rate": 4.2e-05,
+ "loss": 7.582019329071045,
+ "step": 8
+ },
+ {
+ "epoch": 0.1256544502617801,
+ "grad_norm": 0.08265161514282227,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 7.5230793952941895,
+ "step": 9
+ },
+ {
+ "epoch": 0.13961605584642234,
+ "grad_norm": 0.06423710286617279,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 7.473812103271484,
+ "step": 10
+ },
+ {
+ "epoch": 0.15357766143106458,
+ "grad_norm": 0.05132246017456055,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 7.440786838531494,
+ "step": 11
+ },
+ {
+ "epoch": 0.16753926701570682,
+ "grad_norm": 0.040572404861450195,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 7.41290283203125,
+ "step": 12
+ },
+ {
+ "epoch": 0.18150087260034903,
+ "grad_norm": 0.03138705715537071,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 7.39749002456665,
+ "step": 13
+ },
+ {
+ "epoch": 0.19546247818499127,
+ "grad_norm": 0.02553008496761322,
+ "learning_rate": 7.8e-05,
+ "loss": 7.387092590332031,
+ "step": 14
+ },
+ {
+ "epoch": 0.2094240837696335,
+ "grad_norm": 0.022178977727890015,
+ "learning_rate": 8.4e-05,
+ "loss": 7.378993511199951,
+ "step": 15
+ },
+ {
+ "epoch": 0.22338568935427575,
+ "grad_norm": 0.019915293902158737,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 7.37220573425293,
+ "step": 16
+ },
+ {
+ "epoch": 0.23734729493891799,
+ "grad_norm": 0.017390629276633263,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 7.366759300231934,
+ "step": 17
+ },
+ {
+ "epoch": 0.2513089005235602,
+ "grad_norm": 0.015472771599888802,
+ "learning_rate": 0.000102,
+ "loss": 7.3652663230896,
+ "step": 18
+ },
+ {
+ "epoch": 0.26527050610820246,
+ "grad_norm": 0.013599214144051075,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 7.359714984893799,
+ "step": 19
+ },
+ {
+ "epoch": 0.2792321116928447,
+ "grad_norm": 0.012442861683666706,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 7.353302001953125,
+ "step": 20
+ },
+ {
+ "epoch": 0.2931937172774869,
+ "grad_norm": 0.010471965186297894,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 7.357417583465576,
+ "step": 21
+ },
+ {
+ "epoch": 0.30715532286212915,
+ "grad_norm": 0.009647224098443985,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 7.351763725280762,
+ "step": 22
+ },
+ {
+ "epoch": 0.32111692844677137,
+ "grad_norm": 0.008228210732340813,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 7.345434188842773,
+ "step": 23
+ },
+ {
+ "epoch": 0.33507853403141363,
+ "grad_norm": 0.007454514969140291,
+ "learning_rate": 0.000138,
+ "loss": 7.342520236968994,
+ "step": 24
+ },
+ {
+ "epoch": 0.34904013961605584,
+ "grad_norm": 0.0069856527261435986,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 7.338870048522949,
+ "step": 25
+ },
+ {
+ "epoch": 0.36300174520069806,
+ "grad_norm": 0.006091665010899305,
+ "learning_rate": 0.00015,
+ "loss": 7.328121662139893,
+ "step": 26
+ },
+ {
+ "epoch": 0.3769633507853403,
+ "grad_norm": 0.005525320768356323,
+ "learning_rate": 0.000156,
+ "loss": 7.335029602050781,
+ "step": 27
+ },
+ {
+ "epoch": 0.39092495636998253,
+ "grad_norm": 0.0061133443377912045,
+ "learning_rate": 0.000162,
+ "loss": 7.319417953491211,
+ "step": 28
+ },
+ {
+ "epoch": 0.4048865619546248,
+ "grad_norm": 0.006040909793227911,
+ "learning_rate": 0.000168,
+ "loss": 7.326376438140869,
+ "step": 29
+ },
+ {
+ "epoch": 0.418848167539267,
+ "grad_norm": 0.006003216840326786,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 7.319552421569824,
+ "step": 30
+ },
+ {
+ "epoch": 0.4328097731239092,
+ "grad_norm": 0.007096003741025925,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 7.323486328125,
+ "step": 31
+ },
+ {
+ "epoch": 0.4467713787085515,
+ "grad_norm": 0.007725459989160299,
+ "learning_rate": 0.000186,
+ "loss": 7.3190717697143555,
+ "step": 32
+ },
+ {
+ "epoch": 0.4607329842931937,
+ "grad_norm": 0.007338670082390308,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 7.317154884338379,
+ "step": 33
+ },
+ {
+ "epoch": 0.47469458987783597,
+ "grad_norm": 0.007709544617682695,
+ "learning_rate": 0.000198,
+ "loss": 7.311826705932617,
+ "step": 34
+ },
+ {
+ "epoch": 0.4886561954624782,
+ "grad_norm": 0.009249784983694553,
+ "learning_rate": 0.000204,
+ "loss": 7.322498798370361,
+ "step": 35
+ },
+ {
+ "epoch": 0.5026178010471204,
+ "grad_norm": 0.006950767710804939,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 7.314571380615234,
+ "step": 36
+ },
+ {
+ "epoch": 0.5165794066317626,
+ "grad_norm": 0.007583985570818186,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 7.309540748596191,
+ "step": 37
+ },
+ {
+ "epoch": 0.5305410122164049,
+ "grad_norm": 0.014731820672750473,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 7.307004928588867,
+ "step": 38
+ },
+ {
+ "epoch": 0.5445026178010471,
+ "grad_norm": 0.016276473179459572,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 7.299642562866211,
+ "step": 39
+ },
+ {
+ "epoch": 0.5584642233856894,
+ "grad_norm": 0.020018689334392548,
+ "learning_rate": 0.000234,
+ "loss": 7.298466682434082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5724258289703316,
+ "grad_norm": 0.014941536821424961,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 7.289968967437744,
+ "step": 41
+ },
+ {
+ "epoch": 0.5863874345549738,
+ "grad_norm": 0.018299812451004982,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 7.276564598083496,
+ "step": 42
+ },
+ {
+ "epoch": 0.6003490401396161,
+ "grad_norm": 0.01615075021982193,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.2575554847717285,
+ "step": 43
+ },
+ {
+ "epoch": 0.6143106457242583,
+ "grad_norm": 0.013063831254839897,
+ "learning_rate": 0.000258,
+ "loss": 7.253885269165039,
+ "step": 44
+ },
+ {
+ "epoch": 0.6282722513089005,
+ "grad_norm": 0.015211676247417927,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.245779037475586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6422338568935427,
+ "grad_norm": 0.02093500830233097,
+ "learning_rate": 0.00027,
+ "loss": 7.225130081176758,
+ "step": 46
+ },
+ {
+ "epoch": 0.6561954624781849,
+ "grad_norm": 0.01808053068816662,
+ "learning_rate": 0.000276,
+ "loss": 7.209654331207275,
+ "step": 47
+ },
+ {
+ "epoch": 0.6701570680628273,
+ "grad_norm": 0.02209644205868244,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.197124481201172,
+ "step": 48
+ },
+ {
+ "epoch": 0.6841186736474695,
+ "grad_norm": 0.023524878546595573,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.1841936111450195,
+ "step": 49
+ },
+ {
+ "epoch": 0.6980802792321117,
+ "grad_norm": 0.031148087233304977,
+ "learning_rate": 0.000294,
+ "loss": 7.188040733337402,
+ "step": 50
+ },
+ {
+ "epoch": 0.7120418848167539,
+ "grad_norm": 0.02931929938495159,
+ "learning_rate": 0.0003,
+ "loss": 7.17371940612793,
+ "step": 51
+ },
+ {
+ "epoch": 0.7260034904013961,
+ "grad_norm": 0.01611483097076416,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.1714324951171875,
+ "step": 52
+ },
+ {
+ "epoch": 0.7399650959860384,
+ "grad_norm": 0.02619907818734646,
+ "learning_rate": 0.000312,
+ "loss": 7.158261299133301,
+ "step": 53
+ },
+ {
+ "epoch": 0.7539267015706806,
+ "grad_norm": 0.049462560564279556,
+ "learning_rate": 0.000318,
+ "loss": 7.167076110839844,
+ "step": 54
+ },
+ {
+ "epoch": 0.7678883071553229,
+ "grad_norm": 0.09835070371627808,
+ "learning_rate": 0.000324,
+ "loss": 7.154292106628418,
+ "step": 55
+ },
+ {
+ "epoch": 0.7818499127399651,
+ "grad_norm": 0.11406510323286057,
+ "learning_rate": 0.00033,
+ "loss": 7.162016868591309,
+ "step": 56
+ },
+ {
+ "epoch": 0.7958115183246073,
+ "grad_norm": 0.027026303112506866,
+ "learning_rate": 0.000336,
+ "loss": 7.141416549682617,
+ "step": 57
+ },
+ {
+ "epoch": 0.8097731239092496,
+ "grad_norm": 0.06690579652786255,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.13959264755249,
+ "step": 58
+ },
+ {
+ "epoch": 0.8237347294938918,
+ "grad_norm": 0.020919431000947952,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.126513481140137,
+ "step": 59
+ },
+ {
+ "epoch": 0.837696335078534,
+ "grad_norm": 0.05618130415678024,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.124345779418945,
+ "step": 60
+ },
+ {
+ "epoch": 0.8516579406631762,
+ "grad_norm": 0.03402302414178848,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.100615501403809,
+ "step": 61
+ },
+ {
+ "epoch": 0.8656195462478184,
+ "grad_norm": 0.03646906837821007,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.116507530212402,
+ "step": 62
+ },
+ {
+ "epoch": 0.8795811518324608,
+ "grad_norm": 0.03076397068798542,
+ "learning_rate": 0.000372,
+ "loss": 7.097659111022949,
+ "step": 63
+ },
+ {
+ "epoch": 0.893542757417103,
+ "grad_norm": 0.03329646959900856,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.097290992736816,
+ "step": 64
+ },
+ {
+ "epoch": 0.9075043630017452,
+ "grad_norm": 0.01998630352318287,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.092268943786621,
+ "step": 65
+ },
+ {
+ "epoch": 0.9214659685863874,
+ "grad_norm": 0.03278940171003342,
+ "learning_rate": 0.00039,
+ "loss": 7.088292121887207,
+ "step": 66
+ },
+ {
+ "epoch": 0.9354275741710296,
+ "grad_norm": 0.024620421230793,
+ "learning_rate": 0.000396,
+ "loss": 7.078650951385498,
+ "step": 67
+ },
+ {
+ "epoch": 0.9493891797556719,
+ "grad_norm": 0.03750383481383324,
+ "learning_rate": 0.000402,
+ "loss": 7.052778244018555,
+ "step": 68
+ },
+ {
+ "epoch": 0.9633507853403142,
+ "grad_norm": 0.07558075338602066,
+ "learning_rate": 0.000408,
+ "loss": 7.045991897583008,
+ "step": 69
+ },
+ {
+ "epoch": 0.9773123909249564,
+ "grad_norm": 0.16021482646465302,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.023152828216553,
+ "step": 70
+ },
+ {
+ "epoch": 0.9912739965095986,
+ "grad_norm": 0.24981893599033356,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.067540645599365,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.10255520790815353,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 4.425640106201172,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8861758708953857,
+ "eval_runtime": 58.2699,
+ "eval_samples_per_second": 41.908,
+ "eval_steps_per_second": 0.669,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139616055846423,
+ "grad_norm": 0.13374797999858856,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.073373317718506,
+ "step": 73
+ },
+ {
+ "epoch": 1.0279232111692844,
+ "grad_norm": 0.05079387128353119,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.0243306159973145,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418848167539267,
+ "grad_norm": 0.10056068748235703,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.014476299285889,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558464223385688,
+ "grad_norm": 0.06995758414268494,
+ "learning_rate": 0.00045,
+ "loss": 7.011931419372559,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698080279232112,
+ "grad_norm": 0.0483892522752285,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 6.9822797775268555,
+ "step": 77
+ },
+ {
+ "epoch": 1.0837696335078535,
+ "grad_norm": 0.08408170193433762,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 6.964348793029785,
+ "step": 78
+ },
+ {
+ "epoch": 1.0977312390924956,
+ "grad_norm": 0.08027710020542145,
+ "learning_rate": 0.000468,
+ "loss": 6.96852445602417,
+ "step": 79
+ },
+ {
+ "epoch": 1.111692844677138,
+ "grad_norm": 0.05647159367799759,
+ "learning_rate": 0.000474,
+ "loss": 6.944484710693359,
+ "step": 80
+ },
+ {
+ "epoch": 1.12565445026178,
+ "grad_norm": 0.07464330643415451,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 6.9246745109558105,
+ "step": 81
+ },
+ {
+ "epoch": 1.1396160558464223,
+ "grad_norm": 0.04451674595475197,
+ "learning_rate": 0.000486,
+ "loss": 6.895403861999512,
+ "step": 82
+ },
+ {
+ "epoch": 1.1535776614310647,
+ "grad_norm": 0.05820256471633911,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 6.887338638305664,
+ "step": 83
+ },
+ {
+ "epoch": 1.1675392670157068,
+ "grad_norm": 0.02925296127796173,
+ "learning_rate": 0.000498,
+ "loss": 6.857797622680664,
+ "step": 84
+ },
+ {
+ "epoch": 1.181500872600349,
+ "grad_norm": 0.05096987634897232,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 6.854822158813477,
+ "step": 85
+ },
+ {
+ "epoch": 1.1954624781849912,
+ "grad_norm": 0.03483197093009949,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 6.840914726257324,
+ "step": 86
+ },
+ {
+ "epoch": 1.2094240837696335,
+ "grad_norm": 0.03323264792561531,
+ "learning_rate": 0.000516,
+ "loss": 6.834497451782227,
+ "step": 87
+ },
+ {
+ "epoch": 1.2233856893542758,
+ "grad_norm": 0.03290868178009987,
+ "learning_rate": 0.000522,
+ "loss": 6.820746898651123,
+ "step": 88
+ },
+ {
+ "epoch": 1.237347294938918,
+ "grad_norm": 0.025398844853043556,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 6.795746803283691,
+ "step": 89
+ },
+ {
+ "epoch": 1.2513089005235603,
+ "grad_norm": 0.030700596049427986,
+ "learning_rate": 0.000534,
+ "loss": 6.7867889404296875,
+ "step": 90
+ },
+ {
+ "epoch": 1.2652705061082026,
+ "grad_norm": 0.019589420408010483,
+ "learning_rate": 0.00054,
+ "loss": 6.77443790435791,
+ "step": 91
+ },
+ {
+ "epoch": 1.2792321116928447,
+ "grad_norm": 0.03313920646905899,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 6.762271881103516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2931937172774868,
+ "grad_norm": 0.025389468297362328,
+ "learning_rate": 0.000552,
+ "loss": 6.750258445739746,
+ "step": 93
+ },
+ {
+ "epoch": 1.307155322862129,
+ "grad_norm": 0.040296655148267746,
+ "learning_rate": 0.000558,
+ "loss": 6.729808807373047,
+ "step": 94
+ },
+ {
+ "epoch": 1.3211169284467714,
+ "grad_norm": 0.04735933244228363,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 6.737982749938965,
+ "step": 95
+ },
+ {
+ "epoch": 1.3350785340314135,
+ "grad_norm": 0.05529680848121643,
+ "learning_rate": 0.00057,
+ "loss": 6.735714912414551,
+ "step": 96
+ },
+ {
+ "epoch": 1.3490401396160558,
+ "grad_norm": 0.07148554176092148,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 6.708859920501709,
+ "step": 97
+ },
+ {
+ "epoch": 1.3630017452006982,
+ "grad_norm": 0.059708479791879654,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 6.708860397338867,
+ "step": 98
+ },
+ {
+ "epoch": 1.3769633507853403,
+ "grad_norm": 0.027111195027828217,
+ "learning_rate": 0.000588,
+ "loss": 6.686685085296631,
+ "step": 99
+ },
+ {
+ "epoch": 1.3909249563699826,
+ "grad_norm": 0.0500967763364315,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 6.676325798034668,
+ "step": 100
+ },
+ {
+ "epoch": 1.404886561954625,
+ "grad_norm": 0.044565796852111816,
+ "learning_rate": 0.0006,
+ "loss": 6.674530029296875,
+ "step": 101
+ },
+ {
+ "epoch": 1.418848167539267,
+ "grad_norm": 0.03608180209994316,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 6.653741359710693,
+ "step": 102
+ },
+ {
+ "epoch": 1.4328097731239091,
+ "grad_norm": 0.030394982546567917,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 6.643152236938477,
+ "step": 103
+ },
+ {
+ "epoch": 1.4467713787085514,
+ "grad_norm": 0.040185656398534775,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 6.628686904907227,
+ "step": 104
+ },
+ {
+ "epoch": 1.4607329842931938,
+ "grad_norm": 0.021049322560429573,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 6.610864639282227,
+ "step": 105
+ },
+ {
+ "epoch": 1.4746945898778359,
+ "grad_norm": 0.04164277762174606,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 6.602880954742432,
+ "step": 106
+ },
+ {
+ "epoch": 1.4886561954624782,
+ "grad_norm": 0.0389864556491375,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 6.599456787109375,
+ "step": 107
+ },
+ {
+ "epoch": 1.5026178010471205,
+ "grad_norm": 0.05805973336100578,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 6.588700294494629,
+ "step": 108
+ },
+ {
+ "epoch": 1.5165794066317626,
+ "grad_norm": 0.08825036883354187,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 6.587496757507324,
+ "step": 109
+ },
+ {
+ "epoch": 1.530541012216405,
+ "grad_norm": 0.12521523237228394,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 6.626501083374023,
+ "step": 110
+ },
+ {
+ "epoch": 1.5445026178010473,
+ "grad_norm": 0.07429645210504532,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 6.590241432189941,
+ "step": 111
+ },
+ {
+ "epoch": 1.5584642233856894,
+ "grad_norm": 0.06463009119033813,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 6.578714370727539,
+ "step": 112
+ },
+ {
+ "epoch": 1.5724258289703315,
+ "grad_norm": 0.07508203387260437,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 6.561270713806152,
+ "step": 113
+ },
+ {
+ "epoch": 1.5863874345549738,
+ "grad_norm": 0.04125197231769562,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 6.558821678161621,
+ "step": 114
+ },
+ {
+ "epoch": 1.600349040139616,
+ "grad_norm": 0.05457386001944542,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.541611671447754,
+ "step": 115
+ },
+ {
+ "epoch": 1.6143106457242582,
+ "grad_norm": 0.03684385493397713,
+ "learning_rate": 0.000599973080557108,
+ "loss": 6.547738075256348,
+ "step": 116
+ },
+ {
+ "epoch": 1.6282722513089005,
+ "grad_norm": 0.05888746306300163,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.529491424560547,
+ "step": 117
+ },
+ {
+ "epoch": 1.6422338568935428,
+ "grad_norm": 0.056668370962142944,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.519444942474365,
+ "step": 118
+ },
+ {
+ "epoch": 1.656195462478185,
+ "grad_norm": 0.09123896062374115,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.522984981536865,
+ "step": 119
+ },
+ {
+ "epoch": 1.6701570680628273,
+ "grad_norm": 0.12380023300647736,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.515375137329102,
+ "step": 120
+ },
+ {
+ "epoch": 1.6841186736474696,
+ "grad_norm": 0.06953322887420654,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.500491142272949,
+ "step": 121
+ },
+ {
+ "epoch": 1.6980802792321117,
+ "grad_norm": 0.08613869547843933,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.5097198486328125,
+ "step": 122
+ },
+ {
+ "epoch": 1.7120418848167538,
+ "grad_norm": 0.08296383172273636,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.487521171569824,
+ "step": 123
+ },
+ {
+ "epoch": 1.7260034904013961,
+ "grad_norm": 0.05744783207774162,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.485134124755859,
+ "step": 124
+ },
+ {
+ "epoch": 1.7399650959860384,
+ "grad_norm": 0.04108566418290138,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.466182708740234,
+ "step": 125
+ },
+ {
+ "epoch": 1.7539267015706805,
+ "grad_norm": 0.05374931916594505,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.479011535644531,
+ "step": 126
+ },
+ {
+ "epoch": 1.7678883071553229,
+ "grad_norm": 0.03867914527654648,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.465902328491211,
+ "step": 127
+ },
+ {
+ "epoch": 1.7818499127399652,
+ "grad_norm": 0.05477374792098999,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.453082084655762,
+ "step": 128
+ },
+ {
+ "epoch": 1.7958115183246073,
+ "grad_norm": 0.06508298963308334,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.446089744567871,
+ "step": 129
+ },
+ {
+ "epoch": 1.8097731239092496,
+ "grad_norm": 0.06388624757528305,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.451748847961426,
+ "step": 130
+ },
+ {
+ "epoch": 1.823734729493892,
+ "grad_norm": 0.04223188757896423,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.4492387771606445,
+ "step": 131
+ },
+ {
+ "epoch": 1.837696335078534,
+ "grad_norm": 0.04528658092021942,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.409390926361084,
+ "step": 132
+ },
+ {
+ "epoch": 1.8516579406631761,
+ "grad_norm": 0.04962967336177826,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.415774345397949,
+ "step": 133
+ },
+ {
+ "epoch": 1.8656195462478184,
+ "grad_norm": 0.031944390386343,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.411528587341309,
+ "step": 134
+ },
+ {
+ "epoch": 1.8795811518324608,
+ "grad_norm": 0.028411149978637695,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.392329216003418,
+ "step": 135
+ },
+ {
+ "epoch": 1.8935427574171029,
+ "grad_norm": 0.03516148403286934,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.392238616943359,
+ "step": 136
+ },
+ {
+ "epoch": 1.9075043630017452,
+ "grad_norm": 0.027800120413303375,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.39179801940918,
+ "step": 137
+ },
+ {
+ "epoch": 1.9214659685863875,
+ "grad_norm": 0.02362987957894802,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.378960609436035,
+ "step": 138
+ },
+ {
+ "epoch": 1.9354275741710296,
+ "grad_norm": 0.033999308943748474,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.36802339553833,
+ "step": 139
+ },
+ {
+ "epoch": 1.949389179755672,
+ "grad_norm": 0.03290879726409912,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.366308689117432,
+ "step": 140
+ },
+ {
+ "epoch": 1.9633507853403143,
+ "grad_norm": 0.0286843404173851,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.350399971008301,
+ "step": 141
+ },
+ {
+ "epoch": 1.9773123909249564,
+ "grad_norm": 0.027145877480506897,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.33477783203125,
+ "step": 142
+ },
+ {
+ "epoch": 1.9912739965095985,
+ "grad_norm": 0.03530878946185112,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.322859764099121,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.028133150190114975,
+ "learning_rate": 0.000599778806120077,
+ "loss": 3.955288887023926,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7927101254463196,
+ "eval_runtime": 60.8372,
+ "eval_samples_per_second": 40.14,
+ "eval_steps_per_second": 0.641,
+ "step": 144
+ },
+ {
+ "epoch": 2.013961605584642,
+ "grad_norm": 0.03054557554423809,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.311975479125977,
+ "step": 145
+ },
+ {
+ "epoch": 2.0279232111692846,
+ "grad_norm": 0.05406459793448448,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.3030853271484375,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418848167539267,
+ "grad_norm": 0.10911019891500473,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.329056739807129,
+ "step": 147
+ },
+ {
+ "epoch": 2.055846422338569,
+ "grad_norm": 0.13948315382003784,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.362421035766602,
+ "step": 148
+ },
+ {
+ "epoch": 2.0698080279232114,
+ "grad_norm": 0.13176991045475006,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.3888139724731445,
+ "step": 149
+ },
+ {
+ "epoch": 2.0837696335078535,
+ "grad_norm": 0.09615885466337204,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.35154914855957,
+ "step": 150
+ },
+ {
+ "epoch": 2.0977312390924956,
+ "grad_norm": 0.06462561339139938,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.335790634155273,
+ "step": 151
+ },
+ {
+ "epoch": 2.1116928446771377,
+ "grad_norm": 0.05186621472239494,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.324396133422852,
+ "step": 152
+ },
+ {
+ "epoch": 2.1256544502617802,
+ "grad_norm": 0.06095981225371361,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.318549156188965,
+ "step": 153
+ },
+ {
+ "epoch": 2.1396160558464223,
+ "grad_norm": 0.036121442914009094,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.300607681274414,
+ "step": 154
+ },
+ {
+ "epoch": 2.1535776614310644,
+ "grad_norm": 0.055716633796691895,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.307400703430176,
+ "step": 155
+ },
+ {
+ "epoch": 2.167539267015707,
+ "grad_norm": 0.037626396864652634,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.287364959716797,
+ "step": 156
+ },
+ {
+ "epoch": 2.181500872600349,
+ "grad_norm": 0.04329098388552666,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.28244686126709,
+ "step": 157
+ },
+ {
+ "epoch": 2.195462478184991,
+ "grad_norm": 0.031868450343608856,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.2602410316467285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2094240837696333,
+ "grad_norm": 0.03647344186902046,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.263988971710205,
+ "step": 159
+ },
+ {
+ "epoch": 2.223385689354276,
+ "grad_norm": 0.03141765296459198,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.263205051422119,
+ "step": 160
+ },
+ {
+ "epoch": 2.237347294938918,
+ "grad_norm": 0.028253229334950447,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.253107070922852,
+ "step": 161
+ },
+ {
+ "epoch": 2.25130890052356,
+ "grad_norm": 0.03608971834182739,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.2571702003479,
+ "step": 162
+ },
+ {
+ "epoch": 2.2652705061082026,
+ "grad_norm": 0.02644292451441288,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.241732597351074,
+ "step": 163
+ },
+ {
+ "epoch": 2.2792321116928447,
+ "grad_norm": 0.028604455292224884,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.230289459228516,
+ "step": 164
+ },
+ {
+ "epoch": 2.2931937172774868,
+ "grad_norm": 0.023704614490270615,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.226101398468018,
+ "step": 165
+ },
+ {
+ "epoch": 2.3071553228621293,
+ "grad_norm": 0.020691027864813805,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.2096405029296875,
+ "step": 166
+ },
+ {
+ "epoch": 2.3211169284467714,
+ "grad_norm": 0.025598427280783653,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.22075080871582,
+ "step": 167
+ },
+ {
+ "epoch": 2.3350785340314135,
+ "grad_norm": 0.030546002089977264,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.216886520385742,
+ "step": 168
+ },
+ {
+ "epoch": 2.349040139616056,
+ "grad_norm": 0.026066245511174202,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.2022504806518555,
+ "step": 169
+ },
+ {
+ "epoch": 2.363001745200698,
+ "grad_norm": 0.023802239447832108,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.189509391784668,
+ "step": 170
+ },
+ {
+ "epoch": 2.3769633507853403,
+ "grad_norm": 0.029805002734065056,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.186277389526367,
+ "step": 171
+ },
+ {
+ "epoch": 2.3909249563699824,
+ "grad_norm": 0.043327830731868744,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.164876937866211,
+ "step": 172
+ },
+ {
+ "epoch": 2.404886561954625,
+ "grad_norm": 0.058787211775779724,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.179936408996582,
+ "step": 173
+ },
+ {
+ "epoch": 2.418848167539267,
+ "grad_norm": 0.0787397176027298,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.1830244064331055,
+ "step": 174
+ },
+ {
+ "epoch": 2.432809773123909,
+ "grad_norm": 0.0899854376912117,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.190343856811523,
+ "step": 175
+ },
+ {
+ "epoch": 2.4467713787085517,
+ "grad_norm": 0.06520707905292511,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.185057163238525,
+ "step": 176
+ },
+ {
+ "epoch": 2.4607329842931938,
+ "grad_norm": 0.1002504974603653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.171279430389404,
+ "step": 177
+ },
+ {
+ "epoch": 2.474694589877836,
+ "grad_norm": 0.09591004997491837,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.174382209777832,
+ "step": 178
+ },
+ {
+ "epoch": 2.488656195462478,
+ "grad_norm": 0.12127121537923813,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.185717582702637,
+ "step": 179
+ },
+ {
+ "epoch": 2.5026178010471205,
+ "grad_norm": 0.09640590846538544,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.196681976318359,
+ "step": 180
+ },
+ {
+ "epoch": 2.5165794066317626,
+ "grad_norm": 0.048341453075408936,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.158774375915527,
+ "step": 181
+ },
+ {
+ "epoch": 2.530541012216405,
+ "grad_norm": 0.08206334710121155,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.17445707321167,
+ "step": 182
+ },
+ {
+ "epoch": 2.5445026178010473,
+ "grad_norm": 0.038934968411922455,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.170527458190918,
+ "step": 183
+ },
+ {
+ "epoch": 2.5584642233856894,
+ "grad_norm": 0.060987457633018494,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.1501994132995605,
+ "step": 184
+ },
+ {
+ "epoch": 2.5724258289703315,
+ "grad_norm": 0.033269546926021576,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.144186019897461,
+ "step": 185
+ },
+ {
+ "epoch": 2.5863874345549736,
+ "grad_norm": 0.046137236058712006,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.1236677169799805,
+ "step": 186
+ },
+ {
+ "epoch": 2.600349040139616,
+ "grad_norm": 0.038265109062194824,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.121979236602783,
+ "step": 187
+ },
+ {
+ "epoch": 2.614310645724258,
+ "grad_norm": 0.039128001779317856,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.124156951904297,
+ "step": 188
+ },
+ {
+ "epoch": 2.6282722513089007,
+ "grad_norm": 0.025303540751338005,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.108036994934082,
+ "step": 189
+ },
+ {
+ "epoch": 2.642233856893543,
+ "grad_norm": 0.04441308230161667,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.1211161613464355,
+ "step": 190
+ },
+ {
+ "epoch": 2.656195462478185,
+ "grad_norm": 0.03083532303571701,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.094561576843262,
+ "step": 191
+ },
+ {
+ "epoch": 2.670157068062827,
+ "grad_norm": 0.02951296605169773,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.091608047485352,
+ "step": 192
+ },
+ {
+ "epoch": 2.6841186736474696,
+ "grad_norm": 0.02385987527668476,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.087080001831055,
+ "step": 193
+ },
+ {
+ "epoch": 2.6980802792321117,
+ "grad_norm": 0.03315780311822891,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.088132381439209,
+ "step": 194
+ },
+ {
+ "epoch": 2.712041884816754,
+ "grad_norm": 0.02410835586488247,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.081460952758789,
+ "step": 195
+ },
+ {
+ "epoch": 2.7260034904013963,
+ "grad_norm": 0.033069487661123276,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.0696539878845215,
+ "step": 196
+ },
+ {
+ "epoch": 2.7399650959860384,
+ "grad_norm": 0.031288471072912216,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.047612190246582,
+ "step": 197
+ },
+ {
+ "epoch": 2.7539267015706805,
+ "grad_norm": 0.03399677947163582,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.019026756286621,
+ "step": 198
+ },
+ {
+ "epoch": 2.7678883071553226,
+ "grad_norm": 0.031948331743478775,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.050133228302002,
+ "step": 199
+ },
+ {
+ "epoch": 2.781849912739965,
+ "grad_norm": 0.049206435680389404,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.045332431793213,
+ "step": 200
+ },
+ {
+ "epoch": 2.7958115183246073,
+ "grad_norm": 0.07528849691152573,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.059202194213867,
+ "step": 201
+ },
+ {
+ "epoch": 2.80977312390925,
+ "grad_norm": 0.11815470457077026,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.077486038208008,
+ "step": 202
+ },
+ {
+ "epoch": 2.823734729493892,
+ "grad_norm": 0.0867348462343216,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.045670032501221,
+ "step": 203
+ },
+ {
+ "epoch": 2.837696335078534,
+ "grad_norm": 0.05716705322265625,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.034373760223389,
+ "step": 204
+ },
+ {
+ "epoch": 2.851657940663176,
+ "grad_norm": 0.05944284051656723,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.049911975860596,
+ "step": 205
+ },
+ {
+ "epoch": 2.8656195462478182,
+ "grad_norm": 0.03220470994710922,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.039861679077148,
+ "step": 206
+ },
+ {
+ "epoch": 2.8795811518324608,
+ "grad_norm": 0.057081177830696106,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.010714530944824,
+ "step": 207
+ },
+ {
+ "epoch": 2.893542757417103,
+ "grad_norm": 0.04037516191601753,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.018877983093262,
+ "step": 208
+ },
+ {
+ "epoch": 2.9075043630017454,
+ "grad_norm": 0.036492880433797836,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.009493827819824,
+ "step": 209
+ },
+ {
+ "epoch": 2.9214659685863875,
+ "grad_norm": 0.04785023257136345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.0140862464904785,
+ "step": 210
+ },
+ {
+ "epoch": 2.9354275741710296,
+ "grad_norm": 0.07721036672592163,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.030359268188477,
+ "step": 211
+ },
+ {
+ "epoch": 2.9493891797556717,
+ "grad_norm": 0.14220190048217773,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.045989990234375,
+ "step": 212
+ },
+ {
+ "epoch": 2.9633507853403143,
+ "grad_norm": 0.10265158116817474,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.043968200683594,
+ "step": 213
+ },
+ {
+ "epoch": 2.9773123909249564,
+ "grad_norm": 0.051398951560258865,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.022278785705566,
+ "step": 214
+ },
+ {
+ "epoch": 2.9912739965095985,
+ "grad_norm": 0.06044618785381317,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.017296314239502,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.03284750506281853,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 3.7544798851013184,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7531244158744812,
+ "eval_runtime": 60.3986,
+ "eval_samples_per_second": 40.431,
+ "eval_steps_per_second": 0.646,
+ "step": 216
+ },
+ {
+ "epoch": 3.013961605584642,
+ "grad_norm": 0.03895840048789978,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 5.9985857009887695,
+ "step": 217
+ },
+ {
+ "epoch": 3.0279232111692846,
+ "grad_norm": 0.04554992541670799,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 5.984710693359375,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418848167539267,
+ "grad_norm": 0.03359116613864899,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 5.984078407287598,
+ "step": 219
+ },
+ {
+ "epoch": 3.055846422338569,
+ "grad_norm": 0.035171881318092346,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 5.979424476623535,
+ "step": 220
+ },
+ {
+ "epoch": 3.0698080279232114,
+ "grad_norm": 0.032774072140455246,
+ "learning_rate": 0.000598278794687186,
+ "loss": 5.94467830657959,
+ "step": 221
+ },
+ {
+ "epoch": 3.0837696335078535,
+ "grad_norm": 0.032842908054590225,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 5.95986270904541,
+ "step": 222
+ },
+ {
+ "epoch": 3.0977312390924956,
+ "grad_norm": 0.02870488539338112,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 5.951186656951904,
+ "step": 223
+ },
+ {
+ "epoch": 3.1116928446771377,
+ "grad_norm": 0.030777374282479286,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 5.942493438720703,
+ "step": 224
+ },
+ {
+ "epoch": 3.1256544502617802,
+ "grad_norm": 0.026439018547534943,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 5.938275337219238,
+ "step": 225
+ },
+ {
+ "epoch": 3.1396160558464223,
+ "grad_norm": 0.028689472004771233,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 5.931982040405273,
+ "step": 226
+ },
+ {
+ "epoch": 3.1535776614310644,
+ "grad_norm": 0.028100362047553062,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 5.919692039489746,
+ "step": 227
+ },
+ {
+ "epoch": 3.167539267015707,
+ "grad_norm": 0.02407177910208702,
+ "learning_rate": 0.00059807235435179,
+ "loss": 5.909029006958008,
+ "step": 228
+ },
+ {
+ "epoch": 3.181500872600349,
+ "grad_norm": 0.0237748846411705,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 5.9131598472595215,
+ "step": 229
+ },
+ {
+ "epoch": 3.195462478184991,
+ "grad_norm": 0.022418556734919548,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 5.909396171569824,
+ "step": 230
+ },
+ {
+ "epoch": 3.2094240837696333,
+ "grad_norm": 0.024656571447849274,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 5.885243892669678,
+ "step": 231
+ },
+ {
+ "epoch": 3.223385689354276,
+ "grad_norm": 0.02136118710041046,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 5.898958206176758,
+ "step": 232
+ },
+ {
+ "epoch": 3.237347294938918,
+ "grad_norm": 0.025290116667747498,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 5.886986255645752,
+ "step": 233
+ },
+ {
+ "epoch": 3.25130890052356,
+ "grad_norm": 0.02656404674053192,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 5.897647857666016,
+ "step": 234
+ },
+ {
+ "epoch": 3.2652705061082026,
+ "grad_norm": 0.019188793376088142,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 5.875659942626953,
+ "step": 235
+ },
+ {
+ "epoch": 3.2792321116928447,
+ "grad_norm": 0.02308926172554493,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 5.859380722045898,
+ "step": 236
+ },
+ {
+ "epoch": 3.2931937172774868,
+ "grad_norm": 0.029775725677609444,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 5.873961448669434,
+ "step": 237
+ },
+ {
+ "epoch": 3.3071553228621293,
+ "grad_norm": 0.03875259310007095,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 5.863171100616455,
+ "step": 238
+ },
+ {
+ "epoch": 3.3211169284467714,
+ "grad_norm": 0.052446842193603516,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 5.886547088623047,
+ "step": 239
+ },
+ {
+ "epoch": 3.3350785340314135,
+ "grad_norm": 0.05583837628364563,
+ "learning_rate": 0.000597691359743243,
+ "loss": 5.8849287033081055,
+ "step": 240
+ },
+ {
+ "epoch": 3.349040139616056,
+ "grad_norm": 0.034729067236185074,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 5.863419532775879,
+ "step": 241
+ },
+ {
+ "epoch": 3.363001745200698,
+ "grad_norm": 0.03590013459324837,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 5.852169990539551,
+ "step": 242
+ },
+ {
+ "epoch": 3.3769633507853403,
+ "grad_norm": 0.035041507333517075,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 5.836690902709961,
+ "step": 243
+ },
+ {
+ "epoch": 3.3909249563699824,
+ "grad_norm": 0.02857939340174198,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 5.836485862731934,
+ "step": 244
+ },
+ {
+ "epoch": 3.404886561954625,
+ "grad_norm": 0.03176642581820488,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 5.860719203948975,
+ "step": 245
+ },
+ {
+ "epoch": 3.418848167539267,
+ "grad_norm": 0.026105748489499092,
+ "learning_rate": 0.000597488038559129,
+ "loss": 5.8354949951171875,
+ "step": 246
+ },
+ {
+ "epoch": 3.432809773123909,
+ "grad_norm": 0.029094981029629707,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 5.846247673034668,
+ "step": 247
+ },
+ {
+ "epoch": 3.4467713787085517,
+ "grad_norm": 0.02388785220682621,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 5.837607383728027,
+ "step": 248
+ },
+ {
+ "epoch": 3.4607329842931938,
+ "grad_norm": 0.03094472736120224,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 5.832090377807617,
+ "step": 249
+ },
+ {
+ "epoch": 3.474694589877836,
+ "grad_norm": 0.03227658197283745,
+ "learning_rate": 0.000597347745466109,
+ "loss": 5.83722448348999,
+ "step": 250
+ },
+ {
+ "epoch": 3.488656195462478,
+ "grad_norm": 0.042813923209905624,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 5.851722717285156,
+ "step": 251
+ },
+ {
+ "epoch": 3.5026178010471205,
+ "grad_norm": 0.05825525149703026,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.795719146728516,
+ "step": 252
+ },
+ {
+ "epoch": 3.5165794066317626,
+ "grad_norm": 0.0867447778582573,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.813750267028809,
+ "step": 253
+ },
+ {
+ "epoch": 3.530541012216405,
+ "grad_norm": 0.09192710369825363,
+ "learning_rate": 0.000597203657966319,
+ "loss": 5.823488235473633,
+ "step": 254
+ },
+ {
+ "epoch": 3.5445026178010473,
+ "grad_norm": 0.06069248169660568,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 5.814342021942139,
+ "step": 255
+ },
+ {
+ "epoch": 3.5584642233856894,
+ "grad_norm": 0.06259952485561371,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 5.834272384643555,
+ "step": 256
+ },
+ {
+ "epoch": 3.5724258289703315,
+ "grad_norm": 0.0513639897108078,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.807951927185059,
+ "step": 257
+ },
+ {
+ "epoch": 3.5863874345549736,
+ "grad_norm": 0.03327798470854759,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.790238857269287,
+ "step": 258
+ },
+ {
+ "epoch": 3.600349040139616,
+ "grad_norm": 0.03963128477334976,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.7956695556640625,
+ "step": 259
+ },
+ {
+ "epoch": 3.614310645724258,
+ "grad_norm": 0.031153453513979912,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 5.7874531745910645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6282722513089007,
+ "grad_norm": 0.031419433653354645,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 5.779650688171387,
+ "step": 261
+ },
+ {
+ "epoch": 3.642233856893543,
+ "grad_norm": 0.0288684144616127,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.776577949523926,
+ "step": 262
+ },
+ {
+ "epoch": 3.656195462478185,
+ "grad_norm": 0.028315918520092964,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.778571128845215,
+ "step": 263
+ },
+ {
+ "epoch": 3.670157068062827,
+ "grad_norm": 0.024977795779705048,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 5.776470184326172,
+ "step": 264
+ },
+ {
+ "epoch": 3.6841186736474696,
+ "grad_norm": 0.026822632178664207,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.755108833312988,
+ "step": 265
+ },
+ {
+ "epoch": 3.6980802792321117,
+ "grad_norm": 0.02738630585372448,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.74041748046875,
+ "step": 266
+ },
+ {
+ "epoch": 3.712041884816754,
+ "grad_norm": 0.02361280657351017,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.764027118682861,
+ "step": 267
+ },
+ {
+ "epoch": 3.7260034904013963,
+ "grad_norm": 0.029926009476184845,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.780390739440918,
+ "step": 268
+ },
+ {
+ "epoch": 3.7399650959860384,
+ "grad_norm": 0.03519507870078087,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.758105278015137,
+ "step": 269
+ },
+ {
+ "epoch": 3.7539267015706805,
+ "grad_norm": 0.030114633962512016,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.757230758666992,
+ "step": 270
+ },
+ {
+ "epoch": 3.7678883071553226,
+ "grad_norm": 0.02467748522758484,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.7615461349487305,
+ "step": 271
+ },
+ {
+ "epoch": 3.781849912739965,
+ "grad_norm": 0.03425183147192001,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.749448776245117,
+ "step": 272
+ },
+ {
+ "epoch": 3.7958115183246073,
+ "grad_norm": 0.046307582408189774,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.7356977462768555,
+ "step": 273
+ },
+ {
+ "epoch": 3.80977312390925,
+ "grad_norm": 0.054102856665849686,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.741869926452637,
+ "step": 274
+ },
+ {
+ "epoch": 3.823734729493892,
+ "grad_norm": 0.07026689499616623,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.721103668212891,
+ "step": 275
+ },
+ {
+ "epoch": 3.837696335078534,
+ "grad_norm": 0.0733039602637291,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.731414794921875,
+ "step": 276
+ },
+ {
+ "epoch": 3.851657940663176,
+ "grad_norm": 0.05342920497059822,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.734922885894775,
+ "step": 277
+ },
+ {
+ "epoch": 3.8656195462478182,
+ "grad_norm": 0.0504680834710598,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.732746124267578,
+ "step": 278
+ },
+ {
+ "epoch": 3.8795811518324608,
+ "grad_norm": 0.038991592824459076,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.719473838806152,
+ "step": 279
+ },
+ {
+ "epoch": 3.893542757417103,
+ "grad_norm": 0.04434404894709587,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.709481239318848,
+ "step": 280
+ },
+ {
+ "epoch": 3.9075043630017454,
+ "grad_norm": 0.047285184264183044,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.693531036376953,
+ "step": 281
+ },
+ {
+ "epoch": 3.9214659685863875,
+ "grad_norm": 0.045081138610839844,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.728959083557129,
+ "step": 282
+ },
+ {
+ "epoch": 3.9354275741710296,
+ "grad_norm": 0.025133663788437843,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.719696998596191,
+ "step": 283
+ },
+ {
+ "epoch": 3.9493891797556717,
+ "grad_norm": 0.03353596478700638,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.700235366821289,
+ "step": 284
+ },
+ {
+ "epoch": 3.9633507853403143,
+ "grad_norm": 0.035162411630153656,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.684836387634277,
+ "step": 285
+ },
+ {
+ "epoch": 3.9773123909249564,
+ "grad_norm": 0.04025450348854065,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.692097187042236,
+ "step": 286
+ },
+ {
+ "epoch": 3.9912739965095985,
+ "grad_norm": 0.04995397478342056,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.6766862869262695,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.031619708985090256,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 3.5519747734069824,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.714687705039978,
+ "eval_runtime": 59.637,
+ "eval_samples_per_second": 40.948,
+ "eval_steps_per_second": 0.654,
+ "step": 288
+ },
+ {
+ "epoch": 4.013961605584642,
+ "grad_norm": 0.036328598856925964,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.666294097900391,
+ "step": 289
+ },
+ {
+ "epoch": 4.027923211169284,
+ "grad_norm": 0.07326336950063705,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.6879167556762695,
+ "step": 290
+ },
+ {
+ "epoch": 4.041884816753926,
+ "grad_norm": 0.07947526127099991,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.674883842468262,
+ "step": 291
+ },
+ {
+ "epoch": 4.055846422338569,
+ "grad_norm": 0.04599885269999504,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.688086986541748,
+ "step": 292
+ },
+ {
+ "epoch": 4.069808027923211,
+ "grad_norm": 0.0465150885283947,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.649483680725098,
+ "step": 293
+ },
+ {
+ "epoch": 4.0837696335078535,
+ "grad_norm": 0.038542453199625015,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.653055667877197,
+ "step": 294
+ },
+ {
+ "epoch": 4.097731239092496,
+ "grad_norm": 0.03659164905548096,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.65955924987793,
+ "step": 295
+ },
+ {
+ "epoch": 4.111692844677138,
+ "grad_norm": 0.03824038803577423,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.650853157043457,
+ "step": 296
+ },
+ {
+ "epoch": 4.12565445026178,
+ "grad_norm": 0.03591204807162285,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.643499374389648,
+ "step": 297
+ },
+ {
+ "epoch": 4.139616055846423,
+ "grad_norm": 0.03269682824611664,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.657600402832031,
+ "step": 298
+ },
+ {
+ "epoch": 4.153577661431065,
+ "grad_norm": 0.03343537449836731,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.6352081298828125,
+ "step": 299
+ },
+ {
+ "epoch": 4.167539267015707,
+ "grad_norm": 0.034283801913261414,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.626757621765137,
+ "step": 300
+ },
+ {
+ "epoch": 4.181500872600349,
+ "grad_norm": 0.034614723175764084,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.6316680908203125,
+ "step": 301
+ },
+ {
+ "epoch": 4.195462478184991,
+ "grad_norm": 0.0340128131210804,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.648160934448242,
+ "step": 302
+ },
+ {
+ "epoch": 4.209424083769633,
+ "grad_norm": 0.03068280965089798,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.619738578796387,
+ "step": 303
+ },
+ {
+ "epoch": 4.223385689354275,
+ "grad_norm": 0.030269058421254158,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.609238624572754,
+ "step": 304
+ },
+ {
+ "epoch": 4.237347294938918,
+ "grad_norm": 0.03612440824508667,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.602194786071777,
+ "step": 305
+ },
+ {
+ "epoch": 4.2513089005235605,
+ "grad_norm": 0.044292811304330826,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.61159610748291,
+ "step": 306
+ },
+ {
+ "epoch": 4.265270506108203,
+ "grad_norm": 0.04449291527271271,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.6274237632751465,
+ "step": 307
+ },
+ {
+ "epoch": 4.279232111692845,
+ "grad_norm": 0.05014493316411972,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.589034080505371,
+ "step": 308
+ },
+ {
+ "epoch": 4.293193717277487,
+ "grad_norm": 0.05286206305027008,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.600770950317383,
+ "step": 309
+ },
+ {
+ "epoch": 4.307155322862129,
+ "grad_norm": 0.048459846526384354,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.5796098709106445,
+ "step": 310
+ },
+ {
+ "epoch": 4.321116928446771,
+ "grad_norm": 0.05329663306474686,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.590407371520996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335078534031414,
+ "grad_norm": 0.052595119923353195,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.583693504333496,
+ "step": 312
+ },
+ {
+ "epoch": 4.349040139616056,
+ "grad_norm": 0.05164601653814316,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.603281021118164,
+ "step": 313
+ },
+ {
+ "epoch": 4.363001745200698,
+ "grad_norm": 0.04657137766480446,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.591344833374023,
+ "step": 314
+ },
+ {
+ "epoch": 4.37696335078534,
+ "grad_norm": 0.03804599866271019,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.580578804016113,
+ "step": 315
+ },
+ {
+ "epoch": 4.390924956369982,
+ "grad_norm": 0.03835320845246315,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.597836971282959,
+ "step": 316
+ },
+ {
+ "epoch": 4.4048865619546245,
+ "grad_norm": 0.034059133380651474,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.567502021789551,
+ "step": 317
+ },
+ {
+ "epoch": 4.418848167539267,
+ "grad_norm": 0.03350386768579483,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.552022933959961,
+ "step": 318
+ },
+ {
+ "epoch": 4.43280977312391,
+ "grad_norm": 0.032574303448200226,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.559177875518799,
+ "step": 319
+ },
+ {
+ "epoch": 4.446771378708552,
+ "grad_norm": 0.030893854796886444,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.549288272857666,
+ "step": 320
+ },
+ {
+ "epoch": 4.460732984293194,
+ "grad_norm": 0.03255036100745201,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.5400390625,
+ "step": 321
+ },
+ {
+ "epoch": 4.474694589877836,
+ "grad_norm": 0.03009258769452572,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.5328474044799805,
+ "step": 322
+ },
+ {
+ "epoch": 4.488656195462478,
+ "grad_norm": 0.03178320452570915,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.5594482421875,
+ "step": 323
+ },
+ {
+ "epoch": 4.50261780104712,
+ "grad_norm": 0.03360465541481972,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.5402727127075195,
+ "step": 324
+ },
+ {
+ "epoch": 4.516579406631763,
+ "grad_norm": 0.04275049269199371,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.534474849700928,
+ "step": 325
+ },
+ {
+ "epoch": 4.530541012216405,
+ "grad_norm": 0.04140641540288925,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.5324859619140625,
+ "step": 326
+ },
+ {
+ "epoch": 4.544502617801047,
+ "grad_norm": 0.04769204929471016,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.4919352531433105,
+ "step": 327
+ },
+ {
+ "epoch": 4.558464223385689,
+ "grad_norm": 0.04402285814285278,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.520858287811279,
+ "step": 328
+ },
+ {
+ "epoch": 4.5724258289703315,
+ "grad_norm": 0.038526229560375214,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.523778438568115,
+ "step": 329
+ },
+ {
+ "epoch": 4.5863874345549736,
+ "grad_norm": 0.05422377213835716,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.517973899841309,
+ "step": 330
+ },
+ {
+ "epoch": 4.600349040139616,
+ "grad_norm": 0.07236721366643906,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.548083782196045,
+ "step": 331
+ },
+ {
+ "epoch": 4.614310645724259,
+ "grad_norm": 0.06861061602830887,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.535577774047852,
+ "step": 332
+ },
+ {
+ "epoch": 4.628272251308901,
+ "grad_norm": 0.07148340344429016,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.54155158996582,
+ "step": 333
+ },
+ {
+ "epoch": 4.642233856893543,
+ "grad_norm": 0.07838302850723267,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.552852630615234,
+ "step": 334
+ },
+ {
+ "epoch": 4.656195462478185,
+ "grad_norm": 0.10946758091449738,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.555066108703613,
+ "step": 335
+ },
+ {
+ "epoch": 4.670157068062827,
+ "grad_norm": 0.10143530368804932,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.584932804107666,
+ "step": 336
+ },
+ {
+ "epoch": 4.684118673647469,
+ "grad_norm": 0.08926606178283691,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.570462226867676,
+ "step": 337
+ },
+ {
+ "epoch": 4.698080279232112,
+ "grad_norm": 0.07060826569795609,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.5716753005981445,
+ "step": 338
+ },
+ {
+ "epoch": 4.712041884816754,
+ "grad_norm": 0.0418456569314003,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.556496620178223,
+ "step": 339
+ },
+ {
+ "epoch": 4.726003490401396,
+ "grad_norm": 0.05952411890029907,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.567801475524902,
+ "step": 340
+ },
+ {
+ "epoch": 4.739965095986038,
+ "grad_norm": 0.042999885976314545,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.537403106689453,
+ "step": 341
+ },
+ {
+ "epoch": 4.7539267015706805,
+ "grad_norm": 0.04426449537277222,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.552387237548828,
+ "step": 342
+ },
+ {
+ "epoch": 4.767888307155323,
+ "grad_norm": 0.040835145860910416,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.53051233291626,
+ "step": 343
+ },
+ {
+ "epoch": 4.781849912739965,
+ "grad_norm": 0.03819291293621063,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.531174659729004,
+ "step": 344
+ },
+ {
+ "epoch": 4.795811518324607,
+ "grad_norm": 0.030088599771261215,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.499328136444092,
+ "step": 345
+ },
+ {
+ "epoch": 4.80977312390925,
+ "grad_norm": 0.029984362423419952,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.504540920257568,
+ "step": 346
+ },
+ {
+ "epoch": 4.823734729493892,
+ "grad_norm": 0.027402915060520172,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.50899076461792,
+ "step": 347
+ },
+ {
+ "epoch": 4.837696335078534,
+ "grad_norm": 0.025096213445067406,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.480479717254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.851657940663176,
+ "grad_norm": 0.026828132569789886,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.501729965209961,
+ "step": 349
+ },
+ {
+ "epoch": 4.865619546247818,
+ "grad_norm": 0.023715168237686157,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.466860771179199,
+ "step": 350
+ },
+ {
+ "epoch": 4.879581151832461,
+ "grad_norm": 0.024503640830516815,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.491827964782715,
+ "step": 351
+ },
+ {
+ "epoch": 4.893542757417103,
+ "grad_norm": 0.020891105756163597,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.478384971618652,
+ "step": 352
+ },
+ {
+ "epoch": 4.907504363001745,
+ "grad_norm": 0.021373765543103218,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.4612135887146,
+ "step": 353
+ },
+ {
+ "epoch": 4.9214659685863875,
+ "grad_norm": 0.021773340180516243,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.4547319412231445,
+ "step": 354
+ },
+ {
+ "epoch": 4.93542757417103,
+ "grad_norm": 0.019637586548924446,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.431741714477539,
+ "step": 355
+ },
+ {
+ "epoch": 4.949389179755672,
+ "grad_norm": 0.023108696565032005,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.448612213134766,
+ "step": 356
+ },
+ {
+ "epoch": 4.963350785340314,
+ "grad_norm": 0.021935103461146355,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.447112083435059,
+ "step": 357
+ },
+ {
+ "epoch": 4.977312390924956,
+ "grad_norm": 0.023260753601789474,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.436091423034668,
+ "step": 358
+ },
+ {
+ "epoch": 4.991273996509599,
+ "grad_norm": 0.021355057135224342,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.408421993255615,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.014992325566709042,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 3.4004223346710205,
+ "step": 360
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.5361471231819776e+17,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-360/training_args.bin b/runs/l2r90-wd0033-ssl/checkpoint-360/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..313c41733ff55284c06034a59bb885b30e40ffc8
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-360/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7ddd0bdddba68d3e19ab33477000f3fc684fe4ee8e3580eeb8c02646e5fc27ac
+size 4856
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3600/chat_template.jinja b/runs/l2r90-wd0033-ssl/checkpoint-3600/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3600/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3600/config.json b/runs/l2r90-wd0033-ssl/checkpoint-3600/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3600/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3600/generation_config.json b/runs/l2r90-wd0033-ssl/checkpoint-3600/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3600/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3600/model.safetensors b/runs/l2r90-wd0033-ssl/checkpoint-3600/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..8a0f2c3a5da311f46effbca09b95f8b4c8705f93
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3600/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7c0e423ce1c06a48b1b71fddb7518c9f801cc9cd0efd5df48f405c80934461f0
+size 583356232
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3600/optimizer.pt b/runs/l2r90-wd0033-ssl/checkpoint-3600/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..4c3317e9d103c79909482f08e7c25d04155372cf
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3600/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:fae375040526e06fb19c0dc49a1eae654983806664c5dedb8043d8ddfddee77a
+size 1166827898
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3600/rng_state_0.pth b/runs/l2r90-wd0033-ssl/checkpoint-3600/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..eb61f65f3a3f7e24b260771238c24cc304e14e74
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3600/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:21d0b427100597feb74d71fe83df6db25d601564c6f3a36049ad6073b6430e1c
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3600/rng_state_1.pth b/runs/l2r90-wd0033-ssl/checkpoint-3600/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..81cece85458832c36516df1a5dad973faefe38bc
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3600/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f958752c475d37e1c987c806e29e7d8b7c077cbeda2b613ed7a652cd7c1e5e17
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3600/scheduler.pt b/runs/l2r90-wd0033-ssl/checkpoint-3600/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..dff04aa3b724ae7dec72031a9c3d0ecf0b695f1b
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3600/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:187927afe10d5874e052562039160b9509612fe972489c80a97620ccae701300
+size 1064
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3600/tokenizer.json b/runs/l2r90-wd0033-ssl/checkpoint-3600/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3600/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3600/tokenizer_config.json b/runs/l2r90-wd0033-ssl/checkpoint-3600/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3600/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3600/trainer_state.json b/runs/l2r90-wd0033-ssl/checkpoint-3600/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..280dc74f69d765d077db8e7d462e48ad2ac4b31c
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3600/trainer_state.json
@@ -0,0 +1,25610 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 50.0,
+ "eval_steps": 500,
+ "global_step": 3600,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013961605584642234,
+ "grad_norm": 0.26733073592185974,
+ "learning_rate": 0.0,
+ "loss": 7.990724086761475,
+ "step": 1
+ },
+ {
+ "epoch": 0.027923211169284468,
+ "grad_norm": 0.26712340116500854,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 7.990110874176025,
+ "step": 2
+ },
+ {
+ "epoch": 0.041884816753926704,
+ "grad_norm": 0.2646932601928711,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 7.956935405731201,
+ "step": 3
+ },
+ {
+ "epoch": 0.055846422338568937,
+ "grad_norm": 0.24684545397758484,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 7.890539646148682,
+ "step": 4
+ },
+ {
+ "epoch": 0.06980802792321117,
+ "grad_norm": 0.21016643941402435,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 7.8146138191223145,
+ "step": 5
+ },
+ {
+ "epoch": 0.08376963350785341,
+ "grad_norm": 0.17992742359638214,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 7.724307060241699,
+ "step": 6
+ },
+ {
+ "epoch": 0.09773123909249563,
+ "grad_norm": 0.14356893301010132,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 7.645662307739258,
+ "step": 7
+ },
+ {
+ "epoch": 0.11169284467713787,
+ "grad_norm": 0.10830964893102646,
+ "learning_rate": 4.2e-05,
+ "loss": 7.582019329071045,
+ "step": 8
+ },
+ {
+ "epoch": 0.1256544502617801,
+ "grad_norm": 0.08265161514282227,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 7.5230793952941895,
+ "step": 9
+ },
+ {
+ "epoch": 0.13961605584642234,
+ "grad_norm": 0.06423710286617279,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 7.473812103271484,
+ "step": 10
+ },
+ {
+ "epoch": 0.15357766143106458,
+ "grad_norm": 0.05132246017456055,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 7.440786838531494,
+ "step": 11
+ },
+ {
+ "epoch": 0.16753926701570682,
+ "grad_norm": 0.040572404861450195,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 7.41290283203125,
+ "step": 12
+ },
+ {
+ "epoch": 0.18150087260034903,
+ "grad_norm": 0.03138705715537071,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 7.39749002456665,
+ "step": 13
+ },
+ {
+ "epoch": 0.19546247818499127,
+ "grad_norm": 0.02553008496761322,
+ "learning_rate": 7.8e-05,
+ "loss": 7.387092590332031,
+ "step": 14
+ },
+ {
+ "epoch": 0.2094240837696335,
+ "grad_norm": 0.022178977727890015,
+ "learning_rate": 8.4e-05,
+ "loss": 7.378993511199951,
+ "step": 15
+ },
+ {
+ "epoch": 0.22338568935427575,
+ "grad_norm": 0.019915293902158737,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 7.37220573425293,
+ "step": 16
+ },
+ {
+ "epoch": 0.23734729493891799,
+ "grad_norm": 0.017390629276633263,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 7.366759300231934,
+ "step": 17
+ },
+ {
+ "epoch": 0.2513089005235602,
+ "grad_norm": 0.015472771599888802,
+ "learning_rate": 0.000102,
+ "loss": 7.3652663230896,
+ "step": 18
+ },
+ {
+ "epoch": 0.26527050610820246,
+ "grad_norm": 0.013599214144051075,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 7.359714984893799,
+ "step": 19
+ },
+ {
+ "epoch": 0.2792321116928447,
+ "grad_norm": 0.012442861683666706,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 7.353302001953125,
+ "step": 20
+ },
+ {
+ "epoch": 0.2931937172774869,
+ "grad_norm": 0.010471965186297894,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 7.357417583465576,
+ "step": 21
+ },
+ {
+ "epoch": 0.30715532286212915,
+ "grad_norm": 0.009647224098443985,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 7.351763725280762,
+ "step": 22
+ },
+ {
+ "epoch": 0.32111692844677137,
+ "grad_norm": 0.008228210732340813,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 7.345434188842773,
+ "step": 23
+ },
+ {
+ "epoch": 0.33507853403141363,
+ "grad_norm": 0.007454514969140291,
+ "learning_rate": 0.000138,
+ "loss": 7.342520236968994,
+ "step": 24
+ },
+ {
+ "epoch": 0.34904013961605584,
+ "grad_norm": 0.0069856527261435986,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 7.338870048522949,
+ "step": 25
+ },
+ {
+ "epoch": 0.36300174520069806,
+ "grad_norm": 0.006091665010899305,
+ "learning_rate": 0.00015,
+ "loss": 7.328121662139893,
+ "step": 26
+ },
+ {
+ "epoch": 0.3769633507853403,
+ "grad_norm": 0.005525320768356323,
+ "learning_rate": 0.000156,
+ "loss": 7.335029602050781,
+ "step": 27
+ },
+ {
+ "epoch": 0.39092495636998253,
+ "grad_norm": 0.0061133443377912045,
+ "learning_rate": 0.000162,
+ "loss": 7.319417953491211,
+ "step": 28
+ },
+ {
+ "epoch": 0.4048865619546248,
+ "grad_norm": 0.006040909793227911,
+ "learning_rate": 0.000168,
+ "loss": 7.326376438140869,
+ "step": 29
+ },
+ {
+ "epoch": 0.418848167539267,
+ "grad_norm": 0.006003216840326786,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 7.319552421569824,
+ "step": 30
+ },
+ {
+ "epoch": 0.4328097731239092,
+ "grad_norm": 0.007096003741025925,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 7.323486328125,
+ "step": 31
+ },
+ {
+ "epoch": 0.4467713787085515,
+ "grad_norm": 0.007725459989160299,
+ "learning_rate": 0.000186,
+ "loss": 7.3190717697143555,
+ "step": 32
+ },
+ {
+ "epoch": 0.4607329842931937,
+ "grad_norm": 0.007338670082390308,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 7.317154884338379,
+ "step": 33
+ },
+ {
+ "epoch": 0.47469458987783597,
+ "grad_norm": 0.007709544617682695,
+ "learning_rate": 0.000198,
+ "loss": 7.311826705932617,
+ "step": 34
+ },
+ {
+ "epoch": 0.4886561954624782,
+ "grad_norm": 0.009249784983694553,
+ "learning_rate": 0.000204,
+ "loss": 7.322498798370361,
+ "step": 35
+ },
+ {
+ "epoch": 0.5026178010471204,
+ "grad_norm": 0.006950767710804939,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 7.314571380615234,
+ "step": 36
+ },
+ {
+ "epoch": 0.5165794066317626,
+ "grad_norm": 0.007583985570818186,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 7.309540748596191,
+ "step": 37
+ },
+ {
+ "epoch": 0.5305410122164049,
+ "grad_norm": 0.014731820672750473,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 7.307004928588867,
+ "step": 38
+ },
+ {
+ "epoch": 0.5445026178010471,
+ "grad_norm": 0.016276473179459572,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 7.299642562866211,
+ "step": 39
+ },
+ {
+ "epoch": 0.5584642233856894,
+ "grad_norm": 0.020018689334392548,
+ "learning_rate": 0.000234,
+ "loss": 7.298466682434082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5724258289703316,
+ "grad_norm": 0.014941536821424961,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 7.289968967437744,
+ "step": 41
+ },
+ {
+ "epoch": 0.5863874345549738,
+ "grad_norm": 0.018299812451004982,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 7.276564598083496,
+ "step": 42
+ },
+ {
+ "epoch": 0.6003490401396161,
+ "grad_norm": 0.01615075021982193,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.2575554847717285,
+ "step": 43
+ },
+ {
+ "epoch": 0.6143106457242583,
+ "grad_norm": 0.013063831254839897,
+ "learning_rate": 0.000258,
+ "loss": 7.253885269165039,
+ "step": 44
+ },
+ {
+ "epoch": 0.6282722513089005,
+ "grad_norm": 0.015211676247417927,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.245779037475586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6422338568935427,
+ "grad_norm": 0.02093500830233097,
+ "learning_rate": 0.00027,
+ "loss": 7.225130081176758,
+ "step": 46
+ },
+ {
+ "epoch": 0.6561954624781849,
+ "grad_norm": 0.01808053068816662,
+ "learning_rate": 0.000276,
+ "loss": 7.209654331207275,
+ "step": 47
+ },
+ {
+ "epoch": 0.6701570680628273,
+ "grad_norm": 0.02209644205868244,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.197124481201172,
+ "step": 48
+ },
+ {
+ "epoch": 0.6841186736474695,
+ "grad_norm": 0.023524878546595573,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.1841936111450195,
+ "step": 49
+ },
+ {
+ "epoch": 0.6980802792321117,
+ "grad_norm": 0.031148087233304977,
+ "learning_rate": 0.000294,
+ "loss": 7.188040733337402,
+ "step": 50
+ },
+ {
+ "epoch": 0.7120418848167539,
+ "grad_norm": 0.02931929938495159,
+ "learning_rate": 0.0003,
+ "loss": 7.17371940612793,
+ "step": 51
+ },
+ {
+ "epoch": 0.7260034904013961,
+ "grad_norm": 0.01611483097076416,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.1714324951171875,
+ "step": 52
+ },
+ {
+ "epoch": 0.7399650959860384,
+ "grad_norm": 0.02619907818734646,
+ "learning_rate": 0.000312,
+ "loss": 7.158261299133301,
+ "step": 53
+ },
+ {
+ "epoch": 0.7539267015706806,
+ "grad_norm": 0.049462560564279556,
+ "learning_rate": 0.000318,
+ "loss": 7.167076110839844,
+ "step": 54
+ },
+ {
+ "epoch": 0.7678883071553229,
+ "grad_norm": 0.09835070371627808,
+ "learning_rate": 0.000324,
+ "loss": 7.154292106628418,
+ "step": 55
+ },
+ {
+ "epoch": 0.7818499127399651,
+ "grad_norm": 0.11406510323286057,
+ "learning_rate": 0.00033,
+ "loss": 7.162016868591309,
+ "step": 56
+ },
+ {
+ "epoch": 0.7958115183246073,
+ "grad_norm": 0.027026303112506866,
+ "learning_rate": 0.000336,
+ "loss": 7.141416549682617,
+ "step": 57
+ },
+ {
+ "epoch": 0.8097731239092496,
+ "grad_norm": 0.06690579652786255,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.13959264755249,
+ "step": 58
+ },
+ {
+ "epoch": 0.8237347294938918,
+ "grad_norm": 0.020919431000947952,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.126513481140137,
+ "step": 59
+ },
+ {
+ "epoch": 0.837696335078534,
+ "grad_norm": 0.05618130415678024,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.124345779418945,
+ "step": 60
+ },
+ {
+ "epoch": 0.8516579406631762,
+ "grad_norm": 0.03402302414178848,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.100615501403809,
+ "step": 61
+ },
+ {
+ "epoch": 0.8656195462478184,
+ "grad_norm": 0.03646906837821007,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.116507530212402,
+ "step": 62
+ },
+ {
+ "epoch": 0.8795811518324608,
+ "grad_norm": 0.03076397068798542,
+ "learning_rate": 0.000372,
+ "loss": 7.097659111022949,
+ "step": 63
+ },
+ {
+ "epoch": 0.893542757417103,
+ "grad_norm": 0.03329646959900856,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.097290992736816,
+ "step": 64
+ },
+ {
+ "epoch": 0.9075043630017452,
+ "grad_norm": 0.01998630352318287,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.092268943786621,
+ "step": 65
+ },
+ {
+ "epoch": 0.9214659685863874,
+ "grad_norm": 0.03278940171003342,
+ "learning_rate": 0.00039,
+ "loss": 7.088292121887207,
+ "step": 66
+ },
+ {
+ "epoch": 0.9354275741710296,
+ "grad_norm": 0.024620421230793,
+ "learning_rate": 0.000396,
+ "loss": 7.078650951385498,
+ "step": 67
+ },
+ {
+ "epoch": 0.9493891797556719,
+ "grad_norm": 0.03750383481383324,
+ "learning_rate": 0.000402,
+ "loss": 7.052778244018555,
+ "step": 68
+ },
+ {
+ "epoch": 0.9633507853403142,
+ "grad_norm": 0.07558075338602066,
+ "learning_rate": 0.000408,
+ "loss": 7.045991897583008,
+ "step": 69
+ },
+ {
+ "epoch": 0.9773123909249564,
+ "grad_norm": 0.16021482646465302,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.023152828216553,
+ "step": 70
+ },
+ {
+ "epoch": 0.9912739965095986,
+ "grad_norm": 0.24981893599033356,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.067540645599365,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.10255520790815353,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 4.425640106201172,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8861758708953857,
+ "eval_runtime": 58.2699,
+ "eval_samples_per_second": 41.908,
+ "eval_steps_per_second": 0.669,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139616055846423,
+ "grad_norm": 0.13374797999858856,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.073373317718506,
+ "step": 73
+ },
+ {
+ "epoch": 1.0279232111692844,
+ "grad_norm": 0.05079387128353119,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.0243306159973145,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418848167539267,
+ "grad_norm": 0.10056068748235703,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.014476299285889,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558464223385688,
+ "grad_norm": 0.06995758414268494,
+ "learning_rate": 0.00045,
+ "loss": 7.011931419372559,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698080279232112,
+ "grad_norm": 0.0483892522752285,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 6.9822797775268555,
+ "step": 77
+ },
+ {
+ "epoch": 1.0837696335078535,
+ "grad_norm": 0.08408170193433762,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 6.964348793029785,
+ "step": 78
+ },
+ {
+ "epoch": 1.0977312390924956,
+ "grad_norm": 0.08027710020542145,
+ "learning_rate": 0.000468,
+ "loss": 6.96852445602417,
+ "step": 79
+ },
+ {
+ "epoch": 1.111692844677138,
+ "grad_norm": 0.05647159367799759,
+ "learning_rate": 0.000474,
+ "loss": 6.944484710693359,
+ "step": 80
+ },
+ {
+ "epoch": 1.12565445026178,
+ "grad_norm": 0.07464330643415451,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 6.9246745109558105,
+ "step": 81
+ },
+ {
+ "epoch": 1.1396160558464223,
+ "grad_norm": 0.04451674595475197,
+ "learning_rate": 0.000486,
+ "loss": 6.895403861999512,
+ "step": 82
+ },
+ {
+ "epoch": 1.1535776614310647,
+ "grad_norm": 0.05820256471633911,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 6.887338638305664,
+ "step": 83
+ },
+ {
+ "epoch": 1.1675392670157068,
+ "grad_norm": 0.02925296127796173,
+ "learning_rate": 0.000498,
+ "loss": 6.857797622680664,
+ "step": 84
+ },
+ {
+ "epoch": 1.181500872600349,
+ "grad_norm": 0.05096987634897232,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 6.854822158813477,
+ "step": 85
+ },
+ {
+ "epoch": 1.1954624781849912,
+ "grad_norm": 0.03483197093009949,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 6.840914726257324,
+ "step": 86
+ },
+ {
+ "epoch": 1.2094240837696335,
+ "grad_norm": 0.03323264792561531,
+ "learning_rate": 0.000516,
+ "loss": 6.834497451782227,
+ "step": 87
+ },
+ {
+ "epoch": 1.2233856893542758,
+ "grad_norm": 0.03290868178009987,
+ "learning_rate": 0.000522,
+ "loss": 6.820746898651123,
+ "step": 88
+ },
+ {
+ "epoch": 1.237347294938918,
+ "grad_norm": 0.025398844853043556,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 6.795746803283691,
+ "step": 89
+ },
+ {
+ "epoch": 1.2513089005235603,
+ "grad_norm": 0.030700596049427986,
+ "learning_rate": 0.000534,
+ "loss": 6.7867889404296875,
+ "step": 90
+ },
+ {
+ "epoch": 1.2652705061082026,
+ "grad_norm": 0.019589420408010483,
+ "learning_rate": 0.00054,
+ "loss": 6.77443790435791,
+ "step": 91
+ },
+ {
+ "epoch": 1.2792321116928447,
+ "grad_norm": 0.03313920646905899,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 6.762271881103516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2931937172774868,
+ "grad_norm": 0.025389468297362328,
+ "learning_rate": 0.000552,
+ "loss": 6.750258445739746,
+ "step": 93
+ },
+ {
+ "epoch": 1.307155322862129,
+ "grad_norm": 0.040296655148267746,
+ "learning_rate": 0.000558,
+ "loss": 6.729808807373047,
+ "step": 94
+ },
+ {
+ "epoch": 1.3211169284467714,
+ "grad_norm": 0.04735933244228363,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 6.737982749938965,
+ "step": 95
+ },
+ {
+ "epoch": 1.3350785340314135,
+ "grad_norm": 0.05529680848121643,
+ "learning_rate": 0.00057,
+ "loss": 6.735714912414551,
+ "step": 96
+ },
+ {
+ "epoch": 1.3490401396160558,
+ "grad_norm": 0.07148554176092148,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 6.708859920501709,
+ "step": 97
+ },
+ {
+ "epoch": 1.3630017452006982,
+ "grad_norm": 0.059708479791879654,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 6.708860397338867,
+ "step": 98
+ },
+ {
+ "epoch": 1.3769633507853403,
+ "grad_norm": 0.027111195027828217,
+ "learning_rate": 0.000588,
+ "loss": 6.686685085296631,
+ "step": 99
+ },
+ {
+ "epoch": 1.3909249563699826,
+ "grad_norm": 0.0500967763364315,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 6.676325798034668,
+ "step": 100
+ },
+ {
+ "epoch": 1.404886561954625,
+ "grad_norm": 0.044565796852111816,
+ "learning_rate": 0.0006,
+ "loss": 6.674530029296875,
+ "step": 101
+ },
+ {
+ "epoch": 1.418848167539267,
+ "grad_norm": 0.03608180209994316,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 6.653741359710693,
+ "step": 102
+ },
+ {
+ "epoch": 1.4328097731239091,
+ "grad_norm": 0.030394982546567917,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 6.643152236938477,
+ "step": 103
+ },
+ {
+ "epoch": 1.4467713787085514,
+ "grad_norm": 0.040185656398534775,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 6.628686904907227,
+ "step": 104
+ },
+ {
+ "epoch": 1.4607329842931938,
+ "grad_norm": 0.021049322560429573,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 6.610864639282227,
+ "step": 105
+ },
+ {
+ "epoch": 1.4746945898778359,
+ "grad_norm": 0.04164277762174606,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 6.602880954742432,
+ "step": 106
+ },
+ {
+ "epoch": 1.4886561954624782,
+ "grad_norm": 0.0389864556491375,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 6.599456787109375,
+ "step": 107
+ },
+ {
+ "epoch": 1.5026178010471205,
+ "grad_norm": 0.05805973336100578,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 6.588700294494629,
+ "step": 108
+ },
+ {
+ "epoch": 1.5165794066317626,
+ "grad_norm": 0.08825036883354187,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 6.587496757507324,
+ "step": 109
+ },
+ {
+ "epoch": 1.530541012216405,
+ "grad_norm": 0.12521523237228394,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 6.626501083374023,
+ "step": 110
+ },
+ {
+ "epoch": 1.5445026178010473,
+ "grad_norm": 0.07429645210504532,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 6.590241432189941,
+ "step": 111
+ },
+ {
+ "epoch": 1.5584642233856894,
+ "grad_norm": 0.06463009119033813,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 6.578714370727539,
+ "step": 112
+ },
+ {
+ "epoch": 1.5724258289703315,
+ "grad_norm": 0.07508203387260437,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 6.561270713806152,
+ "step": 113
+ },
+ {
+ "epoch": 1.5863874345549738,
+ "grad_norm": 0.04125197231769562,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 6.558821678161621,
+ "step": 114
+ },
+ {
+ "epoch": 1.600349040139616,
+ "grad_norm": 0.05457386001944542,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.541611671447754,
+ "step": 115
+ },
+ {
+ "epoch": 1.6143106457242582,
+ "grad_norm": 0.03684385493397713,
+ "learning_rate": 0.000599973080557108,
+ "loss": 6.547738075256348,
+ "step": 116
+ },
+ {
+ "epoch": 1.6282722513089005,
+ "grad_norm": 0.05888746306300163,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.529491424560547,
+ "step": 117
+ },
+ {
+ "epoch": 1.6422338568935428,
+ "grad_norm": 0.056668370962142944,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.519444942474365,
+ "step": 118
+ },
+ {
+ "epoch": 1.656195462478185,
+ "grad_norm": 0.09123896062374115,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.522984981536865,
+ "step": 119
+ },
+ {
+ "epoch": 1.6701570680628273,
+ "grad_norm": 0.12380023300647736,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.515375137329102,
+ "step": 120
+ },
+ {
+ "epoch": 1.6841186736474696,
+ "grad_norm": 0.06953322887420654,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.500491142272949,
+ "step": 121
+ },
+ {
+ "epoch": 1.6980802792321117,
+ "grad_norm": 0.08613869547843933,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.5097198486328125,
+ "step": 122
+ },
+ {
+ "epoch": 1.7120418848167538,
+ "grad_norm": 0.08296383172273636,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.487521171569824,
+ "step": 123
+ },
+ {
+ "epoch": 1.7260034904013961,
+ "grad_norm": 0.05744783207774162,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.485134124755859,
+ "step": 124
+ },
+ {
+ "epoch": 1.7399650959860384,
+ "grad_norm": 0.04108566418290138,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.466182708740234,
+ "step": 125
+ },
+ {
+ "epoch": 1.7539267015706805,
+ "grad_norm": 0.05374931916594505,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.479011535644531,
+ "step": 126
+ },
+ {
+ "epoch": 1.7678883071553229,
+ "grad_norm": 0.03867914527654648,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.465902328491211,
+ "step": 127
+ },
+ {
+ "epoch": 1.7818499127399652,
+ "grad_norm": 0.05477374792098999,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.453082084655762,
+ "step": 128
+ },
+ {
+ "epoch": 1.7958115183246073,
+ "grad_norm": 0.06508298963308334,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.446089744567871,
+ "step": 129
+ },
+ {
+ "epoch": 1.8097731239092496,
+ "grad_norm": 0.06388624757528305,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.451748847961426,
+ "step": 130
+ },
+ {
+ "epoch": 1.823734729493892,
+ "grad_norm": 0.04223188757896423,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.4492387771606445,
+ "step": 131
+ },
+ {
+ "epoch": 1.837696335078534,
+ "grad_norm": 0.04528658092021942,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.409390926361084,
+ "step": 132
+ },
+ {
+ "epoch": 1.8516579406631761,
+ "grad_norm": 0.04962967336177826,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.415774345397949,
+ "step": 133
+ },
+ {
+ "epoch": 1.8656195462478184,
+ "grad_norm": 0.031944390386343,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.411528587341309,
+ "step": 134
+ },
+ {
+ "epoch": 1.8795811518324608,
+ "grad_norm": 0.028411149978637695,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.392329216003418,
+ "step": 135
+ },
+ {
+ "epoch": 1.8935427574171029,
+ "grad_norm": 0.03516148403286934,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.392238616943359,
+ "step": 136
+ },
+ {
+ "epoch": 1.9075043630017452,
+ "grad_norm": 0.027800120413303375,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.39179801940918,
+ "step": 137
+ },
+ {
+ "epoch": 1.9214659685863875,
+ "grad_norm": 0.02362987957894802,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.378960609436035,
+ "step": 138
+ },
+ {
+ "epoch": 1.9354275741710296,
+ "grad_norm": 0.033999308943748474,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.36802339553833,
+ "step": 139
+ },
+ {
+ "epoch": 1.949389179755672,
+ "grad_norm": 0.03290879726409912,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.366308689117432,
+ "step": 140
+ },
+ {
+ "epoch": 1.9633507853403143,
+ "grad_norm": 0.0286843404173851,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.350399971008301,
+ "step": 141
+ },
+ {
+ "epoch": 1.9773123909249564,
+ "grad_norm": 0.027145877480506897,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.33477783203125,
+ "step": 142
+ },
+ {
+ "epoch": 1.9912739965095985,
+ "grad_norm": 0.03530878946185112,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.322859764099121,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.028133150190114975,
+ "learning_rate": 0.000599778806120077,
+ "loss": 3.955288887023926,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7927101254463196,
+ "eval_runtime": 60.8372,
+ "eval_samples_per_second": 40.14,
+ "eval_steps_per_second": 0.641,
+ "step": 144
+ },
+ {
+ "epoch": 2.013961605584642,
+ "grad_norm": 0.03054557554423809,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.311975479125977,
+ "step": 145
+ },
+ {
+ "epoch": 2.0279232111692846,
+ "grad_norm": 0.05406459793448448,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.3030853271484375,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418848167539267,
+ "grad_norm": 0.10911019891500473,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.329056739807129,
+ "step": 147
+ },
+ {
+ "epoch": 2.055846422338569,
+ "grad_norm": 0.13948315382003784,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.362421035766602,
+ "step": 148
+ },
+ {
+ "epoch": 2.0698080279232114,
+ "grad_norm": 0.13176991045475006,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.3888139724731445,
+ "step": 149
+ },
+ {
+ "epoch": 2.0837696335078535,
+ "grad_norm": 0.09615885466337204,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.35154914855957,
+ "step": 150
+ },
+ {
+ "epoch": 2.0977312390924956,
+ "grad_norm": 0.06462561339139938,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.335790634155273,
+ "step": 151
+ },
+ {
+ "epoch": 2.1116928446771377,
+ "grad_norm": 0.05186621472239494,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.324396133422852,
+ "step": 152
+ },
+ {
+ "epoch": 2.1256544502617802,
+ "grad_norm": 0.06095981225371361,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.318549156188965,
+ "step": 153
+ },
+ {
+ "epoch": 2.1396160558464223,
+ "grad_norm": 0.036121442914009094,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.300607681274414,
+ "step": 154
+ },
+ {
+ "epoch": 2.1535776614310644,
+ "grad_norm": 0.055716633796691895,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.307400703430176,
+ "step": 155
+ },
+ {
+ "epoch": 2.167539267015707,
+ "grad_norm": 0.037626396864652634,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.287364959716797,
+ "step": 156
+ },
+ {
+ "epoch": 2.181500872600349,
+ "grad_norm": 0.04329098388552666,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.28244686126709,
+ "step": 157
+ },
+ {
+ "epoch": 2.195462478184991,
+ "grad_norm": 0.031868450343608856,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.2602410316467285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2094240837696333,
+ "grad_norm": 0.03647344186902046,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.263988971710205,
+ "step": 159
+ },
+ {
+ "epoch": 2.223385689354276,
+ "grad_norm": 0.03141765296459198,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.263205051422119,
+ "step": 160
+ },
+ {
+ "epoch": 2.237347294938918,
+ "grad_norm": 0.028253229334950447,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.253107070922852,
+ "step": 161
+ },
+ {
+ "epoch": 2.25130890052356,
+ "grad_norm": 0.03608971834182739,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.2571702003479,
+ "step": 162
+ },
+ {
+ "epoch": 2.2652705061082026,
+ "grad_norm": 0.02644292451441288,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.241732597351074,
+ "step": 163
+ },
+ {
+ "epoch": 2.2792321116928447,
+ "grad_norm": 0.028604455292224884,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.230289459228516,
+ "step": 164
+ },
+ {
+ "epoch": 2.2931937172774868,
+ "grad_norm": 0.023704614490270615,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.226101398468018,
+ "step": 165
+ },
+ {
+ "epoch": 2.3071553228621293,
+ "grad_norm": 0.020691027864813805,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.2096405029296875,
+ "step": 166
+ },
+ {
+ "epoch": 2.3211169284467714,
+ "grad_norm": 0.025598427280783653,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.22075080871582,
+ "step": 167
+ },
+ {
+ "epoch": 2.3350785340314135,
+ "grad_norm": 0.030546002089977264,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.216886520385742,
+ "step": 168
+ },
+ {
+ "epoch": 2.349040139616056,
+ "grad_norm": 0.026066245511174202,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.2022504806518555,
+ "step": 169
+ },
+ {
+ "epoch": 2.363001745200698,
+ "grad_norm": 0.023802239447832108,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.189509391784668,
+ "step": 170
+ },
+ {
+ "epoch": 2.3769633507853403,
+ "grad_norm": 0.029805002734065056,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.186277389526367,
+ "step": 171
+ },
+ {
+ "epoch": 2.3909249563699824,
+ "grad_norm": 0.043327830731868744,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.164876937866211,
+ "step": 172
+ },
+ {
+ "epoch": 2.404886561954625,
+ "grad_norm": 0.058787211775779724,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.179936408996582,
+ "step": 173
+ },
+ {
+ "epoch": 2.418848167539267,
+ "grad_norm": 0.0787397176027298,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.1830244064331055,
+ "step": 174
+ },
+ {
+ "epoch": 2.432809773123909,
+ "grad_norm": 0.0899854376912117,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.190343856811523,
+ "step": 175
+ },
+ {
+ "epoch": 2.4467713787085517,
+ "grad_norm": 0.06520707905292511,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.185057163238525,
+ "step": 176
+ },
+ {
+ "epoch": 2.4607329842931938,
+ "grad_norm": 0.1002504974603653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.171279430389404,
+ "step": 177
+ },
+ {
+ "epoch": 2.474694589877836,
+ "grad_norm": 0.09591004997491837,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.174382209777832,
+ "step": 178
+ },
+ {
+ "epoch": 2.488656195462478,
+ "grad_norm": 0.12127121537923813,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.185717582702637,
+ "step": 179
+ },
+ {
+ "epoch": 2.5026178010471205,
+ "grad_norm": 0.09640590846538544,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.196681976318359,
+ "step": 180
+ },
+ {
+ "epoch": 2.5165794066317626,
+ "grad_norm": 0.048341453075408936,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.158774375915527,
+ "step": 181
+ },
+ {
+ "epoch": 2.530541012216405,
+ "grad_norm": 0.08206334710121155,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.17445707321167,
+ "step": 182
+ },
+ {
+ "epoch": 2.5445026178010473,
+ "grad_norm": 0.038934968411922455,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.170527458190918,
+ "step": 183
+ },
+ {
+ "epoch": 2.5584642233856894,
+ "grad_norm": 0.060987457633018494,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.1501994132995605,
+ "step": 184
+ },
+ {
+ "epoch": 2.5724258289703315,
+ "grad_norm": 0.033269546926021576,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.144186019897461,
+ "step": 185
+ },
+ {
+ "epoch": 2.5863874345549736,
+ "grad_norm": 0.046137236058712006,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.1236677169799805,
+ "step": 186
+ },
+ {
+ "epoch": 2.600349040139616,
+ "grad_norm": 0.038265109062194824,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.121979236602783,
+ "step": 187
+ },
+ {
+ "epoch": 2.614310645724258,
+ "grad_norm": 0.039128001779317856,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.124156951904297,
+ "step": 188
+ },
+ {
+ "epoch": 2.6282722513089007,
+ "grad_norm": 0.025303540751338005,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.108036994934082,
+ "step": 189
+ },
+ {
+ "epoch": 2.642233856893543,
+ "grad_norm": 0.04441308230161667,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.1211161613464355,
+ "step": 190
+ },
+ {
+ "epoch": 2.656195462478185,
+ "grad_norm": 0.03083532303571701,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.094561576843262,
+ "step": 191
+ },
+ {
+ "epoch": 2.670157068062827,
+ "grad_norm": 0.02951296605169773,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.091608047485352,
+ "step": 192
+ },
+ {
+ "epoch": 2.6841186736474696,
+ "grad_norm": 0.02385987527668476,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.087080001831055,
+ "step": 193
+ },
+ {
+ "epoch": 2.6980802792321117,
+ "grad_norm": 0.03315780311822891,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.088132381439209,
+ "step": 194
+ },
+ {
+ "epoch": 2.712041884816754,
+ "grad_norm": 0.02410835586488247,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.081460952758789,
+ "step": 195
+ },
+ {
+ "epoch": 2.7260034904013963,
+ "grad_norm": 0.033069487661123276,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.0696539878845215,
+ "step": 196
+ },
+ {
+ "epoch": 2.7399650959860384,
+ "grad_norm": 0.031288471072912216,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.047612190246582,
+ "step": 197
+ },
+ {
+ "epoch": 2.7539267015706805,
+ "grad_norm": 0.03399677947163582,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.019026756286621,
+ "step": 198
+ },
+ {
+ "epoch": 2.7678883071553226,
+ "grad_norm": 0.031948331743478775,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.050133228302002,
+ "step": 199
+ },
+ {
+ "epoch": 2.781849912739965,
+ "grad_norm": 0.049206435680389404,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.045332431793213,
+ "step": 200
+ },
+ {
+ "epoch": 2.7958115183246073,
+ "grad_norm": 0.07528849691152573,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.059202194213867,
+ "step": 201
+ },
+ {
+ "epoch": 2.80977312390925,
+ "grad_norm": 0.11815470457077026,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.077486038208008,
+ "step": 202
+ },
+ {
+ "epoch": 2.823734729493892,
+ "grad_norm": 0.0867348462343216,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.045670032501221,
+ "step": 203
+ },
+ {
+ "epoch": 2.837696335078534,
+ "grad_norm": 0.05716705322265625,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.034373760223389,
+ "step": 204
+ },
+ {
+ "epoch": 2.851657940663176,
+ "grad_norm": 0.05944284051656723,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.049911975860596,
+ "step": 205
+ },
+ {
+ "epoch": 2.8656195462478182,
+ "grad_norm": 0.03220470994710922,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.039861679077148,
+ "step": 206
+ },
+ {
+ "epoch": 2.8795811518324608,
+ "grad_norm": 0.057081177830696106,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.010714530944824,
+ "step": 207
+ },
+ {
+ "epoch": 2.893542757417103,
+ "grad_norm": 0.04037516191601753,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.018877983093262,
+ "step": 208
+ },
+ {
+ "epoch": 2.9075043630017454,
+ "grad_norm": 0.036492880433797836,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.009493827819824,
+ "step": 209
+ },
+ {
+ "epoch": 2.9214659685863875,
+ "grad_norm": 0.04785023257136345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.0140862464904785,
+ "step": 210
+ },
+ {
+ "epoch": 2.9354275741710296,
+ "grad_norm": 0.07721036672592163,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.030359268188477,
+ "step": 211
+ },
+ {
+ "epoch": 2.9493891797556717,
+ "grad_norm": 0.14220190048217773,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.045989990234375,
+ "step": 212
+ },
+ {
+ "epoch": 2.9633507853403143,
+ "grad_norm": 0.10265158116817474,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.043968200683594,
+ "step": 213
+ },
+ {
+ "epoch": 2.9773123909249564,
+ "grad_norm": 0.051398951560258865,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.022278785705566,
+ "step": 214
+ },
+ {
+ "epoch": 2.9912739965095985,
+ "grad_norm": 0.06044618785381317,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.017296314239502,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.03284750506281853,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 3.7544798851013184,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7531244158744812,
+ "eval_runtime": 60.3986,
+ "eval_samples_per_second": 40.431,
+ "eval_steps_per_second": 0.646,
+ "step": 216
+ },
+ {
+ "epoch": 3.013961605584642,
+ "grad_norm": 0.03895840048789978,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 5.9985857009887695,
+ "step": 217
+ },
+ {
+ "epoch": 3.0279232111692846,
+ "grad_norm": 0.04554992541670799,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 5.984710693359375,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418848167539267,
+ "grad_norm": 0.03359116613864899,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 5.984078407287598,
+ "step": 219
+ },
+ {
+ "epoch": 3.055846422338569,
+ "grad_norm": 0.035171881318092346,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 5.979424476623535,
+ "step": 220
+ },
+ {
+ "epoch": 3.0698080279232114,
+ "grad_norm": 0.032774072140455246,
+ "learning_rate": 0.000598278794687186,
+ "loss": 5.94467830657959,
+ "step": 221
+ },
+ {
+ "epoch": 3.0837696335078535,
+ "grad_norm": 0.032842908054590225,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 5.95986270904541,
+ "step": 222
+ },
+ {
+ "epoch": 3.0977312390924956,
+ "grad_norm": 0.02870488539338112,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 5.951186656951904,
+ "step": 223
+ },
+ {
+ "epoch": 3.1116928446771377,
+ "grad_norm": 0.030777374282479286,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 5.942493438720703,
+ "step": 224
+ },
+ {
+ "epoch": 3.1256544502617802,
+ "grad_norm": 0.026439018547534943,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 5.938275337219238,
+ "step": 225
+ },
+ {
+ "epoch": 3.1396160558464223,
+ "grad_norm": 0.028689472004771233,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 5.931982040405273,
+ "step": 226
+ },
+ {
+ "epoch": 3.1535776614310644,
+ "grad_norm": 0.028100362047553062,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 5.919692039489746,
+ "step": 227
+ },
+ {
+ "epoch": 3.167539267015707,
+ "grad_norm": 0.02407177910208702,
+ "learning_rate": 0.00059807235435179,
+ "loss": 5.909029006958008,
+ "step": 228
+ },
+ {
+ "epoch": 3.181500872600349,
+ "grad_norm": 0.0237748846411705,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 5.9131598472595215,
+ "step": 229
+ },
+ {
+ "epoch": 3.195462478184991,
+ "grad_norm": 0.022418556734919548,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 5.909396171569824,
+ "step": 230
+ },
+ {
+ "epoch": 3.2094240837696333,
+ "grad_norm": 0.024656571447849274,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 5.885243892669678,
+ "step": 231
+ },
+ {
+ "epoch": 3.223385689354276,
+ "grad_norm": 0.02136118710041046,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 5.898958206176758,
+ "step": 232
+ },
+ {
+ "epoch": 3.237347294938918,
+ "grad_norm": 0.025290116667747498,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 5.886986255645752,
+ "step": 233
+ },
+ {
+ "epoch": 3.25130890052356,
+ "grad_norm": 0.02656404674053192,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 5.897647857666016,
+ "step": 234
+ },
+ {
+ "epoch": 3.2652705061082026,
+ "grad_norm": 0.019188793376088142,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 5.875659942626953,
+ "step": 235
+ },
+ {
+ "epoch": 3.2792321116928447,
+ "grad_norm": 0.02308926172554493,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 5.859380722045898,
+ "step": 236
+ },
+ {
+ "epoch": 3.2931937172774868,
+ "grad_norm": 0.029775725677609444,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 5.873961448669434,
+ "step": 237
+ },
+ {
+ "epoch": 3.3071553228621293,
+ "grad_norm": 0.03875259310007095,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 5.863171100616455,
+ "step": 238
+ },
+ {
+ "epoch": 3.3211169284467714,
+ "grad_norm": 0.052446842193603516,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 5.886547088623047,
+ "step": 239
+ },
+ {
+ "epoch": 3.3350785340314135,
+ "grad_norm": 0.05583837628364563,
+ "learning_rate": 0.000597691359743243,
+ "loss": 5.8849287033081055,
+ "step": 240
+ },
+ {
+ "epoch": 3.349040139616056,
+ "grad_norm": 0.034729067236185074,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 5.863419532775879,
+ "step": 241
+ },
+ {
+ "epoch": 3.363001745200698,
+ "grad_norm": 0.03590013459324837,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 5.852169990539551,
+ "step": 242
+ },
+ {
+ "epoch": 3.3769633507853403,
+ "grad_norm": 0.035041507333517075,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 5.836690902709961,
+ "step": 243
+ },
+ {
+ "epoch": 3.3909249563699824,
+ "grad_norm": 0.02857939340174198,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 5.836485862731934,
+ "step": 244
+ },
+ {
+ "epoch": 3.404886561954625,
+ "grad_norm": 0.03176642581820488,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 5.860719203948975,
+ "step": 245
+ },
+ {
+ "epoch": 3.418848167539267,
+ "grad_norm": 0.026105748489499092,
+ "learning_rate": 0.000597488038559129,
+ "loss": 5.8354949951171875,
+ "step": 246
+ },
+ {
+ "epoch": 3.432809773123909,
+ "grad_norm": 0.029094981029629707,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 5.846247673034668,
+ "step": 247
+ },
+ {
+ "epoch": 3.4467713787085517,
+ "grad_norm": 0.02388785220682621,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 5.837607383728027,
+ "step": 248
+ },
+ {
+ "epoch": 3.4607329842931938,
+ "grad_norm": 0.03094472736120224,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 5.832090377807617,
+ "step": 249
+ },
+ {
+ "epoch": 3.474694589877836,
+ "grad_norm": 0.03227658197283745,
+ "learning_rate": 0.000597347745466109,
+ "loss": 5.83722448348999,
+ "step": 250
+ },
+ {
+ "epoch": 3.488656195462478,
+ "grad_norm": 0.042813923209905624,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 5.851722717285156,
+ "step": 251
+ },
+ {
+ "epoch": 3.5026178010471205,
+ "grad_norm": 0.05825525149703026,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.795719146728516,
+ "step": 252
+ },
+ {
+ "epoch": 3.5165794066317626,
+ "grad_norm": 0.0867447778582573,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.813750267028809,
+ "step": 253
+ },
+ {
+ "epoch": 3.530541012216405,
+ "grad_norm": 0.09192710369825363,
+ "learning_rate": 0.000597203657966319,
+ "loss": 5.823488235473633,
+ "step": 254
+ },
+ {
+ "epoch": 3.5445026178010473,
+ "grad_norm": 0.06069248169660568,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 5.814342021942139,
+ "step": 255
+ },
+ {
+ "epoch": 3.5584642233856894,
+ "grad_norm": 0.06259952485561371,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 5.834272384643555,
+ "step": 256
+ },
+ {
+ "epoch": 3.5724258289703315,
+ "grad_norm": 0.0513639897108078,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.807951927185059,
+ "step": 257
+ },
+ {
+ "epoch": 3.5863874345549736,
+ "grad_norm": 0.03327798470854759,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.790238857269287,
+ "step": 258
+ },
+ {
+ "epoch": 3.600349040139616,
+ "grad_norm": 0.03963128477334976,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.7956695556640625,
+ "step": 259
+ },
+ {
+ "epoch": 3.614310645724258,
+ "grad_norm": 0.031153453513979912,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 5.7874531745910645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6282722513089007,
+ "grad_norm": 0.031419433653354645,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 5.779650688171387,
+ "step": 261
+ },
+ {
+ "epoch": 3.642233856893543,
+ "grad_norm": 0.0288684144616127,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.776577949523926,
+ "step": 262
+ },
+ {
+ "epoch": 3.656195462478185,
+ "grad_norm": 0.028315918520092964,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.778571128845215,
+ "step": 263
+ },
+ {
+ "epoch": 3.670157068062827,
+ "grad_norm": 0.024977795779705048,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 5.776470184326172,
+ "step": 264
+ },
+ {
+ "epoch": 3.6841186736474696,
+ "grad_norm": 0.026822632178664207,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.755108833312988,
+ "step": 265
+ },
+ {
+ "epoch": 3.6980802792321117,
+ "grad_norm": 0.02738630585372448,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.74041748046875,
+ "step": 266
+ },
+ {
+ "epoch": 3.712041884816754,
+ "grad_norm": 0.02361280657351017,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.764027118682861,
+ "step": 267
+ },
+ {
+ "epoch": 3.7260034904013963,
+ "grad_norm": 0.029926009476184845,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.780390739440918,
+ "step": 268
+ },
+ {
+ "epoch": 3.7399650959860384,
+ "grad_norm": 0.03519507870078087,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.758105278015137,
+ "step": 269
+ },
+ {
+ "epoch": 3.7539267015706805,
+ "grad_norm": 0.030114633962512016,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.757230758666992,
+ "step": 270
+ },
+ {
+ "epoch": 3.7678883071553226,
+ "grad_norm": 0.02467748522758484,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.7615461349487305,
+ "step": 271
+ },
+ {
+ "epoch": 3.781849912739965,
+ "grad_norm": 0.03425183147192001,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.749448776245117,
+ "step": 272
+ },
+ {
+ "epoch": 3.7958115183246073,
+ "grad_norm": 0.046307582408189774,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.7356977462768555,
+ "step": 273
+ },
+ {
+ "epoch": 3.80977312390925,
+ "grad_norm": 0.054102856665849686,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.741869926452637,
+ "step": 274
+ },
+ {
+ "epoch": 3.823734729493892,
+ "grad_norm": 0.07026689499616623,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.721103668212891,
+ "step": 275
+ },
+ {
+ "epoch": 3.837696335078534,
+ "grad_norm": 0.0733039602637291,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.731414794921875,
+ "step": 276
+ },
+ {
+ "epoch": 3.851657940663176,
+ "grad_norm": 0.05342920497059822,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.734922885894775,
+ "step": 277
+ },
+ {
+ "epoch": 3.8656195462478182,
+ "grad_norm": 0.0504680834710598,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.732746124267578,
+ "step": 278
+ },
+ {
+ "epoch": 3.8795811518324608,
+ "grad_norm": 0.038991592824459076,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.719473838806152,
+ "step": 279
+ },
+ {
+ "epoch": 3.893542757417103,
+ "grad_norm": 0.04434404894709587,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.709481239318848,
+ "step": 280
+ },
+ {
+ "epoch": 3.9075043630017454,
+ "grad_norm": 0.047285184264183044,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.693531036376953,
+ "step": 281
+ },
+ {
+ "epoch": 3.9214659685863875,
+ "grad_norm": 0.045081138610839844,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.728959083557129,
+ "step": 282
+ },
+ {
+ "epoch": 3.9354275741710296,
+ "grad_norm": 0.025133663788437843,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.719696998596191,
+ "step": 283
+ },
+ {
+ "epoch": 3.9493891797556717,
+ "grad_norm": 0.03353596478700638,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.700235366821289,
+ "step": 284
+ },
+ {
+ "epoch": 3.9633507853403143,
+ "grad_norm": 0.035162411630153656,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.684836387634277,
+ "step": 285
+ },
+ {
+ "epoch": 3.9773123909249564,
+ "grad_norm": 0.04025450348854065,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.692097187042236,
+ "step": 286
+ },
+ {
+ "epoch": 3.9912739965095985,
+ "grad_norm": 0.04995397478342056,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.6766862869262695,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.031619708985090256,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 3.5519747734069824,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.714687705039978,
+ "eval_runtime": 59.637,
+ "eval_samples_per_second": 40.948,
+ "eval_steps_per_second": 0.654,
+ "step": 288
+ },
+ {
+ "epoch": 4.013961605584642,
+ "grad_norm": 0.036328598856925964,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.666294097900391,
+ "step": 289
+ },
+ {
+ "epoch": 4.027923211169284,
+ "grad_norm": 0.07326336950063705,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.6879167556762695,
+ "step": 290
+ },
+ {
+ "epoch": 4.041884816753926,
+ "grad_norm": 0.07947526127099991,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.674883842468262,
+ "step": 291
+ },
+ {
+ "epoch": 4.055846422338569,
+ "grad_norm": 0.04599885269999504,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.688086986541748,
+ "step": 292
+ },
+ {
+ "epoch": 4.069808027923211,
+ "grad_norm": 0.0465150885283947,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.649483680725098,
+ "step": 293
+ },
+ {
+ "epoch": 4.0837696335078535,
+ "grad_norm": 0.038542453199625015,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.653055667877197,
+ "step": 294
+ },
+ {
+ "epoch": 4.097731239092496,
+ "grad_norm": 0.03659164905548096,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.65955924987793,
+ "step": 295
+ },
+ {
+ "epoch": 4.111692844677138,
+ "grad_norm": 0.03824038803577423,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.650853157043457,
+ "step": 296
+ },
+ {
+ "epoch": 4.12565445026178,
+ "grad_norm": 0.03591204807162285,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.643499374389648,
+ "step": 297
+ },
+ {
+ "epoch": 4.139616055846423,
+ "grad_norm": 0.03269682824611664,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.657600402832031,
+ "step": 298
+ },
+ {
+ "epoch": 4.153577661431065,
+ "grad_norm": 0.03343537449836731,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.6352081298828125,
+ "step": 299
+ },
+ {
+ "epoch": 4.167539267015707,
+ "grad_norm": 0.034283801913261414,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.626757621765137,
+ "step": 300
+ },
+ {
+ "epoch": 4.181500872600349,
+ "grad_norm": 0.034614723175764084,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.6316680908203125,
+ "step": 301
+ },
+ {
+ "epoch": 4.195462478184991,
+ "grad_norm": 0.0340128131210804,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.648160934448242,
+ "step": 302
+ },
+ {
+ "epoch": 4.209424083769633,
+ "grad_norm": 0.03068280965089798,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.619738578796387,
+ "step": 303
+ },
+ {
+ "epoch": 4.223385689354275,
+ "grad_norm": 0.030269058421254158,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.609238624572754,
+ "step": 304
+ },
+ {
+ "epoch": 4.237347294938918,
+ "grad_norm": 0.03612440824508667,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.602194786071777,
+ "step": 305
+ },
+ {
+ "epoch": 4.2513089005235605,
+ "grad_norm": 0.044292811304330826,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.61159610748291,
+ "step": 306
+ },
+ {
+ "epoch": 4.265270506108203,
+ "grad_norm": 0.04449291527271271,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.6274237632751465,
+ "step": 307
+ },
+ {
+ "epoch": 4.279232111692845,
+ "grad_norm": 0.05014493316411972,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.589034080505371,
+ "step": 308
+ },
+ {
+ "epoch": 4.293193717277487,
+ "grad_norm": 0.05286206305027008,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.600770950317383,
+ "step": 309
+ },
+ {
+ "epoch": 4.307155322862129,
+ "grad_norm": 0.048459846526384354,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.5796098709106445,
+ "step": 310
+ },
+ {
+ "epoch": 4.321116928446771,
+ "grad_norm": 0.05329663306474686,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.590407371520996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335078534031414,
+ "grad_norm": 0.052595119923353195,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.583693504333496,
+ "step": 312
+ },
+ {
+ "epoch": 4.349040139616056,
+ "grad_norm": 0.05164601653814316,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.603281021118164,
+ "step": 313
+ },
+ {
+ "epoch": 4.363001745200698,
+ "grad_norm": 0.04657137766480446,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.591344833374023,
+ "step": 314
+ },
+ {
+ "epoch": 4.37696335078534,
+ "grad_norm": 0.03804599866271019,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.580578804016113,
+ "step": 315
+ },
+ {
+ "epoch": 4.390924956369982,
+ "grad_norm": 0.03835320845246315,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.597836971282959,
+ "step": 316
+ },
+ {
+ "epoch": 4.4048865619546245,
+ "grad_norm": 0.034059133380651474,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.567502021789551,
+ "step": 317
+ },
+ {
+ "epoch": 4.418848167539267,
+ "grad_norm": 0.03350386768579483,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.552022933959961,
+ "step": 318
+ },
+ {
+ "epoch": 4.43280977312391,
+ "grad_norm": 0.032574303448200226,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.559177875518799,
+ "step": 319
+ },
+ {
+ "epoch": 4.446771378708552,
+ "grad_norm": 0.030893854796886444,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.549288272857666,
+ "step": 320
+ },
+ {
+ "epoch": 4.460732984293194,
+ "grad_norm": 0.03255036100745201,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.5400390625,
+ "step": 321
+ },
+ {
+ "epoch": 4.474694589877836,
+ "grad_norm": 0.03009258769452572,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.5328474044799805,
+ "step": 322
+ },
+ {
+ "epoch": 4.488656195462478,
+ "grad_norm": 0.03178320452570915,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.5594482421875,
+ "step": 323
+ },
+ {
+ "epoch": 4.50261780104712,
+ "grad_norm": 0.03360465541481972,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.5402727127075195,
+ "step": 324
+ },
+ {
+ "epoch": 4.516579406631763,
+ "grad_norm": 0.04275049269199371,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.534474849700928,
+ "step": 325
+ },
+ {
+ "epoch": 4.530541012216405,
+ "grad_norm": 0.04140641540288925,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.5324859619140625,
+ "step": 326
+ },
+ {
+ "epoch": 4.544502617801047,
+ "grad_norm": 0.04769204929471016,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.4919352531433105,
+ "step": 327
+ },
+ {
+ "epoch": 4.558464223385689,
+ "grad_norm": 0.04402285814285278,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.520858287811279,
+ "step": 328
+ },
+ {
+ "epoch": 4.5724258289703315,
+ "grad_norm": 0.038526229560375214,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.523778438568115,
+ "step": 329
+ },
+ {
+ "epoch": 4.5863874345549736,
+ "grad_norm": 0.05422377213835716,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.517973899841309,
+ "step": 330
+ },
+ {
+ "epoch": 4.600349040139616,
+ "grad_norm": 0.07236721366643906,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.548083782196045,
+ "step": 331
+ },
+ {
+ "epoch": 4.614310645724259,
+ "grad_norm": 0.06861061602830887,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.535577774047852,
+ "step": 332
+ },
+ {
+ "epoch": 4.628272251308901,
+ "grad_norm": 0.07148340344429016,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.54155158996582,
+ "step": 333
+ },
+ {
+ "epoch": 4.642233856893543,
+ "grad_norm": 0.07838302850723267,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.552852630615234,
+ "step": 334
+ },
+ {
+ "epoch": 4.656195462478185,
+ "grad_norm": 0.10946758091449738,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.555066108703613,
+ "step": 335
+ },
+ {
+ "epoch": 4.670157068062827,
+ "grad_norm": 0.10143530368804932,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.584932804107666,
+ "step": 336
+ },
+ {
+ "epoch": 4.684118673647469,
+ "grad_norm": 0.08926606178283691,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.570462226867676,
+ "step": 337
+ },
+ {
+ "epoch": 4.698080279232112,
+ "grad_norm": 0.07060826569795609,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.5716753005981445,
+ "step": 338
+ },
+ {
+ "epoch": 4.712041884816754,
+ "grad_norm": 0.0418456569314003,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.556496620178223,
+ "step": 339
+ },
+ {
+ "epoch": 4.726003490401396,
+ "grad_norm": 0.05952411890029907,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.567801475524902,
+ "step": 340
+ },
+ {
+ "epoch": 4.739965095986038,
+ "grad_norm": 0.042999885976314545,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.537403106689453,
+ "step": 341
+ },
+ {
+ "epoch": 4.7539267015706805,
+ "grad_norm": 0.04426449537277222,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.552387237548828,
+ "step": 342
+ },
+ {
+ "epoch": 4.767888307155323,
+ "grad_norm": 0.040835145860910416,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.53051233291626,
+ "step": 343
+ },
+ {
+ "epoch": 4.781849912739965,
+ "grad_norm": 0.03819291293621063,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.531174659729004,
+ "step": 344
+ },
+ {
+ "epoch": 4.795811518324607,
+ "grad_norm": 0.030088599771261215,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.499328136444092,
+ "step": 345
+ },
+ {
+ "epoch": 4.80977312390925,
+ "grad_norm": 0.029984362423419952,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.504540920257568,
+ "step": 346
+ },
+ {
+ "epoch": 4.823734729493892,
+ "grad_norm": 0.027402915060520172,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.50899076461792,
+ "step": 347
+ },
+ {
+ "epoch": 4.837696335078534,
+ "grad_norm": 0.025096213445067406,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.480479717254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.851657940663176,
+ "grad_norm": 0.026828132569789886,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.501729965209961,
+ "step": 349
+ },
+ {
+ "epoch": 4.865619546247818,
+ "grad_norm": 0.023715168237686157,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.466860771179199,
+ "step": 350
+ },
+ {
+ "epoch": 4.879581151832461,
+ "grad_norm": 0.024503640830516815,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.491827964782715,
+ "step": 351
+ },
+ {
+ "epoch": 4.893542757417103,
+ "grad_norm": 0.020891105756163597,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.478384971618652,
+ "step": 352
+ },
+ {
+ "epoch": 4.907504363001745,
+ "grad_norm": 0.021373765543103218,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.4612135887146,
+ "step": 353
+ },
+ {
+ "epoch": 4.9214659685863875,
+ "grad_norm": 0.021773340180516243,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.4547319412231445,
+ "step": 354
+ },
+ {
+ "epoch": 4.93542757417103,
+ "grad_norm": 0.019637586548924446,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.431741714477539,
+ "step": 355
+ },
+ {
+ "epoch": 4.949389179755672,
+ "grad_norm": 0.023108696565032005,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.448612213134766,
+ "step": 356
+ },
+ {
+ "epoch": 4.963350785340314,
+ "grad_norm": 0.021935103461146355,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.447112083435059,
+ "step": 357
+ },
+ {
+ "epoch": 4.977312390924956,
+ "grad_norm": 0.023260753601789474,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.436091423034668,
+ "step": 358
+ },
+ {
+ "epoch": 4.991273996509599,
+ "grad_norm": 0.021355057135224342,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.408421993255615,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.014992325566709042,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 3.4004223346710205,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6834986805915833,
+ "eval_runtime": 60.4148,
+ "eval_samples_per_second": 40.421,
+ "eval_steps_per_second": 0.646,
+ "step": 360
+ },
+ {
+ "epoch": 5.013961605584642,
+ "grad_norm": 0.02471218630671501,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.396238803863525,
+ "step": 361
+ },
+ {
+ "epoch": 5.027923211169284,
+ "grad_norm": 0.027672411873936653,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.4048919677734375,
+ "step": 362
+ },
+ {
+ "epoch": 5.041884816753926,
+ "grad_norm": 0.031560689210891724,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.399665832519531,
+ "step": 363
+ },
+ {
+ "epoch": 5.055846422338569,
+ "grad_norm": 0.03459061309695244,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.396360397338867,
+ "step": 364
+ },
+ {
+ "epoch": 5.069808027923211,
+ "grad_norm": 0.04040960222482681,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.405309677124023,
+ "step": 365
+ },
+ {
+ "epoch": 5.0837696335078535,
+ "grad_norm": 0.05444670096039772,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.408894062042236,
+ "step": 366
+ },
+ {
+ "epoch": 5.097731239092496,
+ "grad_norm": 0.05816745012998581,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.403566360473633,
+ "step": 367
+ },
+ {
+ "epoch": 5.111692844677138,
+ "grad_norm": 0.06330069154500961,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.395961761474609,
+ "step": 368
+ },
+ {
+ "epoch": 5.12565445026178,
+ "grad_norm": 0.07458917796611786,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.400968551635742,
+ "step": 369
+ },
+ {
+ "epoch": 5.139616055846423,
+ "grad_norm": 0.06231356039643288,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.412624359130859,
+ "step": 370
+ },
+ {
+ "epoch": 5.153577661431065,
+ "grad_norm": 0.05212346464395523,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.394794464111328,
+ "step": 371
+ },
+ {
+ "epoch": 5.167539267015707,
+ "grad_norm": 0.050003375858068466,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.380125045776367,
+ "step": 372
+ },
+ {
+ "epoch": 5.181500872600349,
+ "grad_norm": 0.048393815755844116,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.390806198120117,
+ "step": 373
+ },
+ {
+ "epoch": 5.195462478184991,
+ "grad_norm": 0.04200794920325279,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.387467384338379,
+ "step": 374
+ },
+ {
+ "epoch": 5.209424083769633,
+ "grad_norm": 0.04931448772549629,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.37552547454834,
+ "step": 375
+ },
+ {
+ "epoch": 5.223385689354275,
+ "grad_norm": 0.03851857781410217,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.373126983642578,
+ "step": 376
+ },
+ {
+ "epoch": 5.237347294938918,
+ "grad_norm": 0.03458717465400696,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.346826553344727,
+ "step": 377
+ },
+ {
+ "epoch": 5.2513089005235605,
+ "grad_norm": 0.031107677146792412,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.353341102600098,
+ "step": 378
+ },
+ {
+ "epoch": 5.265270506108203,
+ "grad_norm": 0.03077058307826519,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.358076095581055,
+ "step": 379
+ },
+ {
+ "epoch": 5.279232111692845,
+ "grad_norm": 0.028170401230454445,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.354019641876221,
+ "step": 380
+ },
+ {
+ "epoch": 5.293193717277487,
+ "grad_norm": 0.02871990203857422,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.368258476257324,
+ "step": 381
+ },
+ {
+ "epoch": 5.307155322862129,
+ "grad_norm": 0.03046584501862526,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.342505931854248,
+ "step": 382
+ },
+ {
+ "epoch": 5.321116928446771,
+ "grad_norm": 0.02781563811004162,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.339411735534668,
+ "step": 383
+ },
+ {
+ "epoch": 5.335078534031414,
+ "grad_norm": 0.028825288638472557,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.341948509216309,
+ "step": 384
+ },
+ {
+ "epoch": 5.349040139616056,
+ "grad_norm": 0.026684079319238663,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.340881824493408,
+ "step": 385
+ },
+ {
+ "epoch": 5.363001745200698,
+ "grad_norm": 0.02429167740046978,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.342733383178711,
+ "step": 386
+ },
+ {
+ "epoch": 5.37696335078534,
+ "grad_norm": 0.02855239436030388,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.313878536224365,
+ "step": 387
+ },
+ {
+ "epoch": 5.390924956369982,
+ "grad_norm": 0.029651548713445663,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.312665939331055,
+ "step": 388
+ },
+ {
+ "epoch": 5.4048865619546245,
+ "grad_norm": 0.027191132307052612,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3086838722229,
+ "step": 389
+ },
+ {
+ "epoch": 5.418848167539267,
+ "grad_norm": 0.026719525456428528,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.305359840393066,
+ "step": 390
+ },
+ {
+ "epoch": 5.43280977312391,
+ "grad_norm": 0.026763100177049637,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.3164381980896,
+ "step": 391
+ },
+ {
+ "epoch": 5.446771378708552,
+ "grad_norm": 0.03125109151005745,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.3083953857421875,
+ "step": 392
+ },
+ {
+ "epoch": 5.460732984293194,
+ "grad_norm": 0.0394827201962471,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.275996208190918,
+ "step": 393
+ },
+ {
+ "epoch": 5.474694589877836,
+ "grad_norm": 0.040579475462436676,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.292995929718018,
+ "step": 394
+ },
+ {
+ "epoch": 5.488656195462478,
+ "grad_norm": 0.037223003804683685,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.298526763916016,
+ "step": 395
+ },
+ {
+ "epoch": 5.50261780104712,
+ "grad_norm": 0.04207765311002731,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.287985801696777,
+ "step": 396
+ },
+ {
+ "epoch": 5.516579406631763,
+ "grad_norm": 0.048299964517354965,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.244608402252197,
+ "step": 397
+ },
+ {
+ "epoch": 5.530541012216405,
+ "grad_norm": 0.04650307446718216,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.285453796386719,
+ "step": 398
+ },
+ {
+ "epoch": 5.544502617801047,
+ "grad_norm": 0.045339711010456085,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.312187671661377,
+ "step": 399
+ },
+ {
+ "epoch": 5.558464223385689,
+ "grad_norm": 0.041067708283662796,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.292692184448242,
+ "step": 400
+ },
+ {
+ "epoch": 5.5724258289703315,
+ "grad_norm": 0.05455026403069496,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.253364562988281,
+ "step": 401
+ },
+ {
+ "epoch": 5.5863874345549736,
+ "grad_norm": 0.06660915911197662,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.262646675109863,
+ "step": 402
+ },
+ {
+ "epoch": 5.600349040139616,
+ "grad_norm": 0.06634793430566788,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.274322986602783,
+ "step": 403
+ },
+ {
+ "epoch": 5.614310645724259,
+ "grad_norm": 0.08996330946683884,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.287951469421387,
+ "step": 404
+ },
+ {
+ "epoch": 5.628272251308901,
+ "grad_norm": 0.08526470512151718,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.3007097244262695,
+ "step": 405
+ },
+ {
+ "epoch": 5.642233856893543,
+ "grad_norm": 0.08766836673021317,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.332538604736328,
+ "step": 406
+ },
+ {
+ "epoch": 5.656195462478185,
+ "grad_norm": 0.07055048644542694,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.296169281005859,
+ "step": 407
+ },
+ {
+ "epoch": 5.670157068062827,
+ "grad_norm": 0.07520575076341629,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.309683322906494,
+ "step": 408
+ },
+ {
+ "epoch": 5.684118673647469,
+ "grad_norm": 0.06115756183862686,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.311301231384277,
+ "step": 409
+ },
+ {
+ "epoch": 5.698080279232112,
+ "grad_norm": 0.05862593650817871,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.28066349029541,
+ "step": 410
+ },
+ {
+ "epoch": 5.712041884816754,
+ "grad_norm": 0.06169601157307625,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.286840438842773,
+ "step": 411
+ },
+ {
+ "epoch": 5.726003490401396,
+ "grad_norm": 0.05270719155669212,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.302594184875488,
+ "step": 412
+ },
+ {
+ "epoch": 5.739965095986038,
+ "grad_norm": 0.03912634402513504,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.262563705444336,
+ "step": 413
+ },
+ {
+ "epoch": 5.7539267015706805,
+ "grad_norm": 0.03215811401605606,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.269310474395752,
+ "step": 414
+ },
+ {
+ "epoch": 5.767888307155323,
+ "grad_norm": 0.03236516937613487,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.268766403198242,
+ "step": 415
+ },
+ {
+ "epoch": 5.781849912739965,
+ "grad_norm": 0.027429452165961266,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.283801555633545,
+ "step": 416
+ },
+ {
+ "epoch": 5.795811518324607,
+ "grad_norm": 0.02990538813173771,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.22291374206543,
+ "step": 417
+ },
+ {
+ "epoch": 5.80977312390925,
+ "grad_norm": 0.030410146340727806,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.2281599044799805,
+ "step": 418
+ },
+ {
+ "epoch": 5.823734729493892,
+ "grad_norm": 0.030068187043070793,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.250294208526611,
+ "step": 419
+ },
+ {
+ "epoch": 5.837696335078534,
+ "grad_norm": 0.02800513058900833,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.226508617401123,
+ "step": 420
+ },
+ {
+ "epoch": 5.851657940663176,
+ "grad_norm": 0.023823332041502,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.247838020324707,
+ "step": 421
+ },
+ {
+ "epoch": 5.865619546247818,
+ "grad_norm": 0.024974249303340912,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.2392473220825195,
+ "step": 422
+ },
+ {
+ "epoch": 5.879581151832461,
+ "grad_norm": 0.023320546373724937,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.209543704986572,
+ "step": 423
+ },
+ {
+ "epoch": 5.893542757417103,
+ "grad_norm": 0.024201491847634315,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.219725608825684,
+ "step": 424
+ },
+ {
+ "epoch": 5.907504363001745,
+ "grad_norm": 0.021649666130542755,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.203141212463379,
+ "step": 425
+ },
+ {
+ "epoch": 5.9214659685863875,
+ "grad_norm": 0.0248002577573061,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.179058074951172,
+ "step": 426
+ },
+ {
+ "epoch": 5.93542757417103,
+ "grad_norm": 0.028549587354063988,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.202126502990723,
+ "step": 427
+ },
+ {
+ "epoch": 5.949389179755672,
+ "grad_norm": 0.0331057533621788,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.208107948303223,
+ "step": 428
+ },
+ {
+ "epoch": 5.963350785340314,
+ "grad_norm": 0.031843364238739014,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.184683799743652,
+ "step": 429
+ },
+ {
+ "epoch": 5.977312390924956,
+ "grad_norm": 0.024927377700805664,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.180486679077148,
+ "step": 430
+ },
+ {
+ "epoch": 5.991273996509599,
+ "grad_norm": 0.026685267686843872,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.186746597290039,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.018545351922512054,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 3.217043876647949,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.653471827507019,
+ "eval_runtime": 61.1586,
+ "eval_samples_per_second": 39.929,
+ "eval_steps_per_second": 0.638,
+ "step": 432
+ },
+ {
+ "epoch": 6.013961605584642,
+ "grad_norm": 0.025291118770837784,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.156534194946289,
+ "step": 433
+ },
+ {
+ "epoch": 6.027923211169284,
+ "grad_norm": 0.03774752840399742,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.144041538238525,
+ "step": 434
+ },
+ {
+ "epoch": 6.041884816753926,
+ "grad_norm": 0.04244692996144295,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.158862113952637,
+ "step": 435
+ },
+ {
+ "epoch": 6.055846422338569,
+ "grad_norm": 0.039782389998435974,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.166165351867676,
+ "step": 436
+ },
+ {
+ "epoch": 6.069808027923211,
+ "grad_norm": 0.05489896610379219,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.155766487121582,
+ "step": 437
+ },
+ {
+ "epoch": 6.0837696335078535,
+ "grad_norm": 0.059844326227903366,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.153733253479004,
+ "step": 438
+ },
+ {
+ "epoch": 6.097731239092496,
+ "grad_norm": 0.05539938807487488,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.130112648010254,
+ "step": 439
+ },
+ {
+ "epoch": 6.111692844677138,
+ "grad_norm": 0.05283183231949806,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.136910915374756,
+ "step": 440
+ },
+ {
+ "epoch": 6.12565445026178,
+ "grad_norm": 0.054407261312007904,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.165216445922852,
+ "step": 441
+ },
+ {
+ "epoch": 6.139616055846423,
+ "grad_norm": 0.0509788915514946,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.140888214111328,
+ "step": 442
+ },
+ {
+ "epoch": 6.153577661431065,
+ "grad_norm": 0.05018748342990875,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.13115930557251,
+ "step": 443
+ },
+ {
+ "epoch": 6.167539267015707,
+ "grad_norm": 0.047012437134981155,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.146507263183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181500872600349,
+ "grad_norm": 0.04943335801362991,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.135608673095703,
+ "step": 445
+ },
+ {
+ "epoch": 6.195462478184991,
+ "grad_norm": 0.05304260924458504,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.136394023895264,
+ "step": 446
+ },
+ {
+ "epoch": 6.209424083769633,
+ "grad_norm": 0.04554047808051109,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.124661922454834,
+ "step": 447
+ },
+ {
+ "epoch": 6.223385689354275,
+ "grad_norm": 0.040672432631254196,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.129773139953613,
+ "step": 448
+ },
+ {
+ "epoch": 6.237347294938918,
+ "grad_norm": 0.03442879393696785,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.128028392791748,
+ "step": 449
+ },
+ {
+ "epoch": 6.2513089005235605,
+ "grad_norm": 0.03927934914827347,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.092485427856445,
+ "step": 450
+ },
+ {
+ "epoch": 6.265270506108203,
+ "grad_norm": 0.029380977153778076,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.077669143676758,
+ "step": 451
+ },
+ {
+ "epoch": 6.279232111692845,
+ "grad_norm": 0.03331821784377098,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.138857841491699,
+ "step": 452
+ },
+ {
+ "epoch": 6.293193717277487,
+ "grad_norm": 0.02939118817448616,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.082937240600586,
+ "step": 453
+ },
+ {
+ "epoch": 6.307155322862129,
+ "grad_norm": 0.02730412967503071,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.093183517456055,
+ "step": 454
+ },
+ {
+ "epoch": 6.321116928446771,
+ "grad_norm": 0.02964780665934086,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.1028151512146,
+ "step": 455
+ },
+ {
+ "epoch": 6.335078534031414,
+ "grad_norm": 0.026350129395723343,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.066362380981445,
+ "step": 456
+ },
+ {
+ "epoch": 6.349040139616056,
+ "grad_norm": 0.027010295540094376,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.070037841796875,
+ "step": 457
+ },
+ {
+ "epoch": 6.363001745200698,
+ "grad_norm": 0.02906128577888012,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.107708930969238,
+ "step": 458
+ },
+ {
+ "epoch": 6.37696335078534,
+ "grad_norm": 0.028619443997740746,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.073182582855225,
+ "step": 459
+ },
+ {
+ "epoch": 6.390924956369982,
+ "grad_norm": 0.029981834813952446,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.076602935791016,
+ "step": 460
+ },
+ {
+ "epoch": 6.4048865619546245,
+ "grad_norm": 0.03255423903465271,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.063301086425781,
+ "step": 461
+ },
+ {
+ "epoch": 6.418848167539267,
+ "grad_norm": 0.03863866627216339,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.041812896728516,
+ "step": 462
+ },
+ {
+ "epoch": 6.43280977312391,
+ "grad_norm": 0.03860306739807129,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.06298828125,
+ "step": 463
+ },
+ {
+ "epoch": 6.446771378708552,
+ "grad_norm": 0.03378435596823692,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.0434370040893555,
+ "step": 464
+ },
+ {
+ "epoch": 6.460732984293194,
+ "grad_norm": 0.03276759013533592,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.036187171936035,
+ "step": 465
+ },
+ {
+ "epoch": 6.474694589877836,
+ "grad_norm": 0.04130513221025467,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.039098739624023,
+ "step": 466
+ },
+ {
+ "epoch": 6.488656195462478,
+ "grad_norm": 0.05389940366148949,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.062798976898193,
+ "step": 467
+ },
+ {
+ "epoch": 6.50261780104712,
+ "grad_norm": 0.05881704390048981,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.067028522491455,
+ "step": 468
+ },
+ {
+ "epoch": 6.516579406631763,
+ "grad_norm": 0.061897799372673035,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.056947231292725,
+ "step": 469
+ },
+ {
+ "epoch": 6.530541012216405,
+ "grad_norm": 0.06549323350191116,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.089262008666992,
+ "step": 470
+ },
+ {
+ "epoch": 6.544502617801047,
+ "grad_norm": 0.048082590103149414,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.055979251861572,
+ "step": 471
+ },
+ {
+ "epoch": 6.558464223385689,
+ "grad_norm": 0.03768027946352959,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.0660600662231445,
+ "step": 472
+ },
+ {
+ "epoch": 6.5724258289703315,
+ "grad_norm": 0.043922603130340576,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 5.036745071411133,
+ "step": 473
+ },
+ {
+ "epoch": 6.5863874345549736,
+ "grad_norm": 0.044848836958408356,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.073650360107422,
+ "step": 474
+ },
+ {
+ "epoch": 6.600349040139616,
+ "grad_norm": 0.05841026455163956,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.053045272827148,
+ "step": 475
+ },
+ {
+ "epoch": 6.614310645724259,
+ "grad_norm": 0.0650017261505127,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 5.052206039428711,
+ "step": 476
+ },
+ {
+ "epoch": 6.628272251308901,
+ "grad_norm": 0.048979002982378006,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 5.054277420043945,
+ "step": 477
+ },
+ {
+ "epoch": 6.642233856893543,
+ "grad_norm": 0.05085799843072891,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 5.063844680786133,
+ "step": 478
+ },
+ {
+ "epoch": 6.656195462478185,
+ "grad_norm": 0.04336017370223999,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 5.045474052429199,
+ "step": 479
+ },
+ {
+ "epoch": 6.670157068062827,
+ "grad_norm": 0.04113380238413811,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 5.031338214874268,
+ "step": 480
+ },
+ {
+ "epoch": 6.684118673647469,
+ "grad_norm": 0.03199538215994835,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 5.030025959014893,
+ "step": 481
+ },
+ {
+ "epoch": 6.698080279232112,
+ "grad_norm": 0.03043113648891449,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 5.033163070678711,
+ "step": 482
+ },
+ {
+ "epoch": 6.712041884816754,
+ "grad_norm": 0.031694378703832626,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 5.004739761352539,
+ "step": 483
+ },
+ {
+ "epoch": 6.726003490401396,
+ "grad_norm": 0.030243348330259323,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 5.012566089630127,
+ "step": 484
+ },
+ {
+ "epoch": 6.739965095986038,
+ "grad_norm": 0.02937416359782219,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 5.039803504943848,
+ "step": 485
+ },
+ {
+ "epoch": 6.7539267015706805,
+ "grad_norm": 0.029616443440318108,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 5.026117324829102,
+ "step": 486
+ },
+ {
+ "epoch": 6.767888307155323,
+ "grad_norm": 0.029567131772637367,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 5.006569862365723,
+ "step": 487
+ },
+ {
+ "epoch": 6.781849912739965,
+ "grad_norm": 0.029443370178341866,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.9941487312316895,
+ "step": 488
+ },
+ {
+ "epoch": 6.795811518324607,
+ "grad_norm": 0.03434352949261665,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.976431846618652,
+ "step": 489
+ },
+ {
+ "epoch": 6.80977312390925,
+ "grad_norm": 0.03845234215259552,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.993168354034424,
+ "step": 490
+ },
+ {
+ "epoch": 6.823734729493892,
+ "grad_norm": 0.05058208107948303,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.030162334442139,
+ "step": 491
+ },
+ {
+ "epoch": 6.837696335078534,
+ "grad_norm": 0.05961710214614868,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 5.007053375244141,
+ "step": 492
+ },
+ {
+ "epoch": 6.851657940663176,
+ "grad_norm": 0.056259628385305405,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 5.020356178283691,
+ "step": 493
+ },
+ {
+ "epoch": 6.865619546247818,
+ "grad_norm": 0.048628658056259155,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 5.037652969360352,
+ "step": 494
+ },
+ {
+ "epoch": 6.879581151832461,
+ "grad_norm": 0.040836453437805176,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.990390777587891,
+ "step": 495
+ },
+ {
+ "epoch": 6.893542757417103,
+ "grad_norm": 0.04112066701054573,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 5.017822742462158,
+ "step": 496
+ },
+ {
+ "epoch": 6.907504363001745,
+ "grad_norm": 0.03956829756498337,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.993971824645996,
+ "step": 497
+ },
+ {
+ "epoch": 6.9214659685863875,
+ "grad_norm": 0.03887917473912239,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.982889175415039,
+ "step": 498
+ },
+ {
+ "epoch": 6.93542757417103,
+ "grad_norm": 0.031012587249279022,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.961717128753662,
+ "step": 499
+ },
+ {
+ "epoch": 6.949389179755672,
+ "grad_norm": 0.03443489596247673,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.9904046058654785,
+ "step": 500
+ },
+ {
+ "epoch": 6.963350785340314,
+ "grad_norm": 0.035175010561943054,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.983619213104248,
+ "step": 501
+ },
+ {
+ "epoch": 6.977312390924956,
+ "grad_norm": 0.031724050641059875,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 5.000522613525391,
+ "step": 502
+ },
+ {
+ "epoch": 6.991273996509599,
+ "grad_norm": 0.032037995755672455,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.947064399719238,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.01883271522819996,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 3.112166166305542,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6296830177307129,
+ "eval_runtime": 59.5495,
+ "eval_samples_per_second": 41.008,
+ "eval_steps_per_second": 0.655,
+ "step": 504
+ },
+ {
+ "epoch": 7.013961605584642,
+ "grad_norm": 0.031091563403606415,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.919820785522461,
+ "step": 505
+ },
+ {
+ "epoch": 7.027923211169284,
+ "grad_norm": 0.04074421525001526,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.945765972137451,
+ "step": 506
+ },
+ {
+ "epoch": 7.041884816753926,
+ "grad_norm": 0.04225052148103714,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.935208797454834,
+ "step": 507
+ },
+ {
+ "epoch": 7.055846422338569,
+ "grad_norm": 0.0399051196873188,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.951314926147461,
+ "step": 508
+ },
+ {
+ "epoch": 7.069808027923211,
+ "grad_norm": 0.04785890504717827,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.92843770980835,
+ "step": 509
+ },
+ {
+ "epoch": 7.0837696335078535,
+ "grad_norm": 0.04922301694750786,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.901020050048828,
+ "step": 510
+ },
+ {
+ "epoch": 7.097731239092496,
+ "grad_norm": 0.05307295173406601,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.961151599884033,
+ "step": 511
+ },
+ {
+ "epoch": 7.111692844677138,
+ "grad_norm": 0.07003362476825714,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.9427876472473145,
+ "step": 512
+ },
+ {
+ "epoch": 7.12565445026178,
+ "grad_norm": 0.06498776376247406,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.947838306427002,
+ "step": 513
+ },
+ {
+ "epoch": 7.139616055846423,
+ "grad_norm": 0.058723628520965576,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.938859939575195,
+ "step": 514
+ },
+ {
+ "epoch": 7.153577661431065,
+ "grad_norm": 0.06590746343135834,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.962221145629883,
+ "step": 515
+ },
+ {
+ "epoch": 7.167539267015707,
+ "grad_norm": 0.052714601159095764,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.931011199951172,
+ "step": 516
+ },
+ {
+ "epoch": 7.181500872600349,
+ "grad_norm": 0.04296804964542389,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.93586540222168,
+ "step": 517
+ },
+ {
+ "epoch": 7.195462478184991,
+ "grad_norm": 0.03699173778295517,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.910763263702393,
+ "step": 518
+ },
+ {
+ "epoch": 7.209424083769633,
+ "grad_norm": 0.0317913293838501,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.9256439208984375,
+ "step": 519
+ },
+ {
+ "epoch": 7.223385689354275,
+ "grad_norm": 0.030313925817608833,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.938329696655273,
+ "step": 520
+ },
+ {
+ "epoch": 7.237347294938918,
+ "grad_norm": 0.02879229187965393,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.924966812133789,
+ "step": 521
+ },
+ {
+ "epoch": 7.2513089005235605,
+ "grad_norm": 0.02689451351761818,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.9079742431640625,
+ "step": 522
+ },
+ {
+ "epoch": 7.265270506108203,
+ "grad_norm": 0.028403930366039276,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.913238525390625,
+ "step": 523
+ },
+ {
+ "epoch": 7.279232111692845,
+ "grad_norm": 0.025912854820489883,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.896944046020508,
+ "step": 524
+ },
+ {
+ "epoch": 7.293193717277487,
+ "grad_norm": 0.024899760261178017,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.884920120239258,
+ "step": 525
+ },
+ {
+ "epoch": 7.307155322862129,
+ "grad_norm": 0.026912573724985123,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.9181060791015625,
+ "step": 526
+ },
+ {
+ "epoch": 7.321116928446771,
+ "grad_norm": 0.029539721086621284,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.914867401123047,
+ "step": 527
+ },
+ {
+ "epoch": 7.335078534031414,
+ "grad_norm": 0.028527872636914253,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.907217979431152,
+ "step": 528
+ },
+ {
+ "epoch": 7.349040139616056,
+ "grad_norm": 0.025879016146063805,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.905783176422119,
+ "step": 529
+ },
+ {
+ "epoch": 7.363001745200698,
+ "grad_norm": 0.026395199820399284,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.880505561828613,
+ "step": 530
+ },
+ {
+ "epoch": 7.37696335078534,
+ "grad_norm": 0.025072351098060608,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.870449066162109,
+ "step": 531
+ },
+ {
+ "epoch": 7.390924956369982,
+ "grad_norm": 0.02751019410789013,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.888952255249023,
+ "step": 532
+ },
+ {
+ "epoch": 7.4048865619546245,
+ "grad_norm": 0.025810658931732178,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.894552230834961,
+ "step": 533
+ },
+ {
+ "epoch": 7.418848167539267,
+ "grad_norm": 0.02399870567023754,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.8768229484558105,
+ "step": 534
+ },
+ {
+ "epoch": 7.43280977312391,
+ "grad_norm": 0.024639368057250977,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.9030914306640625,
+ "step": 535
+ },
+ {
+ "epoch": 7.446771378708552,
+ "grad_norm": 0.02543291077017784,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.86467170715332,
+ "step": 536
+ },
+ {
+ "epoch": 7.460732984293194,
+ "grad_norm": 0.02702544629573822,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.888283729553223,
+ "step": 537
+ },
+ {
+ "epoch": 7.474694589877836,
+ "grad_norm": 0.028904743492603302,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.866081237792969,
+ "step": 538
+ },
+ {
+ "epoch": 7.488656195462478,
+ "grad_norm": 0.03055848367512226,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.890739440917969,
+ "step": 539
+ },
+ {
+ "epoch": 7.50261780104712,
+ "grad_norm": 0.03410084918141365,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.862918853759766,
+ "step": 540
+ },
+ {
+ "epoch": 7.516579406631763,
+ "grad_norm": 0.03738495334982872,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.864728927612305,
+ "step": 541
+ },
+ {
+ "epoch": 7.530541012216405,
+ "grad_norm": 0.037646882236003876,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.865185737609863,
+ "step": 542
+ },
+ {
+ "epoch": 7.544502617801047,
+ "grad_norm": 0.03418618440628052,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.866114139556885,
+ "step": 543
+ },
+ {
+ "epoch": 7.558464223385689,
+ "grad_norm": 0.036924004554748535,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.883186340332031,
+ "step": 544
+ },
+ {
+ "epoch": 7.5724258289703315,
+ "grad_norm": 0.03590598329901695,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.863725662231445,
+ "step": 545
+ },
+ {
+ "epoch": 7.5863874345549736,
+ "grad_norm": 0.037363454699516296,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.841030120849609,
+ "step": 546
+ },
+ {
+ "epoch": 7.600349040139616,
+ "grad_norm": 0.046537864953279495,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.873837471008301,
+ "step": 547
+ },
+ {
+ "epoch": 7.614310645724259,
+ "grad_norm": 0.057226769626140594,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.865679740905762,
+ "step": 548
+ },
+ {
+ "epoch": 7.628272251308901,
+ "grad_norm": 0.06175961345434189,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.888645648956299,
+ "step": 549
+ },
+ {
+ "epoch": 7.642233856893543,
+ "grad_norm": 0.05063694715499878,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.877828598022461,
+ "step": 550
+ },
+ {
+ "epoch": 7.656195462478185,
+ "grad_norm": 0.03950628638267517,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.88637638092041,
+ "step": 551
+ },
+ {
+ "epoch": 7.670157068062827,
+ "grad_norm": 0.04183337092399597,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.8597917556762695,
+ "step": 552
+ },
+ {
+ "epoch": 7.684118673647469,
+ "grad_norm": 0.04642999917268753,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.880380153656006,
+ "step": 553
+ },
+ {
+ "epoch": 7.698080279232112,
+ "grad_norm": 0.05333106964826584,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.853932857513428,
+ "step": 554
+ },
+ {
+ "epoch": 7.712041884816754,
+ "grad_norm": 0.04764975979924202,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.870235443115234,
+ "step": 555
+ },
+ {
+ "epoch": 7.726003490401396,
+ "grad_norm": 0.043072231113910675,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.822697639465332,
+ "step": 556
+ },
+ {
+ "epoch": 7.739965095986038,
+ "grad_norm": 0.03913857042789459,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.849332809448242,
+ "step": 557
+ },
+ {
+ "epoch": 7.7539267015706805,
+ "grad_norm": 0.034836407750844955,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.842280387878418,
+ "step": 558
+ },
+ {
+ "epoch": 7.767888307155323,
+ "grad_norm": 0.033344294875860214,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.837811470031738,
+ "step": 559
+ },
+ {
+ "epoch": 7.781849912739965,
+ "grad_norm": 0.03342701122164726,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.876040458679199,
+ "step": 560
+ },
+ {
+ "epoch": 7.795811518324607,
+ "grad_norm": 0.034638747572898865,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.830340385437012,
+ "step": 561
+ },
+ {
+ "epoch": 7.80977312390925,
+ "grad_norm": 0.03358135372400284,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.828607559204102,
+ "step": 562
+ },
+ {
+ "epoch": 7.823734729493892,
+ "grad_norm": 0.036001041531562805,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.827905654907227,
+ "step": 563
+ },
+ {
+ "epoch": 7.837696335078534,
+ "grad_norm": 0.038079630583524704,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.866955280303955,
+ "step": 564
+ },
+ {
+ "epoch": 7.851657940663176,
+ "grad_norm": 0.03728470206260681,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.828039169311523,
+ "step": 565
+ },
+ {
+ "epoch": 7.865619546247818,
+ "grad_norm": 0.03920375928282738,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.841631889343262,
+ "step": 566
+ },
+ {
+ "epoch": 7.879581151832461,
+ "grad_norm": 0.032353926450014114,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.861110210418701,
+ "step": 567
+ },
+ {
+ "epoch": 7.893542757417103,
+ "grad_norm": 0.03241422027349472,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.818441390991211,
+ "step": 568
+ },
+ {
+ "epoch": 7.907504363001745,
+ "grad_norm": 0.03879080340266228,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.823749542236328,
+ "step": 569
+ },
+ {
+ "epoch": 7.9214659685863875,
+ "grad_norm": 0.03486498072743416,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.830697059631348,
+ "step": 570
+ },
+ {
+ "epoch": 7.93542757417103,
+ "grad_norm": 0.03279713913798332,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.815736770629883,
+ "step": 571
+ },
+ {
+ "epoch": 7.949389179755672,
+ "grad_norm": 0.03313925117254257,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.8364033699035645,
+ "step": 572
+ },
+ {
+ "epoch": 7.963350785340314,
+ "grad_norm": 0.03405062481760979,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.816010475158691,
+ "step": 573
+ },
+ {
+ "epoch": 7.977312390924956,
+ "grad_norm": 0.03259236365556717,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.813413619995117,
+ "step": 574
+ },
+ {
+ "epoch": 7.991273996509599,
+ "grad_norm": 0.030408985912799835,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.833840370178223,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.022680889815092087,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 2.99565052986145,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6126086115837097,
+ "eval_runtime": 60.2662,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 576
+ },
+ {
+ "epoch": 8.013961605584642,
+ "grad_norm": 0.029732126742601395,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.7816314697265625,
+ "step": 577
+ },
+ {
+ "epoch": 8.027923211169284,
+ "grad_norm": 0.04357453063130379,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.7575225830078125,
+ "step": 578
+ },
+ {
+ "epoch": 8.041884816753926,
+ "grad_norm": 0.051903221756219864,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.802519798278809,
+ "step": 579
+ },
+ {
+ "epoch": 8.055846422338568,
+ "grad_norm": 0.0539076030254364,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.799004554748535,
+ "step": 580
+ },
+ {
+ "epoch": 8.06980802792321,
+ "grad_norm": 0.043642807751894,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.7896528244018555,
+ "step": 581
+ },
+ {
+ "epoch": 8.083769633507853,
+ "grad_norm": 0.0469554178416729,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.765430927276611,
+ "step": 582
+ },
+ {
+ "epoch": 8.097731239092496,
+ "grad_norm": 0.042060062289237976,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.784132957458496,
+ "step": 583
+ },
+ {
+ "epoch": 8.111692844677139,
+ "grad_norm": 0.035101912915706635,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.767234802246094,
+ "step": 584
+ },
+ {
+ "epoch": 8.12565445026178,
+ "grad_norm": 0.042576879262924194,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.762052536010742,
+ "step": 585
+ },
+ {
+ "epoch": 8.139616055846423,
+ "grad_norm": 0.04231419041752815,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.752150535583496,
+ "step": 586
+ },
+ {
+ "epoch": 8.153577661431065,
+ "grad_norm": 0.046912599354982376,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.771881103515625,
+ "step": 587
+ },
+ {
+ "epoch": 8.167539267015707,
+ "grad_norm": 0.051271144300699234,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.773284912109375,
+ "step": 588
+ },
+ {
+ "epoch": 8.181500872600349,
+ "grad_norm": 0.04655958339571953,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.769917964935303,
+ "step": 589
+ },
+ {
+ "epoch": 8.195462478184991,
+ "grad_norm": 0.04463687539100647,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.77938175201416,
+ "step": 590
+ },
+ {
+ "epoch": 8.209424083769633,
+ "grad_norm": 0.047959886491298676,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.785962104797363,
+ "step": 591
+ },
+ {
+ "epoch": 8.223385689354275,
+ "grad_norm": 0.04414986073970795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.786281585693359,
+ "step": 592
+ },
+ {
+ "epoch": 8.237347294938917,
+ "grad_norm": 0.044150348752737045,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.777070045471191,
+ "step": 593
+ },
+ {
+ "epoch": 8.25130890052356,
+ "grad_norm": 0.037002332508563995,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.77547550201416,
+ "step": 594
+ },
+ {
+ "epoch": 8.265270506108202,
+ "grad_norm": 0.03417367860674858,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.782101154327393,
+ "step": 595
+ },
+ {
+ "epoch": 8.279232111692846,
+ "grad_norm": 0.03069210797548294,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.745595455169678,
+ "step": 596
+ },
+ {
+ "epoch": 8.293193717277488,
+ "grad_norm": 0.03228718787431717,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.7663655281066895,
+ "step": 597
+ },
+ {
+ "epoch": 8.30715532286213,
+ "grad_norm": 0.028603695333003998,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.767274856567383,
+ "step": 598
+ },
+ {
+ "epoch": 8.321116928446772,
+ "grad_norm": 0.03098377026617527,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.7853684425354,
+ "step": 599
+ },
+ {
+ "epoch": 8.335078534031414,
+ "grad_norm": 0.029242439195513725,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.738029956817627,
+ "step": 600
+ },
+ {
+ "epoch": 8.349040139616056,
+ "grad_norm": 0.02751081995666027,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.76801872253418,
+ "step": 601
+ },
+ {
+ "epoch": 8.363001745200698,
+ "grad_norm": 0.02838125079870224,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.743884086608887,
+ "step": 602
+ },
+ {
+ "epoch": 8.37696335078534,
+ "grad_norm": 0.029203062877058983,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.746668815612793,
+ "step": 603
+ },
+ {
+ "epoch": 8.390924956369982,
+ "grad_norm": 0.02679944597184658,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.760764122009277,
+ "step": 604
+ },
+ {
+ "epoch": 8.404886561954624,
+ "grad_norm": 0.028955848887562752,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.7353668212890625,
+ "step": 605
+ },
+ {
+ "epoch": 8.418848167539267,
+ "grad_norm": 0.02880851738154888,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.766837120056152,
+ "step": 606
+ },
+ {
+ "epoch": 8.432809773123909,
+ "grad_norm": 0.02495497092604637,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.78959321975708,
+ "step": 607
+ },
+ {
+ "epoch": 8.44677137870855,
+ "grad_norm": 0.026064753532409668,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.736433029174805,
+ "step": 608
+ },
+ {
+ "epoch": 8.460732984293193,
+ "grad_norm": 0.027566388249397278,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.782382965087891,
+ "step": 609
+ },
+ {
+ "epoch": 8.474694589877837,
+ "grad_norm": 0.02967359498143196,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.735286712646484,
+ "step": 610
+ },
+ {
+ "epoch": 8.488656195462479,
+ "grad_norm": 0.03079676441848278,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.725196838378906,
+ "step": 611
+ },
+ {
+ "epoch": 8.502617801047121,
+ "grad_norm": 0.032531846314668655,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.719333648681641,
+ "step": 612
+ },
+ {
+ "epoch": 8.516579406631763,
+ "grad_norm": 0.03120352514088154,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.723845958709717,
+ "step": 613
+ },
+ {
+ "epoch": 8.530541012216405,
+ "grad_norm": 0.03328424692153931,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.7669782638549805,
+ "step": 614
+ },
+ {
+ "epoch": 8.544502617801047,
+ "grad_norm": 0.035601649433374405,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.714925765991211,
+ "step": 615
+ },
+ {
+ "epoch": 8.55846422338569,
+ "grad_norm": 0.03675796464085579,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.715317726135254,
+ "step": 616
+ },
+ {
+ "epoch": 8.572425828970331,
+ "grad_norm": 0.03701988235116005,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.724606990814209,
+ "step": 617
+ },
+ {
+ "epoch": 8.586387434554974,
+ "grad_norm": 0.036719705909490585,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.746044635772705,
+ "step": 618
+ },
+ {
+ "epoch": 8.600349040139616,
+ "grad_norm": 0.034756604582071304,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.74885368347168,
+ "step": 619
+ },
+ {
+ "epoch": 8.614310645724258,
+ "grad_norm": 0.03237922117114067,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.721858501434326,
+ "step": 620
+ },
+ {
+ "epoch": 8.6282722513089,
+ "grad_norm": 0.028420142829418182,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.733814239501953,
+ "step": 621
+ },
+ {
+ "epoch": 8.642233856893542,
+ "grad_norm": 0.026203038170933723,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.726088523864746,
+ "step": 622
+ },
+ {
+ "epoch": 8.656195462478184,
+ "grad_norm": 0.02627231739461422,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.727700233459473,
+ "step": 623
+ },
+ {
+ "epoch": 8.670157068062828,
+ "grad_norm": 0.02920261025428772,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.731454849243164,
+ "step": 624
+ },
+ {
+ "epoch": 8.68411867364747,
+ "grad_norm": 0.02792907878756523,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.734154224395752,
+ "step": 625
+ },
+ {
+ "epoch": 8.698080279232112,
+ "grad_norm": 0.026770325377583504,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.715337753295898,
+ "step": 626
+ },
+ {
+ "epoch": 8.712041884816754,
+ "grad_norm": 0.027346540242433548,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.716512680053711,
+ "step": 627
+ },
+ {
+ "epoch": 8.726003490401396,
+ "grad_norm": 0.031965773552656174,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.7072038650512695,
+ "step": 628
+ },
+ {
+ "epoch": 8.739965095986038,
+ "grad_norm": 0.03708025813102722,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.749785900115967,
+ "step": 629
+ },
+ {
+ "epoch": 8.75392670157068,
+ "grad_norm": 0.04846490919589996,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.735586166381836,
+ "step": 630
+ },
+ {
+ "epoch": 8.767888307155323,
+ "grad_norm": 0.053352758288383484,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.720524787902832,
+ "step": 631
+ },
+ {
+ "epoch": 8.781849912739965,
+ "grad_norm": 0.040559202432632446,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.719901084899902,
+ "step": 632
+ },
+ {
+ "epoch": 8.795811518324607,
+ "grad_norm": 0.03333032503724098,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.708139419555664,
+ "step": 633
+ },
+ {
+ "epoch": 8.809773123909249,
+ "grad_norm": 0.030532941222190857,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.728184223175049,
+ "step": 634
+ },
+ {
+ "epoch": 8.823734729493891,
+ "grad_norm": 0.03337649628520012,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.726020336151123,
+ "step": 635
+ },
+ {
+ "epoch": 8.837696335078533,
+ "grad_norm": 0.03352045267820358,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.7196855545043945,
+ "step": 636
+ },
+ {
+ "epoch": 8.851657940663177,
+ "grad_norm": 0.03662872314453125,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.732823371887207,
+ "step": 637
+ },
+ {
+ "epoch": 8.86561954624782,
+ "grad_norm": 0.037085745483636856,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.728443145751953,
+ "step": 638
+ },
+ {
+ "epoch": 8.879581151832461,
+ "grad_norm": 0.036635689437389374,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.713988780975342,
+ "step": 639
+ },
+ {
+ "epoch": 8.893542757417103,
+ "grad_norm": 0.03719659522175789,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.715950965881348,
+ "step": 640
+ },
+ {
+ "epoch": 8.907504363001745,
+ "grad_norm": 0.03719159588217735,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.7124924659729,
+ "step": 641
+ },
+ {
+ "epoch": 8.921465968586388,
+ "grad_norm": 0.04275721684098244,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.708404064178467,
+ "step": 642
+ },
+ {
+ "epoch": 8.93542757417103,
+ "grad_norm": 0.04204598814249039,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.715372085571289,
+ "step": 643
+ },
+ {
+ "epoch": 8.949389179755672,
+ "grad_norm": 0.03938259929418564,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.729745864868164,
+ "step": 644
+ },
+ {
+ "epoch": 8.963350785340314,
+ "grad_norm": 0.036226850003004074,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.733749866485596,
+ "step": 645
+ },
+ {
+ "epoch": 8.977312390924956,
+ "grad_norm": 0.03542874753475189,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.706825256347656,
+ "step": 646
+ },
+ {
+ "epoch": 8.991273996509598,
+ "grad_norm": 0.029055381193757057,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.718567848205566,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.01990380510687828,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 2.943183183670044,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.601699709892273,
+ "eval_runtime": 60.0221,
+ "eval_samples_per_second": 40.685,
+ "eval_steps_per_second": 0.65,
+ "step": 648
+ },
+ {
+ "epoch": 9.013961605584642,
+ "grad_norm": 0.030904127284884453,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.660327911376953,
+ "step": 649
+ },
+ {
+ "epoch": 9.027923211169284,
+ "grad_norm": 0.0441313236951828,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.701565742492676,
+ "step": 650
+ },
+ {
+ "epoch": 9.041884816753926,
+ "grad_norm": 0.047547727823257446,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.666440486907959,
+ "step": 651
+ },
+ {
+ "epoch": 9.055846422338568,
+ "grad_norm": 0.04549463838338852,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.660548210144043,
+ "step": 652
+ },
+ {
+ "epoch": 9.06980802792321,
+ "grad_norm": 0.050346389412879944,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.640229225158691,
+ "step": 653
+ },
+ {
+ "epoch": 9.083769633507853,
+ "grad_norm": 0.058505453169345856,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.673881530761719,
+ "step": 654
+ },
+ {
+ "epoch": 9.097731239092496,
+ "grad_norm": 0.051989227533340454,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.693943977355957,
+ "step": 655
+ },
+ {
+ "epoch": 9.111692844677139,
+ "grad_norm": 0.052564945071935654,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.644315242767334,
+ "step": 656
+ },
+ {
+ "epoch": 9.12565445026178,
+ "grad_norm": 0.05647841840982437,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.681768417358398,
+ "step": 657
+ },
+ {
+ "epoch": 9.139616055846423,
+ "grad_norm": 0.058640867471694946,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.676457405090332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153577661431065,
+ "grad_norm": 0.05386025458574295,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.659588813781738,
+ "step": 659
+ },
+ {
+ "epoch": 9.167539267015707,
+ "grad_norm": 0.048699021339416504,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.680031776428223,
+ "step": 660
+ },
+ {
+ "epoch": 9.181500872600349,
+ "grad_norm": 0.04374176263809204,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.672671318054199,
+ "step": 661
+ },
+ {
+ "epoch": 9.195462478184991,
+ "grad_norm": 0.042337436228990555,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.673431873321533,
+ "step": 662
+ },
+ {
+ "epoch": 9.209424083769633,
+ "grad_norm": 0.04564661160111427,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.632595062255859,
+ "step": 663
+ },
+ {
+ "epoch": 9.223385689354275,
+ "grad_norm": 0.04404835030436516,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.678891181945801,
+ "step": 664
+ },
+ {
+ "epoch": 9.237347294938917,
+ "grad_norm": 0.040642790496349335,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.633418560028076,
+ "step": 665
+ },
+ {
+ "epoch": 9.25130890052356,
+ "grad_norm": 0.036492060869932175,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.670109748840332,
+ "step": 666
+ },
+ {
+ "epoch": 9.265270506108202,
+ "grad_norm": 0.03735879808664322,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.608570098876953,
+ "step": 667
+ },
+ {
+ "epoch": 9.279232111692846,
+ "grad_norm": 0.034563641995191574,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.632505416870117,
+ "step": 668
+ },
+ {
+ "epoch": 9.293193717277488,
+ "grad_norm": 0.03522484749555588,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.6475629806518555,
+ "step": 669
+ },
+ {
+ "epoch": 9.30715532286213,
+ "grad_norm": 0.03216472268104553,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.627873420715332,
+ "step": 670
+ },
+ {
+ "epoch": 9.321116928446772,
+ "grad_norm": 0.029155759140849113,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.652706146240234,
+ "step": 671
+ },
+ {
+ "epoch": 9.335078534031414,
+ "grad_norm": 0.031060541048645973,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.644545078277588,
+ "step": 672
+ },
+ {
+ "epoch": 9.349040139616056,
+ "grad_norm": 0.034979186952114105,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.642645359039307,
+ "step": 673
+ },
+ {
+ "epoch": 9.363001745200698,
+ "grad_norm": 0.035258810967206955,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.647383689880371,
+ "step": 674
+ },
+ {
+ "epoch": 9.37696335078534,
+ "grad_norm": 0.031552575528621674,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.667779445648193,
+ "step": 675
+ },
+ {
+ "epoch": 9.390924956369982,
+ "grad_norm": 0.03193951025605202,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.63625431060791,
+ "step": 676
+ },
+ {
+ "epoch": 9.404886561954624,
+ "grad_norm": 0.029811711981892586,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.658000946044922,
+ "step": 677
+ },
+ {
+ "epoch": 9.418848167539267,
+ "grad_norm": 0.028464138507843018,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.6545515060424805,
+ "step": 678
+ },
+ {
+ "epoch": 9.432809773123909,
+ "grad_norm": 0.0315735749900341,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.641153335571289,
+ "step": 679
+ },
+ {
+ "epoch": 9.44677137870855,
+ "grad_norm": 0.033412106335163116,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.626931667327881,
+ "step": 680
+ },
+ {
+ "epoch": 9.460732984293193,
+ "grad_norm": 0.03757626190781593,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.671254634857178,
+ "step": 681
+ },
+ {
+ "epoch": 9.474694589877837,
+ "grad_norm": 0.03910793736577034,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.651954174041748,
+ "step": 682
+ },
+ {
+ "epoch": 9.488656195462479,
+ "grad_norm": 0.04015549272298813,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.613324165344238,
+ "step": 683
+ },
+ {
+ "epoch": 9.502617801047121,
+ "grad_norm": 0.040689144283533096,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.639243125915527,
+ "step": 684
+ },
+ {
+ "epoch": 9.516579406631763,
+ "grad_norm": 0.04075361415743828,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.649250030517578,
+ "step": 685
+ },
+ {
+ "epoch": 9.530541012216405,
+ "grad_norm": 0.03914607688784599,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.66444730758667,
+ "step": 686
+ },
+ {
+ "epoch": 9.544502617801047,
+ "grad_norm": 0.033253733068704605,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.647335052490234,
+ "step": 687
+ },
+ {
+ "epoch": 9.55846422338569,
+ "grad_norm": 0.03179008886218071,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.6727800369262695,
+ "step": 688
+ },
+ {
+ "epoch": 9.572425828970331,
+ "grad_norm": 0.03477025404572487,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.655778884887695,
+ "step": 689
+ },
+ {
+ "epoch": 9.586387434554974,
+ "grad_norm": 0.030094800516963005,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.680851936340332,
+ "step": 690
+ },
+ {
+ "epoch": 9.600349040139616,
+ "grad_norm": 0.030153103172779083,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.654638290405273,
+ "step": 691
+ },
+ {
+ "epoch": 9.614310645724258,
+ "grad_norm": 0.029009278863668442,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.625916481018066,
+ "step": 692
+ },
+ {
+ "epoch": 9.6282722513089,
+ "grad_norm": 0.027248440310359,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.656370162963867,
+ "step": 693
+ },
+ {
+ "epoch": 9.642233856893542,
+ "grad_norm": 0.027727462351322174,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.631473541259766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656195462478184,
+ "grad_norm": 0.026710307225584984,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.636524200439453,
+ "step": 695
+ },
+ {
+ "epoch": 9.670157068062828,
+ "grad_norm": 0.02726181596517563,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.629701614379883,
+ "step": 696
+ },
+ {
+ "epoch": 9.68411867364747,
+ "grad_norm": 0.027146652340888977,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.640240669250488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698080279232112,
+ "grad_norm": 0.027165887877345085,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.64462947845459,
+ "step": 698
+ },
+ {
+ "epoch": 9.712041884816754,
+ "grad_norm": 0.03025982715189457,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.6271820068359375,
+ "step": 699
+ },
+ {
+ "epoch": 9.726003490401396,
+ "grad_norm": 0.02837025746703148,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.629697799682617,
+ "step": 700
+ },
+ {
+ "epoch": 9.739965095986038,
+ "grad_norm": 0.026777995750308037,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.646146774291992,
+ "step": 701
+ },
+ {
+ "epoch": 9.75392670157068,
+ "grad_norm": 0.024626802653074265,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.641671657562256,
+ "step": 702
+ },
+ {
+ "epoch": 9.767888307155323,
+ "grad_norm": 0.025329895317554474,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.626345157623291,
+ "step": 703
+ },
+ {
+ "epoch": 9.781849912739965,
+ "grad_norm": 0.027804844081401825,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.621457099914551,
+ "step": 704
+ },
+ {
+ "epoch": 9.795811518324607,
+ "grad_norm": 0.028763437643647194,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.6124114990234375,
+ "step": 705
+ },
+ {
+ "epoch": 9.809773123909249,
+ "grad_norm": 0.030769916251301765,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.604399681091309,
+ "step": 706
+ },
+ {
+ "epoch": 9.823734729493891,
+ "grad_norm": 0.03060949593782425,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.640804290771484,
+ "step": 707
+ },
+ {
+ "epoch": 9.837696335078533,
+ "grad_norm": 0.029832778498530388,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.593376159667969,
+ "step": 708
+ },
+ {
+ "epoch": 9.851657940663177,
+ "grad_norm": 0.02952776849269867,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.619533538818359,
+ "step": 709
+ },
+ {
+ "epoch": 9.86561954624782,
+ "grad_norm": 0.02894212119281292,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.640516757965088,
+ "step": 710
+ },
+ {
+ "epoch": 9.879581151832461,
+ "grad_norm": 0.026435624808073044,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.6342997550964355,
+ "step": 711
+ },
+ {
+ "epoch": 9.893542757417103,
+ "grad_norm": 0.02591666206717491,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.630884170532227,
+ "step": 712
+ },
+ {
+ "epoch": 9.907504363001745,
+ "grad_norm": 0.02702309936285019,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.632085800170898,
+ "step": 713
+ },
+ {
+ "epoch": 9.921465968586388,
+ "grad_norm": 0.026766657829284668,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.6231231689453125,
+ "step": 714
+ },
+ {
+ "epoch": 9.93542757417103,
+ "grad_norm": 0.026666559278964996,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.601619243621826,
+ "step": 715
+ },
+ {
+ "epoch": 9.949389179755672,
+ "grad_norm": 0.026801064610481262,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.641724109649658,
+ "step": 716
+ },
+ {
+ "epoch": 9.963350785340314,
+ "grad_norm": 0.03152798116207123,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.642511367797852,
+ "step": 717
+ },
+ {
+ "epoch": 9.977312390924956,
+ "grad_norm": 0.035406943410634995,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.635779857635498,
+ "step": 718
+ },
+ {
+ "epoch": 9.991273996509598,
+ "grad_norm": 0.037504684180021286,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.629463195800781,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02263174019753933,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 2.876767635345459,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.5942904949188232,
+ "eval_runtime": 60.2086,
+ "eval_samples_per_second": 40.559,
+ "eval_steps_per_second": 0.648,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03707852587103844,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.5765533447265625,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.05454680323600769,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.576363563537598,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.04832114651799202,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.586915969848633,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.050469402223825455,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.5265092849731445,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0622565895318985,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.592833518981934,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0640949010848999,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.5855560302734375,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.06502196192741394,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.591236114501953,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.06167149916291237,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.56747579574585,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.06747374683618546,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.60452938079834,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.07210271060466766,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.561588764190674,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.0733448714017868,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.610766410827637,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.0731528028845787,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.589570045471191,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.0822535902261734,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.631077289581299,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.06921929866075516,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.628293991088867,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.07443586736917496,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.627321243286133,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.07153813540935516,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.641068935394287,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.069196417927742,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.615875720977783,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06483697891235352,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.6165313720703125,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.05482150986790657,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.599890232086182,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.05364019051194191,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.61989164352417,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.053216829895973206,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.564305305480957,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05740506947040558,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.582683563232422,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.05587657541036606,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.61264705657959,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.050252825021743774,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.648050308227539,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.050955042243003845,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.615367889404297,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04903746768832207,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.617378234863281,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04252410680055618,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.609130859375,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.03801053389906883,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.625149250030518,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.03422372043132782,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.606106281280518,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.030507879331707954,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.6125993728637695,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.0340375155210495,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.5690202713012695,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.030507374554872513,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.6042985916137695,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027383938431739807,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.585570812225342,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.028792303055524826,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.566786766052246,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02866457775235176,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.566376686096191,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.028657952323555946,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.581750869750977,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.028227943927049637,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.576788902282715,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.024757765233516693,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.59373664855957,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.027442635968327522,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.570978164672852,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.02590034157037735,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.571286201477051,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.025696570053696632,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.542209625244141,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.025478003546595573,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.5900421142578125,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.02563696913421154,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.544609546661377,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.025326019152998924,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.5808515548706055,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.024502627551555634,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.560579776763916,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.02687518671154976,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.567154884338379,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.025280235335230827,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.560011863708496,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.026266315951943398,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.5758442878723145,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.026058891788125038,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.583309173583984,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.02620471641421318,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.547737121582031,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02626362442970276,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.556955337524414,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.024503158405423164,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.580735206604004,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02580532431602478,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.564000129699707,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.024437161162495613,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.545594215393066,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.026664750650525093,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.56415319442749,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.02499818615615368,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.550689697265625,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.027514580637216568,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.5548248291015625,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.026799403131008148,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.548295021057129,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.02685476280748844,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.5660600662231445,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.026104001328349113,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.556082248687744,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.027479661628603935,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.5580244064331055,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.02638247236609459,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.550073623657227,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.025853842496871948,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.578088760375977,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.02636682614684105,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.58807373046875,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.02684398740530014,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.561017036437988,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.028538964688777924,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.548762321472168,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.028327809646725655,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.548969745635986,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.025131452828645706,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.564047813415527,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.026498543098568916,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.553997039794922,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.030081963166594505,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.5787787437438965,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.03201213851571083,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.566424369812012,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.020518779754638672,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8421616554260254,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.5899484157562256,
+ "eval_runtime": 60.2671,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03692612797021866,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.492715835571289,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.050258126109838486,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.480323791503906,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.04179152473807335,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.484371185302734,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.04160204157233238,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.4905524253845215,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.03974860534071922,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.503803253173828,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.037031810730695724,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.494629859924316,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.038762833923101425,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.495111465454102,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.0412563718855381,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.490113735198975,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.041371624916791916,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.490643501281738,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04151557758450508,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.49881649017334,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04037577286362648,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.492178916931152,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.03676290065050125,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.477460861206055,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.038055796176195145,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.509150981903076,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.041152238845825195,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.47133731842041,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.03728092834353447,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.508072376251221,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.035338856279850006,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.505619049072266,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.037813492119312286,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.509489059448242,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.03789111226797104,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.508646488189697,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.03880828619003296,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.52567720413208,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.03611127287149429,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.498889446258545,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.036903638392686844,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.498601913452148,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03620576485991478,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.490840911865234,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03555903211236,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.506166458129883,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.03441924974322319,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.498001575469971,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.0334339402616024,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.491082191467285,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.031418606638908386,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.527408599853516,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03073921985924244,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.514107704162598,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.029589885845780373,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.48323917388916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.02974501997232437,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.51338529586792,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.029247552156448364,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.504205226898193,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.03132032975554466,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.505674362182617,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03088037669658661,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.477198600769043,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.031054604798555374,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.498721122741699,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03293661028146744,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.527826309204102,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.03181406110525131,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.495999336242676,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.029380083084106445,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.530351638793945,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.029296347871422768,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.492367744445801,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.02958036959171295,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.493037700653076,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03078395500779152,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.47504186630249,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03280661255121231,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.518409729003906,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.03323954716324806,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.460753917694092,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.032254140824079514,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.474285125732422,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.03018898330628872,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.5162353515625,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.02802245132625103,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.538684844970703,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.02821732498705387,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.4850287437438965,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.029189350083470345,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.50770378112793,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.02851196378469467,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.499988555908203,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.02741030789911747,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.499852180480957,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.028708625584840775,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.465211868286133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.028519680723547935,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.502464294433594,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.027846841141581535,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.503248691558838,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030456233769655228,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.490391731262207,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029952052980661392,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.50056266784668,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.02925778552889824,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.49215030670166,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.029679695144295692,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.523841857910156,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.031178541481494904,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.499871730804443,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.031264498829841614,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.50546932220459,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03482109308242798,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.494480133056641,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.03341105580329895,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.532899856567383,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028623664751648903,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.5016279220581055,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028600146993994713,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.511589050292969,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.029127901419997215,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.502058982849121,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.027898237109184265,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.493182182312012,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.028683282434940338,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.519793510437012,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.02887023612856865,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.457289695739746,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.02811296097934246,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.49398946762085,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.028485111892223358,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.478344440460205,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02914245054125786,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.487666130065918,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.028030497953295708,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.518488883972168,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.027973609045147896,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.515271186828613,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.027458522468805313,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.490856647491455,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.01909293793141842,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.822120189666748,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5866912603378296,
+ "eval_runtime": 61.3185,
+ "eval_samples_per_second": 39.825,
+ "eval_steps_per_second": 0.636,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03683992475271225,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.438363075256348,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.05042923241853714,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.443560600280762,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.055914171040058136,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.422282695770264,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.057007886469364166,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.442644119262695,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.05999710038304329,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.44780969619751,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06472347676753998,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.485074996948242,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.06923902779817581,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.457515716552734,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.0710320770740509,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.465476989746094,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.06942574679851532,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.426226615905762,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.06867868453264236,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.438092231750488,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.0688796266913414,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.446664810180664,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.0690719485282898,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.456482887268066,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.06068912148475647,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.4730730056762695,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.05411486327648163,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.4596452713012695,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.056623928248882294,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.418197154998779,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.05634097009897232,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.482892036437988,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.05056057870388031,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.457589149475098,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.04460986703634262,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.451439380645752,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.045022401958703995,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.452855110168457,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.04054102674126625,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.417323112487793,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.03682882338762283,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.465713977813721,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.03528819605708122,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.4502339363098145,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.03568802773952484,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.422399520874023,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.0363088883459568,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.447678089141846,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.03460463136434555,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.437664985656738,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.034751344472169876,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.489151954650879,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.033658601343631744,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.4471635818481445,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.03433239087462425,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.455332279205322,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.032074104994535446,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.435157775878906,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.032168325036764145,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.435118675231934,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.03007678873836994,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.42781925201416,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03094375506043434,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.433105945587158,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.032186444848775864,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.383000373840332,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.030610304325819016,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.473257064819336,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.030257435515522957,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.453041076660156,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.0301323514431715,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.396578788757324,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.030934512615203857,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.445223808288574,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.029907869175076485,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.442287445068359,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03163841366767883,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.434650421142578,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.029627414420247078,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.454374313354492,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.031026875600218773,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.413784027099609,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03099491260945797,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.458438873291016,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.02981731668114662,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.438560962677002,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.02927570603787899,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.44493293762207,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.029779717326164246,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.445121765136719,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.02909068390727043,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.4736127853393555,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.029217200353741646,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.434604644775391,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.029872911050915718,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.454591751098633,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.028742222115397453,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.468222141265869,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.030313383787870407,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.45759391784668,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.03014662116765976,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.446331977844238,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029481127858161926,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.478209495544434,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.028485266491770744,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.444445610046387,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.02857700176537037,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.430604934692383,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.029038602486252785,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.43759298324585,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.029576119035482407,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.474579811096191,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.02897879108786583,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.440483093261719,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.029799485579133034,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.4839277267456055,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.030710609629750252,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.466696739196777,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.030270954594016075,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.455197334289551,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.03170182928442955,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.454664707183838,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030684730038046837,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.437619209289551,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.030972573906183243,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.431796073913574,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028914963826537132,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.460618019104004,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.029810264706611633,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.407312393188477,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.029806939885020256,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.466520309448242,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.03020886518061161,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.427349090576172,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030598826706409454,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.427891731262207,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03192012757062912,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.448113918304443,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.030492989346385002,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.450060844421387,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03188254311680794,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.43007230758667,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02081947773694992,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.782536506652832,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5860040187835693,
+ "eval_runtime": 60.6725,
+ "eval_samples_per_second": 40.249,
+ "eval_steps_per_second": 0.643,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.036721356213092804,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.358441352844238,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.056633755564689636,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.361527919769287,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.06175022944808006,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.3827924728393555,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05159056559205055,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.354856491088867,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.04885871335864067,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.359966278076172,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.04407693073153496,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.364612579345703,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.04891050234436989,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.344021320343018,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.04643657058477402,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.366066932678223,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.05241481214761734,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.3390607833862305,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.05837438255548477,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.390261650085449,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.058482322841882706,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.3856201171875,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.06507553160190582,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.414223670959473,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.0575365275144577,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.402576446533203,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05055592209100723,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.389563083648682,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.047167956829071045,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.402750015258789,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.04473405331373215,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.358315467834473,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.042440690100193024,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.392551898956299,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04465719312429428,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.373558044433594,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.04337528720498085,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.37944221496582,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.04322915896773338,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.38493013381958,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.039873406291007996,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.389793395996094,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.0413178876042366,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.427964210510254,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.04567280039191246,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.404172897338867,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04154926910996437,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.391359329223633,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.03774876892566681,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.381457328796387,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03588850423693657,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.392526626586914,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.03757261857390404,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.39273738861084,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03849692642688751,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.409083366394043,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03787294402718544,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.3776044845581055,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.034496866166591644,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.413462162017822,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.03973205015063286,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.359318733215332,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.036192040890455246,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.392844200134277,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.03436632081866264,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.421384811401367,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.035658642649650574,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.386757850646973,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.03434896469116211,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.364978313446045,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03642290085554123,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.385326385498047,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.03427736461162567,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.406352996826172,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.034845270216464996,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.370953559875488,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03390975296497345,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.3799943923950195,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.03134961053729057,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.419156551361084,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.03512260690331459,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.404660701751709,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03319067507982254,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.388535022735596,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03113507106900215,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.403522968292236,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03559132665395737,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.382323265075684,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.03579750657081604,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.37620210647583,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.033675212413072586,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.411443710327148,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.03220067173242569,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.443470001220703,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031610481441020966,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.406038284301758,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.02958764135837555,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.41604471206665,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.0318237729370594,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4253387451171875,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03260074555873871,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.370792388916016,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.03359333053231239,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.397639274597168,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032516930252313614,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.4140777587890625,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.03203396871685982,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.408872604370117,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.03201935440301895,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.402975082397461,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.03330211713910103,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.394018173217773,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.0348181314766407,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.380587100982666,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.03174913674592972,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.372378349304199,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.029864462092518806,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.42018985748291,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.033312153071165085,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.385973930358887,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03203144669532776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.395167827606201,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.031578030437231064,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.3982062339782715,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.030885562300682068,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.400731086730957,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03235667943954468,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.389830589294434,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.032664328813552856,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.401277542114258,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.030354196205735207,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.422520637512207,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.031162403523921967,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.393032073974609,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.032187070697546005,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.464164733886719,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.03445067256689072,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.368991851806641,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03398612141609192,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.418752670288086,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.029730670154094696,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.41429328918457,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022440090775489807,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7315595149993896,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5853753089904785,
+ "eval_runtime": 60.7174,
+ "eval_samples_per_second": 40.219,
+ "eval_steps_per_second": 0.642,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.03754177689552307,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.304028511047363,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.051616180688142776,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.34659481048584,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055258627980947495,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.310445785522461,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.049489911645650864,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3416619300842285,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05377533286809921,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.3143310546875,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.05443088337779045,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.320600986480713,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.050222691148519516,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.334789752960205,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.051904670894145966,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.316756248474121,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.05376426503062248,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.274116516113281,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.053313180804252625,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.321177005767822,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.058234598487615585,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.3169264793396,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05730101093649864,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.307013511657715,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.058140840381383896,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.332961082458496,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05654815956950188,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.315462112426758,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.05129384249448776,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.338476181030273,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.047652702778577805,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.321232795715332,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.050912003964185715,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.318971633911133,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.0514022521674633,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.3334150314331055,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.05457981303334236,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.325802803039551,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04678885266184807,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.307281494140625,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.05048161372542381,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.299831390380859,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.05334261804819107,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.313231468200684,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.05055771768093109,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.36859130859375,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04955561086535454,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.301530361175537,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.04548008739948273,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.339517593383789,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.046585116535425186,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.333036422729492,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.0503716766834259,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.364710807800293,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.040569406002759933,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.369089126586914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.03754054754972458,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.3349432945251465,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03775954991579056,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.333134174346924,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.04066266492009163,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.354358673095703,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.040490977466106415,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.360981464385986,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.0426657609641552,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.357726097106934,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.0415605828166008,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.363457679748535,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.04020509123802185,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.372447967529297,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.04101516678929329,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.292452812194824,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.03546427935361862,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.33892297744751,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.03661105036735535,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.329778671264648,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03565476834774017,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.351012229919434,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.03591311350464821,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.376980781555176,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.03665630891919136,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.354767799377441,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.0371236577630043,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.374916076660156,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.03464849665760994,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.3713059425354,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03345152735710144,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.350670337677002,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.03357568010687828,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.3573713302612305,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03503486514091492,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.362038612365723,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.03455005958676338,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.3263773918151855,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.03173859417438507,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.351345062255859,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03301836550235748,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.398364067077637,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.033733855932950974,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.335268020629883,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034080155193805695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.371492385864258,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.034396376460790634,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.382076740264893,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.03303401917219162,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.36623477935791,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03297369182109833,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3148393630981445,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.034819211810827255,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.358110427856445,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.033389780670404434,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.367427825927734,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.032363444566726685,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.3813676834106445,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03509577363729477,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.386520862579346,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.034685730934143066,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.33738374710083,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.034755002707242966,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.360980987548828,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03461688384413719,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.396650314331055,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03392560034990311,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.356130123138428,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.03342147544026375,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.359145164489746,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.032852694392204285,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.3796162605285645,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03472859039902687,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.361110687255859,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03277098760008812,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.361811637878418,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.03341882303357124,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.369524955749512,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.032347049564123154,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.369096755981445,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03161193057894707,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.339844703674316,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.03322301432490349,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.376675605773926,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03419092297554016,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.346864223480225,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02461826801300049,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7236275672912598,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.5859372019767761,
+ "eval_runtime": 59.0139,
+ "eval_samples_per_second": 41.38,
+ "eval_steps_per_second": 0.661,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013961605584642,
+ "grad_norm": 0.039559684693813324,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 4.289841651916504,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027923211169284,
+ "grad_norm": 0.05554218962788582,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 4.255925178527832,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041884816753926,
+ "grad_norm": 0.05330996215343475,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 4.272125244140625,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055846422338568,
+ "grad_norm": 0.053247954696416855,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 4.290765285491943,
+ "step": 1084
+ },
+ {
+ "epoch": 15.06980802792321,
+ "grad_norm": 0.05471150949597359,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 4.271744728088379,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083769633507853,
+ "grad_norm": 0.05976719409227371,
+ "learning_rate": 0.000491285979372924,
+ "loss": 4.28203821182251,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097731239092496,
+ "grad_norm": 0.05369880795478821,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 4.275313377380371,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111692844677139,
+ "grad_norm": 0.052529267966747284,
+ "learning_rate": 0.000490873338088198,
+ "loss": 4.25678825378418,
+ "step": 1088
+ },
+ {
+ "epoch": 15.12565445026178,
+ "grad_norm": 0.05184522643685341,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 4.260510444641113,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139616055846423,
+ "grad_norm": 0.051606856286525726,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 4.283233642578125,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153577661431065,
+ "grad_norm": 0.05318903923034668,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 4.286595344543457,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167539267015707,
+ "grad_norm": 0.058525051921606064,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 4.282374382019043,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181500872600349,
+ "grad_norm": 0.054776228964328766,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 4.2913126945495605,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195462478184991,
+ "grad_norm": 0.04907633364200592,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 4.297187805175781,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209424083769633,
+ "grad_norm": 0.05163136497139931,
+ "learning_rate": 0.000489424334303338,
+ "loss": 4.2968645095825195,
+ "step": 1095
+ },
+ {
+ "epoch": 15.223385689354275,
+ "grad_norm": 0.053157489746809006,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 4.285839080810547,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237347294938917,
+ "grad_norm": 0.05251535773277283,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 4.311246871948242,
+ "step": 1097
+ },
+ {
+ "epoch": 15.25130890052356,
+ "grad_norm": 0.048785097897052765,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 4.300364971160889,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265270506108202,
+ "grad_norm": 0.050393108278512955,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 4.299890995025635,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279232111692846,
+ "grad_norm": 0.04894037917256355,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 4.312449932098389,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293193717277488,
+ "grad_norm": 0.05274952948093414,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 4.2889814376831055,
+ "step": 1101
+ },
+ {
+ "epoch": 15.30715532286213,
+ "grad_norm": 0.05946570262312889,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 4.2731170654296875,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321116928446772,
+ "grad_norm": 0.054245397448539734,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 4.272463798522949,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335078534031414,
+ "grad_norm": 0.042092181742191315,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 4.299953937530518,
+ "step": 1104
+ },
+ {
+ "epoch": 15.349040139616056,
+ "grad_norm": 0.040222618728876114,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 4.292524814605713,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363001745200698,
+ "grad_norm": 0.04298696666955948,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 4.31342887878418,
+ "step": 1106
+ },
+ {
+ "epoch": 15.37696335078534,
+ "grad_norm": 0.04271050915122032,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 4.277663707733154,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390924956369982,
+ "grad_norm": 0.042364370077848434,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 4.311102390289307,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404886561954624,
+ "grad_norm": 0.04343319311738014,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 4.311610221862793,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418848167539267,
+ "grad_norm": 0.041321512311697006,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 4.29677677154541,
+ "step": 1110
+ },
+ {
+ "epoch": 15.432809773123909,
+ "grad_norm": 0.04107942432165146,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 4.3163018226623535,
+ "step": 1111
+ },
+ {
+ "epoch": 15.44677137870855,
+ "grad_norm": 0.03844515234231949,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 4.311301231384277,
+ "step": 1112
+ },
+ {
+ "epoch": 15.460732984293193,
+ "grad_norm": 0.03731007128953934,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 4.31325626373291,
+ "step": 1113
+ },
+ {
+ "epoch": 15.474694589877837,
+ "grad_norm": 0.03807826340198517,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 4.296201705932617,
+ "step": 1114
+ },
+ {
+ "epoch": 15.488656195462479,
+ "grad_norm": 0.040137600153684616,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 4.320395469665527,
+ "step": 1115
+ },
+ {
+ "epoch": 15.502617801047121,
+ "grad_norm": 0.038484394550323486,
+ "learning_rate": 0.000485033394934934,
+ "loss": 4.313858985900879,
+ "step": 1116
+ },
+ {
+ "epoch": 15.516579406631763,
+ "grad_norm": 0.03603338822722435,
+ "learning_rate": 0.000484822676912974,
+ "loss": 4.297666549682617,
+ "step": 1117
+ },
+ {
+ "epoch": 15.530541012216405,
+ "grad_norm": 0.038647182285785675,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 4.305002689361572,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544502617801047,
+ "grad_norm": 0.03907139599323273,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 4.329583644866943,
+ "step": 1119
+ },
+ {
+ "epoch": 15.55846422338569,
+ "grad_norm": 0.03621622920036316,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 4.290896415710449,
+ "step": 1120
+ },
+ {
+ "epoch": 15.572425828970331,
+ "grad_norm": 0.037496887147426605,
+ "learning_rate": 0.000483978341612154,
+ "loss": 4.33336877822876,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586387434554974,
+ "grad_norm": 0.03778311610221863,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 4.285354137420654,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600349040139616,
+ "grad_norm": 0.03782997280359268,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 4.29697847366333,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614310645724258,
+ "grad_norm": 0.03665657714009285,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 4.320103168487549,
+ "step": 1124
+ },
+ {
+ "epoch": 15.6282722513089,
+ "grad_norm": 0.03796127438545227,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 4.317153453826904,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642233856893542,
+ "grad_norm": 0.036425214260816574,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 4.293642520904541,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656195462478184,
+ "grad_norm": 0.03629784658551216,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 4.291100025177002,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670157068062828,
+ "grad_norm": 0.03674762696027756,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 4.295144557952881,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68411867364747,
+ "grad_norm": 0.036291178315877914,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 4.313133239746094,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698080279232112,
+ "grad_norm": 0.03817860782146454,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 4.294533729553223,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712041884816754,
+ "grad_norm": 0.036819685250520706,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 4.299263954162598,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726003490401396,
+ "grad_norm": 0.035103823989629745,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 4.287391185760498,
+ "step": 1132
+ },
+ {
+ "epoch": 15.739965095986038,
+ "grad_norm": 0.035701096057891846,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 4.318216323852539,
+ "step": 1133
+ },
+ {
+ "epoch": 15.75392670157068,
+ "grad_norm": 0.03639856353402138,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 4.301465034484863,
+ "step": 1134
+ },
+ {
+ "epoch": 15.767888307155323,
+ "grad_norm": 0.03593013063073158,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 4.311532974243164,
+ "step": 1135
+ },
+ {
+ "epoch": 15.781849912739965,
+ "grad_norm": 0.03564770147204399,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 4.3400726318359375,
+ "step": 1136
+ },
+ {
+ "epoch": 15.795811518324607,
+ "grad_norm": 0.03598494827747345,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 4.30976676940918,
+ "step": 1137
+ },
+ {
+ "epoch": 15.809773123909249,
+ "grad_norm": 0.034391652792692184,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 4.328464508056641,
+ "step": 1138
+ },
+ {
+ "epoch": 15.823734729493891,
+ "grad_norm": 0.03445672616362572,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 4.310579299926758,
+ "step": 1139
+ },
+ {
+ "epoch": 15.837696335078533,
+ "grad_norm": 0.03643282502889633,
+ "learning_rate": 0.000479936075927473,
+ "loss": 4.320905685424805,
+ "step": 1140
+ },
+ {
+ "epoch": 15.851657940663177,
+ "grad_norm": 0.03489850088953972,
+ "learning_rate": 0.000479721889242305,
+ "loss": 4.317229270935059,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86561954624782,
+ "grad_norm": 0.03609171137213707,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 4.309203147888184,
+ "step": 1142
+ },
+ {
+ "epoch": 15.879581151832461,
+ "grad_norm": 0.03705338388681412,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 4.287387371063232,
+ "step": 1143
+ },
+ {
+ "epoch": 15.893542757417103,
+ "grad_norm": 0.03697717934846878,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 4.302860260009766,
+ "step": 1144
+ },
+ {
+ "epoch": 15.907504363001745,
+ "grad_norm": 0.035338208079338074,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 4.3152265548706055,
+ "step": 1145
+ },
+ {
+ "epoch": 15.921465968586388,
+ "grad_norm": 0.03548634052276611,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 4.327335357666016,
+ "step": 1146
+ },
+ {
+ "epoch": 15.93542757417103,
+ "grad_norm": 0.035512328147888184,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 4.3067803382873535,
+ "step": 1147
+ },
+ {
+ "epoch": 15.949389179755672,
+ "grad_norm": 0.03799289092421532,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 4.320701599121094,
+ "step": 1148
+ },
+ {
+ "epoch": 15.963350785340314,
+ "grad_norm": 0.03866339474916458,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 4.321087837219238,
+ "step": 1149
+ },
+ {
+ "epoch": 15.977312390924956,
+ "grad_norm": 0.036002084612846375,
+ "learning_rate": 0.000477787822648007,
+ "loss": 4.287045478820801,
+ "step": 1150
+ },
+ {
+ "epoch": 15.991273996509598,
+ "grad_norm": 0.03433441370725632,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 4.2562665939331055,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.02580358274281025,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 2.67783260345459,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.5870338082313538,
+ "eval_runtime": 60.0469,
+ "eval_samples_per_second": 40.668,
+ "eval_steps_per_second": 0.649,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013961605584644,
+ "grad_norm": 0.04105810821056366,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 4.224295616149902,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027923211169284,
+ "grad_norm": 0.049538832157850266,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 4.222184181213379,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041884816753928,
+ "grad_norm": 0.05089569091796875,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 4.2067060470581055,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05584642233857,
+ "grad_norm": 0.04600589722394943,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 4.238527774810791,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069808027923212,
+ "grad_norm": 0.043311409652233124,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 4.242960453033447,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083769633507853,
+ "grad_norm": 0.047897014766931534,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 4.199069023132324,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097731239092496,
+ "grad_norm": 0.05633604899048805,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 4.211665153503418,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111692844677137,
+ "grad_norm": 0.056110311299562454,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 4.20163631439209,
+ "step": 1160
+ },
+ {
+ "epoch": 16.12565445026178,
+ "grad_norm": 0.05426819995045662,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 4.21889591217041,
+ "step": 1161
+ },
+ {
+ "epoch": 16.13961605584642,
+ "grad_norm": 0.05489835515618324,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 4.220406532287598,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153577661431065,
+ "grad_norm": 0.053941354155540466,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 4.231909275054932,
+ "step": 1163
+ },
+ {
+ "epoch": 16.167539267015705,
+ "grad_norm": 0.058548808097839355,
+ "learning_rate": 0.000474756648870413,
+ "loss": 4.237003326416016,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18150087260035,
+ "grad_norm": 0.05370597913861275,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 4.240085124969482,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195462478184993,
+ "grad_norm": 0.05684991553425789,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 4.203842639923096,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209424083769633,
+ "grad_norm": 0.060745254158973694,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 4.247998237609863,
+ "step": 1167
+ },
+ {
+ "epoch": 16.223385689354277,
+ "grad_norm": 0.058296725153923035,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 4.226161003112793,
+ "step": 1168
+ },
+ {
+ "epoch": 16.237347294938917,
+ "grad_norm": 0.05310584232211113,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 4.229870796203613,
+ "step": 1169
+ },
+ {
+ "epoch": 16.25130890052356,
+ "grad_norm": 0.04710390046238899,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 4.240084648132324,
+ "step": 1170
+ },
+ {
+ "epoch": 16.2652705061082,
+ "grad_norm": 0.05003240332007408,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 4.240972518920898,
+ "step": 1171
+ },
+ {
+ "epoch": 16.279232111692846,
+ "grad_norm": 0.05552851781249046,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 4.248024940490723,
+ "step": 1172
+ },
+ {
+ "epoch": 16.293193717277486,
+ "grad_norm": 0.049967192113399506,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 4.279229164123535,
+ "step": 1173
+ },
+ {
+ "epoch": 16.30715532286213,
+ "grad_norm": 0.04883681237697601,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 4.269807815551758,
+ "step": 1174
+ },
+ {
+ "epoch": 16.32111692844677,
+ "grad_norm": 0.04931480810046196,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 4.232419490814209,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335078534031414,
+ "grad_norm": 0.048746511340141296,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 4.267911911010742,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349040139616054,
+ "grad_norm": 0.04678189381957054,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 4.262979507446289,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363001745200698,
+ "grad_norm": 0.048893529921770096,
+ "learning_rate": 0.000471698352726896,
+ "loss": 4.254408359527588,
+ "step": 1178
+ },
+ {
+ "epoch": 16.376963350785342,
+ "grad_norm": 0.045027732849121094,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 4.2445902824401855,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390924956369982,
+ "grad_norm": 0.04324870929121971,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 4.2685136795043945,
+ "step": 1180
+ },
+ {
+ "epoch": 16.404886561954626,
+ "grad_norm": 0.0470522940158844,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 4.237505912780762,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418848167539267,
+ "grad_norm": 0.04896962642669678,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 4.2523345947265625,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43280977312391,
+ "grad_norm": 0.04562787711620331,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 4.2192792892456055,
+ "step": 1183
+ },
+ {
+ "epoch": 16.44677137870855,
+ "grad_norm": 0.04414575919508934,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 4.250238418579102,
+ "step": 1184
+ },
+ {
+ "epoch": 16.460732984293195,
+ "grad_norm": 0.04363945871591568,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 4.256250381469727,
+ "step": 1185
+ },
+ {
+ "epoch": 16.474694589877835,
+ "grad_norm": 0.04159797728061676,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 4.274903297424316,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48865619546248,
+ "grad_norm": 0.04248544201254845,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 4.278222560882568,
+ "step": 1187
+ },
+ {
+ "epoch": 16.50261780104712,
+ "grad_norm": 0.04379890486598015,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 4.267505645751953,
+ "step": 1188
+ },
+ {
+ "epoch": 16.516579406631763,
+ "grad_norm": 0.04319411888718605,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 4.2607269287109375,
+ "step": 1189
+ },
+ {
+ "epoch": 16.530541012216403,
+ "grad_norm": 0.04143810644745827,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 4.267977714538574,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544502617801047,
+ "grad_norm": 0.03914840891957283,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 4.260498523712158,
+ "step": 1191
+ },
+ {
+ "epoch": 16.55846422338569,
+ "grad_norm": 0.04107543081045151,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 4.2550482749938965,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57242582897033,
+ "grad_norm": 0.041316065937280655,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 4.3075456619262695,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586387434554975,
+ "grad_norm": 0.04069342464208603,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 4.234392166137695,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600349040139616,
+ "grad_norm": 0.03816857561469078,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 4.251038551330566,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61431064572426,
+ "grad_norm": 0.04095374792814255,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 4.284790992736816,
+ "step": 1196
+ },
+ {
+ "epoch": 16.6282722513089,
+ "grad_norm": 0.03788160905241966,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 4.255716323852539,
+ "step": 1197
+ },
+ {
+ "epoch": 16.642233856893544,
+ "grad_norm": 0.03784061595797539,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 4.269606590270996,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656195462478184,
+ "grad_norm": 0.04087168723344803,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 4.231165885925293,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670157068062828,
+ "grad_norm": 0.040546853095293045,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 4.288422584533691,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68411867364747,
+ "grad_norm": 0.03777262568473816,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 4.277492523193359,
+ "step": 1201
+ },
+ {
+ "epoch": 16.698080279232112,
+ "grad_norm": 0.039857733994722366,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 4.239383697509766,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712041884816752,
+ "grad_norm": 0.04387792944908142,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 4.300265312194824,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726003490401396,
+ "grad_norm": 0.04214833676815033,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 4.255545616149902,
+ "step": 1204
+ },
+ {
+ "epoch": 16.739965095986037,
+ "grad_norm": 0.04290056228637695,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 4.275530815124512,
+ "step": 1205
+ },
+ {
+ "epoch": 16.75392670157068,
+ "grad_norm": 0.03975554555654526,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 4.282685279846191,
+ "step": 1206
+ },
+ {
+ "epoch": 16.767888307155324,
+ "grad_norm": 0.039464011788368225,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 4.285504341125488,
+ "step": 1207
+ },
+ {
+ "epoch": 16.781849912739965,
+ "grad_norm": 0.039714012295007706,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 4.273477554321289,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79581151832461,
+ "grad_norm": 0.03936111181974411,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 4.304205894470215,
+ "step": 1209
+ },
+ {
+ "epoch": 16.80977312390925,
+ "grad_norm": 0.03782064467668533,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 4.307382583618164,
+ "step": 1210
+ },
+ {
+ "epoch": 16.823734729493893,
+ "grad_norm": 0.040340717881917953,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 4.308620452880859,
+ "step": 1211
+ },
+ {
+ "epoch": 16.837696335078533,
+ "grad_norm": 0.0403442308306694,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 4.248626708984375,
+ "step": 1212
+ },
+ {
+ "epoch": 16.851657940663177,
+ "grad_norm": 0.038351960480213165,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 4.261248588562012,
+ "step": 1213
+ },
+ {
+ "epoch": 16.865619546247817,
+ "grad_norm": 0.03965333476662636,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 4.278291702270508,
+ "step": 1214
+ },
+ {
+ "epoch": 16.87958115183246,
+ "grad_norm": 0.040909092873334885,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 4.241847991943359,
+ "step": 1215
+ },
+ {
+ "epoch": 16.8935427574171,
+ "grad_norm": 0.038684047758579254,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 4.281346321105957,
+ "step": 1216
+ },
+ {
+ "epoch": 16.907504363001745,
+ "grad_norm": 0.03698103874921799,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 4.3085432052612305,
+ "step": 1217
+ },
+ {
+ "epoch": 16.921465968586386,
+ "grad_norm": 0.03908613696694374,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 4.24220085144043,
+ "step": 1218
+ },
+ {
+ "epoch": 16.93542757417103,
+ "grad_norm": 0.03873388469219208,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 4.282486915588379,
+ "step": 1219
+ },
+ {
+ "epoch": 16.949389179755673,
+ "grad_norm": 0.03709883987903595,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 4.279009819030762,
+ "step": 1220
+ },
+ {
+ "epoch": 16.963350785340314,
+ "grad_norm": 0.03937513753771782,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 4.273211479187012,
+ "step": 1221
+ },
+ {
+ "epoch": 16.977312390924958,
+ "grad_norm": 0.038478728383779526,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 4.281154155731201,
+ "step": 1222
+ },
+ {
+ "epoch": 16.991273996509598,
+ "grad_norm": 0.03913680836558342,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 4.269218444824219,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.02641749382019043,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 2.6905035972595215,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.5879416465759277,
+ "eval_runtime": 60.2807,
+ "eval_samples_per_second": 40.511,
+ "eval_steps_per_second": 0.647,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013961605584644,
+ "grad_norm": 0.048571839928627014,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 4.2008538246154785,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027923211169284,
+ "grad_norm": 0.0640900656580925,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 4.180826663970947,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041884816753928,
+ "grad_norm": 0.06458134204149246,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 4.199562072753906,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05584642233857,
+ "grad_norm": 0.0563119500875473,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 4.171557426452637,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069808027923212,
+ "grad_norm": 0.050914082676172256,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 4.183539390563965,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083769633507853,
+ "grad_norm": 0.053884491324424744,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 4.210450172424316,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097731239092496,
+ "grad_norm": 0.05301345884799957,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 4.1553192138671875,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111692844677137,
+ "grad_norm": 0.051138825714588165,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 4.158510208129883,
+ "step": 1232
+ },
+ {
+ "epoch": 17.12565445026178,
+ "grad_norm": 0.05355675891041756,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 4.167689323425293,
+ "step": 1233
+ },
+ {
+ "epoch": 17.13961605584642,
+ "grad_norm": 0.06062275916337967,
+ "learning_rate": 0.000459203665939781,
+ "loss": 4.170838356018066,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153577661431065,
+ "grad_norm": 0.058666616678237915,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 4.187441825866699,
+ "step": 1235
+ },
+ {
+ "epoch": 17.167539267015705,
+ "grad_norm": 0.057622168213129044,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 4.190746784210205,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18150087260035,
+ "grad_norm": 0.069090336561203,
+ "learning_rate": 0.000458522886595554,
+ "loss": 4.224946022033691,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195462478184993,
+ "grad_norm": 0.06552847474813461,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 4.204067230224609,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209424083769633,
+ "grad_norm": 0.06602177768945694,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 4.189286708831787,
+ "step": 1239
+ },
+ {
+ "epoch": 17.223385689354277,
+ "grad_norm": 0.06751416623592377,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 4.208927154541016,
+ "step": 1240
+ },
+ {
+ "epoch": 17.237347294938917,
+ "grad_norm": 0.0706794261932373,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 4.212923049926758,
+ "step": 1241
+ },
+ {
+ "epoch": 17.25130890052356,
+ "grad_norm": 0.05346836522221565,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 4.2221784591674805,
+ "step": 1242
+ },
+ {
+ "epoch": 17.2652705061082,
+ "grad_norm": 0.05395734682679176,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 4.196428298950195,
+ "step": 1243
+ },
+ {
+ "epoch": 17.279232111692846,
+ "grad_norm": 0.05496550351381302,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 4.208031177520752,
+ "step": 1244
+ },
+ {
+ "epoch": 17.293193717277486,
+ "grad_norm": 0.05320895463228226,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 4.232810974121094,
+ "step": 1245
+ },
+ {
+ "epoch": 17.30715532286213,
+ "grad_norm": 0.05064759403467178,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 4.207826614379883,
+ "step": 1246
+ },
+ {
+ "epoch": 17.32111692844677,
+ "grad_norm": 0.04287239536643028,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 4.195648670196533,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335078534031414,
+ "grad_norm": 0.04527301341295242,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 4.19954776763916,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349040139616054,
+ "grad_norm": 0.04421805590391159,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 4.248671054840088,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363001745200698,
+ "grad_norm": 0.04487175494432449,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 4.2072954177856445,
+ "step": 1250
+ },
+ {
+ "epoch": 17.376963350785342,
+ "grad_norm": 0.04415794834494591,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 4.222837448120117,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390924956369982,
+ "grad_norm": 0.04474013298749924,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 4.213104248046875,
+ "step": 1252
+ },
+ {
+ "epoch": 17.404886561954626,
+ "grad_norm": 0.042631037533283234,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 4.209726333618164,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418848167539267,
+ "grad_norm": 0.04424271360039711,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 4.202699661254883,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43280977312391,
+ "grad_norm": 0.04151812940835953,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 4.190732955932617,
+ "step": 1255
+ },
+ {
+ "epoch": 17.44677137870855,
+ "grad_norm": 0.041588131338357925,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 4.222360610961914,
+ "step": 1256
+ },
+ {
+ "epoch": 17.460732984293195,
+ "grad_norm": 0.042378414422273636,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 4.198424339294434,
+ "step": 1257
+ },
+ {
+ "epoch": 17.474694589877835,
+ "grad_norm": 0.041907574981451035,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 4.224804401397705,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48865619546248,
+ "grad_norm": 0.043774355202913284,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 4.2276611328125,
+ "step": 1259
+ },
+ {
+ "epoch": 17.50261780104712,
+ "grad_norm": 0.04238573834300041,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 4.228109359741211,
+ "step": 1260
+ },
+ {
+ "epoch": 17.516579406631763,
+ "grad_norm": 0.045351386070251465,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 4.208807468414307,
+ "step": 1261
+ },
+ {
+ "epoch": 17.530541012216403,
+ "grad_norm": 0.04300729185342789,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 4.257306098937988,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544502617801047,
+ "grad_norm": 0.041442908346652985,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 4.208004951477051,
+ "step": 1263
+ },
+ {
+ "epoch": 17.55846422338569,
+ "grad_norm": 0.04337337985634804,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 4.23393440246582,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57242582897033,
+ "grad_norm": 0.04335462674498558,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 4.228087425231934,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586387434554975,
+ "grad_norm": 0.03987223282456398,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 4.213971138000488,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600349040139616,
+ "grad_norm": 0.041760094463825226,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 4.248366355895996,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61431064572426,
+ "grad_norm": 0.04420867934823036,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 4.278865814208984,
+ "step": 1268
+ },
+ {
+ "epoch": 17.6282722513089,
+ "grad_norm": 0.04244982823729515,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 4.223642349243164,
+ "step": 1269
+ },
+ {
+ "epoch": 17.642233856893544,
+ "grad_norm": 0.04280664771795273,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 4.231023788452148,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656195462478184,
+ "grad_norm": 0.04294133558869362,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 4.222380638122559,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670157068062828,
+ "grad_norm": 0.04179249703884125,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 4.215641975402832,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68411867364747,
+ "grad_norm": 0.04038798063993454,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 4.228580474853516,
+ "step": 1273
+ },
+ {
+ "epoch": 17.698080279232112,
+ "grad_norm": 0.04211528226733208,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 4.245234489440918,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712041884816752,
+ "grad_norm": 0.04102989658713341,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 4.218495845794678,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726003490401396,
+ "grad_norm": 0.04038692265748978,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 4.2466511726379395,
+ "step": 1276
+ },
+ {
+ "epoch": 17.739965095986037,
+ "grad_norm": 0.03958791121840477,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 4.237762451171875,
+ "step": 1277
+ },
+ {
+ "epoch": 17.75392670157068,
+ "grad_norm": 0.04129907861351967,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 4.238739013671875,
+ "step": 1278
+ },
+ {
+ "epoch": 17.767888307155324,
+ "grad_norm": 0.042581796646118164,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 4.235558032989502,
+ "step": 1279
+ },
+ {
+ "epoch": 17.781849912739965,
+ "grad_norm": 0.04339978098869324,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 4.226593971252441,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79581151832461,
+ "grad_norm": 0.03808023780584335,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 4.238213062286377,
+ "step": 1281
+ },
+ {
+ "epoch": 17.80977312390925,
+ "grad_norm": 0.04022229090332985,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 4.214224815368652,
+ "step": 1282
+ },
+ {
+ "epoch": 17.823734729493893,
+ "grad_norm": 0.04379156231880188,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 4.253068447113037,
+ "step": 1283
+ },
+ {
+ "epoch": 17.837696335078533,
+ "grad_norm": 0.04232964664697647,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 4.249903202056885,
+ "step": 1284
+ },
+ {
+ "epoch": 17.851657940663177,
+ "grad_norm": 0.040980905294418335,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 4.252640724182129,
+ "step": 1285
+ },
+ {
+ "epoch": 17.865619546247817,
+ "grad_norm": 0.04015526548027992,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 4.234862327575684,
+ "step": 1286
+ },
+ {
+ "epoch": 17.87958115183246,
+ "grad_norm": 0.04273828864097595,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 4.216028213500977,
+ "step": 1287
+ },
+ {
+ "epoch": 17.8935427574171,
+ "grad_norm": 0.04210563376545906,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 4.23359489440918,
+ "step": 1288
+ },
+ {
+ "epoch": 17.907504363001745,
+ "grad_norm": 0.038766685873270035,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 4.220360279083252,
+ "step": 1289
+ },
+ {
+ "epoch": 17.921465968586386,
+ "grad_norm": 0.03921609744429588,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 4.253805160522461,
+ "step": 1290
+ },
+ {
+ "epoch": 17.93542757417103,
+ "grad_norm": 0.04071544483304024,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 4.211221694946289,
+ "step": 1291
+ },
+ {
+ "epoch": 17.949389179755673,
+ "grad_norm": 0.03996877372264862,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 4.241739273071289,
+ "step": 1292
+ },
+ {
+ "epoch": 17.963350785340314,
+ "grad_norm": 0.03878853842616081,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 4.237001419067383,
+ "step": 1293
+ },
+ {
+ "epoch": 17.977312390924958,
+ "grad_norm": 0.03889666125178337,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 4.244929790496826,
+ "step": 1294
+ },
+ {
+ "epoch": 17.991273996509598,
+ "grad_norm": 0.03876441717147827,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 4.236083030700684,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.027286000549793243,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 2.636739730834961,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.5896694660186768,
+ "eval_runtime": 60.4742,
+ "eval_samples_per_second": 40.381,
+ "eval_steps_per_second": 0.645,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013961605584644,
+ "grad_norm": 0.04279542714357376,
+ "learning_rate": 0.000444677103403854,
+ "loss": 4.136401653289795,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027923211169284,
+ "grad_norm": 0.051685452461242676,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 4.159584045410156,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041884816753928,
+ "grad_norm": 0.05411457270383835,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 4.160367965698242,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05584642233857,
+ "grad_norm": 0.048574138432741165,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 4.134189128875732,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069808027923212,
+ "grad_norm": 0.04528478533029556,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 4.1022725105285645,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083769633507853,
+ "grad_norm": 0.04474090039730072,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 4.16152286529541,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097731239092496,
+ "grad_norm": 0.04882791265845299,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 4.1553144454956055,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111692844677137,
+ "grad_norm": 0.04831654578447342,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 4.147022724151611,
+ "step": 1304
+ },
+ {
+ "epoch": 18.12565445026178,
+ "grad_norm": 0.047365427017211914,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 4.139405250549316,
+ "step": 1305
+ },
+ {
+ "epoch": 18.13961605584642,
+ "grad_norm": 0.0514012910425663,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 4.115568161010742,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153577661431065,
+ "grad_norm": 0.05225161090493202,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 4.166429042816162,
+ "step": 1307
+ },
+ {
+ "epoch": 18.167539267015705,
+ "grad_norm": 0.05575249716639519,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 4.1414289474487305,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18150087260035,
+ "grad_norm": 0.05405383184552193,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 4.138680934906006,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195462478184993,
+ "grad_norm": 0.05185031518340111,
+ "learning_rate": 0.000441621628895411,
+ "loss": 4.159243583679199,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209424083769633,
+ "grad_norm": 0.05070916935801506,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 4.1844892501831055,
+ "step": 1311
+ },
+ {
+ "epoch": 18.223385689354277,
+ "grad_norm": 0.05037331581115723,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 4.132781028747559,
+ "step": 1312
+ },
+ {
+ "epoch": 18.237347294938917,
+ "grad_norm": 0.05082111805677414,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 4.143031120300293,
+ "step": 1313
+ },
+ {
+ "epoch": 18.25130890052356,
+ "grad_norm": 0.04742169380187988,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 4.137580871582031,
+ "step": 1314
+ },
+ {
+ "epoch": 18.2652705061082,
+ "grad_norm": 0.049042731523513794,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 4.159721374511719,
+ "step": 1315
+ },
+ {
+ "epoch": 18.279232111692846,
+ "grad_norm": 0.045649368315935135,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 4.144008636474609,
+ "step": 1316
+ },
+ {
+ "epoch": 18.293193717277486,
+ "grad_norm": 0.0473787859082222,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 4.164243698120117,
+ "step": 1317
+ },
+ {
+ "epoch": 18.30715532286213,
+ "grad_norm": 0.047626055777072906,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 4.164441108703613,
+ "step": 1318
+ },
+ {
+ "epoch": 18.32111692844677,
+ "grad_norm": 0.04738573357462883,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 4.180338382720947,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335078534031414,
+ "grad_norm": 0.04712498560547829,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 4.146285533905029,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349040139616054,
+ "grad_norm": 0.045505691319704056,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 4.162572860717773,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363001745200698,
+ "grad_norm": 0.0489528588950634,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 4.160312175750732,
+ "step": 1322
+ },
+ {
+ "epoch": 18.376963350785342,
+ "grad_norm": 0.047517575323581696,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 4.164361953735352,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390924956369982,
+ "grad_norm": 0.044214535504579544,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 4.182154655456543,
+ "step": 1324
+ },
+ {
+ "epoch": 18.404886561954626,
+ "grad_norm": 0.043775733560323715,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 4.130522727966309,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418848167539267,
+ "grad_norm": 0.044364530593156815,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 4.151247024536133,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43280977312391,
+ "grad_norm": 0.046320103108882904,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 4.204159736633301,
+ "step": 1327
+ },
+ {
+ "epoch": 18.44677137870855,
+ "grad_norm": 0.04802565276622772,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 4.182550430297852,
+ "step": 1328
+ },
+ {
+ "epoch": 18.460732984293195,
+ "grad_norm": 0.049340926110744476,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 4.163658142089844,
+ "step": 1329
+ },
+ {
+ "epoch": 18.474694589877835,
+ "grad_norm": 0.046701837331056595,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 4.170445442199707,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48865619546248,
+ "grad_norm": 0.04746273159980774,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 4.220232963562012,
+ "step": 1331
+ },
+ {
+ "epoch": 18.50261780104712,
+ "grad_norm": 0.04644489288330078,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 4.190027236938477,
+ "step": 1332
+ },
+ {
+ "epoch": 18.516579406631763,
+ "grad_norm": 0.04394262284040451,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 4.159570693969727,
+ "step": 1333
+ },
+ {
+ "epoch": 18.530541012216403,
+ "grad_norm": 0.04609063267707825,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 4.1657586097717285,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544502617801047,
+ "grad_norm": 0.0450211800634861,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 4.184865474700928,
+ "step": 1335
+ },
+ {
+ "epoch": 18.55846422338569,
+ "grad_norm": 0.0475253090262413,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 4.185857772827148,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57242582897033,
+ "grad_norm": 0.04332203418016434,
+ "learning_rate": 0.000435215814386134,
+ "loss": 4.160165309906006,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586387434554975,
+ "grad_norm": 0.0427570678293705,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 4.1727495193481445,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600349040139616,
+ "grad_norm": 0.04651174694299698,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 4.204797744750977,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61431064572426,
+ "grad_norm": 0.04324701428413391,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 4.186931610107422,
+ "step": 1340
+ },
+ {
+ "epoch": 18.6282722513089,
+ "grad_norm": 0.044644393026828766,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 4.209554195404053,
+ "step": 1341
+ },
+ {
+ "epoch": 18.642233856893544,
+ "grad_norm": 0.042982228100299835,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 4.205570220947266,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656195462478184,
+ "grad_norm": 0.04385090991854668,
+ "learning_rate": 0.000433781621332752,
+ "loss": 4.177936553955078,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670157068062828,
+ "grad_norm": 0.04771510139107704,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 4.1795806884765625,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68411867364747,
+ "grad_norm": 0.04551433399319649,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 4.192194938659668,
+ "step": 1345
+ },
+ {
+ "epoch": 18.698080279232112,
+ "grad_norm": 0.043501488864421844,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 4.177781105041504,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712041884816752,
+ "grad_norm": 0.04394034296274185,
+ "learning_rate": 0.000432823382695327,
+ "loss": 4.206354141235352,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726003490401396,
+ "grad_norm": 0.04911153391003609,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 4.205874919891357,
+ "step": 1348
+ },
+ {
+ "epoch": 18.739965095986037,
+ "grad_norm": 0.04602045193314552,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 4.1947021484375,
+ "step": 1349
+ },
+ {
+ "epoch": 18.75392670157068,
+ "grad_norm": 0.04393904283642769,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 4.188337802886963,
+ "step": 1350
+ },
+ {
+ "epoch": 18.767888307155324,
+ "grad_norm": 0.04526393488049507,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 4.175235748291016,
+ "step": 1351
+ },
+ {
+ "epoch": 18.781849912739965,
+ "grad_norm": 0.041930247098207474,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 4.210986137390137,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79581151832461,
+ "grad_norm": 0.044262565672397614,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 4.176301002502441,
+ "step": 1353
+ },
+ {
+ "epoch": 18.80977312390925,
+ "grad_norm": 0.044190581887960434,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 4.1888837814331055,
+ "step": 1354
+ },
+ {
+ "epoch": 18.823734729493893,
+ "grad_norm": 0.045371655374765396,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 4.182244777679443,
+ "step": 1355
+ },
+ {
+ "epoch": 18.837696335078533,
+ "grad_norm": 0.04302282631397247,
+ "learning_rate": 0.000430661245733797,
+ "loss": 4.189350128173828,
+ "step": 1356
+ },
+ {
+ "epoch": 18.851657940663177,
+ "grad_norm": 0.04402372986078262,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 4.233372211456299,
+ "step": 1357
+ },
+ {
+ "epoch": 18.865619546247817,
+ "grad_norm": 0.04397045075893402,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 4.21682071685791,
+ "step": 1358
+ },
+ {
+ "epoch": 18.87958115183246,
+ "grad_norm": 0.04319734871387482,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 4.204183578491211,
+ "step": 1359
+ },
+ {
+ "epoch": 18.8935427574171,
+ "grad_norm": 0.04600509628653526,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 4.206524848937988,
+ "step": 1360
+ },
+ {
+ "epoch": 18.907504363001745,
+ "grad_norm": 0.04364705830812454,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 4.213656425476074,
+ "step": 1361
+ },
+ {
+ "epoch": 18.921465968586386,
+ "grad_norm": 0.044004205614328384,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 4.1841583251953125,
+ "step": 1362
+ },
+ {
+ "epoch": 18.93542757417103,
+ "grad_norm": 0.04480595886707306,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 4.166214942932129,
+ "step": 1363
+ },
+ {
+ "epoch": 18.949389179755673,
+ "grad_norm": 0.04260887950658798,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 4.198726654052734,
+ "step": 1364
+ },
+ {
+ "epoch": 18.963350785340314,
+ "grad_norm": 0.04513927176594734,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 4.2122392654418945,
+ "step": 1365
+ },
+ {
+ "epoch": 18.977312390924958,
+ "grad_norm": 0.04286042973399162,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 4.192383766174316,
+ "step": 1366
+ },
+ {
+ "epoch": 18.991273996509598,
+ "grad_norm": 0.04287346079945564,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 4.221860408782959,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.030670644715428352,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 2.624823570251465,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.5910831689834595,
+ "eval_runtime": 60.5517,
+ "eval_samples_per_second": 40.329,
+ "eval_steps_per_second": 0.644,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013961605584644,
+ "grad_norm": 0.046196095645427704,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 4.104501724243164,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027923211169284,
+ "grad_norm": 0.05348595231771469,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 4.109098434448242,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041884816753928,
+ "grad_norm": 0.05256905406713486,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 4.115832328796387,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05584642233857,
+ "grad_norm": 0.05241430550813675,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 4.079610824584961,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069808027923212,
+ "grad_norm": 0.04983370006084442,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 4.095335006713867,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083769633507853,
+ "grad_norm": 0.04900007322430611,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 4.093306541442871,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097731239092496,
+ "grad_norm": 0.04995666816830635,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 4.111114501953125,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111692844677137,
+ "grad_norm": 0.04875681549310684,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 4.091944694519043,
+ "step": 1376
+ },
+ {
+ "epoch": 19.12565445026178,
+ "grad_norm": 0.05283811315894127,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 4.112549304962158,
+ "step": 1377
+ },
+ {
+ "epoch": 19.13961605584642,
+ "grad_norm": 0.04801992326974869,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 4.108818054199219,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153577661431065,
+ "grad_norm": 0.04751930385828018,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 4.089441299438477,
+ "step": 1379
+ },
+ {
+ "epoch": 19.167539267015705,
+ "grad_norm": 0.045653726905584335,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 4.07722282409668,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18150087260035,
+ "grad_norm": 0.045616209506988525,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 4.097578048706055,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195462478184993,
+ "grad_norm": 0.04929365590214729,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 4.082221984863281,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209424083769633,
+ "grad_norm": 0.04915834963321686,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 4.123330116271973,
+ "step": 1383
+ },
+ {
+ "epoch": 19.223385689354277,
+ "grad_norm": 0.0470132976770401,
+ "learning_rate": 0.000423881957237287,
+ "loss": 4.113927841186523,
+ "step": 1384
+ },
+ {
+ "epoch": 19.237347294938917,
+ "grad_norm": 0.04695259779691696,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 4.096814155578613,
+ "step": 1385
+ },
+ {
+ "epoch": 19.25130890052356,
+ "grad_norm": 0.04730398207902908,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 4.122525215148926,
+ "step": 1386
+ },
+ {
+ "epoch": 19.2652705061082,
+ "grad_norm": 0.050295207649469376,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 4.117993354797363,
+ "step": 1387
+ },
+ {
+ "epoch": 19.279232111692846,
+ "grad_norm": 0.05082492157816887,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 4.14082145690918,
+ "step": 1388
+ },
+ {
+ "epoch": 19.293193717277486,
+ "grad_norm": 0.04960232973098755,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 4.139887809753418,
+ "step": 1389
+ },
+ {
+ "epoch": 19.30715532286213,
+ "grad_norm": 0.04851173236966133,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 4.138492584228516,
+ "step": 1390
+ },
+ {
+ "epoch": 19.32111692844677,
+ "grad_norm": 0.047838468104600906,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 4.118064880371094,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335078534031414,
+ "grad_norm": 0.05148220434784889,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 4.136533737182617,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349040139616054,
+ "grad_norm": 0.053177181631326675,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 4.1125168800354,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363001745200698,
+ "grad_norm": 0.04861171543598175,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 4.1437153816223145,
+ "step": 1394
+ },
+ {
+ "epoch": 19.376963350785342,
+ "grad_norm": 0.05194924771785736,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 4.097813606262207,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390924956369982,
+ "grad_norm": 0.049609847366809845,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 4.068940162658691,
+ "step": 1396
+ },
+ {
+ "epoch": 19.404886561954626,
+ "grad_norm": 0.047042325139045715,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 4.121333122253418,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418848167539267,
+ "grad_norm": 0.047646693885326385,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 4.114943981170654,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43280977312391,
+ "grad_norm": 0.04676460474729538,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 4.139252185821533,
+ "step": 1399
+ },
+ {
+ "epoch": 19.44677137870855,
+ "grad_norm": 0.04905956983566284,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 4.135970115661621,
+ "step": 1400
+ },
+ {
+ "epoch": 19.460732984293195,
+ "grad_norm": 0.050025034695863724,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 4.15662956237793,
+ "step": 1401
+ },
+ {
+ "epoch": 19.474694589877835,
+ "grad_norm": 0.04740425944328308,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 4.140295505523682,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48865619546248,
+ "grad_norm": 0.046904709190130234,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 4.14083194732666,
+ "step": 1403
+ },
+ {
+ "epoch": 19.50261780104712,
+ "grad_norm": 0.048393651843070984,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 4.161823272705078,
+ "step": 1404
+ },
+ {
+ "epoch": 19.516579406631763,
+ "grad_norm": 0.04766222462058067,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 4.151817798614502,
+ "step": 1405
+ },
+ {
+ "epoch": 19.530541012216403,
+ "grad_norm": 0.04797709733247757,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 4.173160552978516,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544502617801047,
+ "grad_norm": 0.05012711137533188,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 4.161989688873291,
+ "step": 1407
+ },
+ {
+ "epoch": 19.55846422338569,
+ "grad_norm": 0.04726000875234604,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 4.136838436126709,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57242582897033,
+ "grad_norm": 0.046251531690359116,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 4.1526994705200195,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586387434554975,
+ "grad_norm": 0.047958049923181534,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 4.13817024230957,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600349040139616,
+ "grad_norm": 0.046333350241184235,
+ "learning_rate": 0.000417272427439646,
+ "loss": 4.160187721252441,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61431064572426,
+ "grad_norm": 0.0488678440451622,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 4.166190147399902,
+ "step": 1412
+ },
+ {
+ "epoch": 19.6282722513089,
+ "grad_norm": 0.04784323647618294,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 4.164414405822754,
+ "step": 1413
+ },
+ {
+ "epoch": 19.642233856893544,
+ "grad_norm": 0.05064673349261284,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 4.149814128875732,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656195462478184,
+ "grad_norm": 0.05072085186839104,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 4.167759895324707,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670157068062828,
+ "grad_norm": 0.0499395951628685,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 4.162792205810547,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68411867364747,
+ "grad_norm": 0.046766918152570724,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 4.1416521072387695,
+ "step": 1417
+ },
+ {
+ "epoch": 19.698080279232112,
+ "grad_norm": 0.04523273929953575,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 4.161751747131348,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712041884816752,
+ "grad_norm": 0.046106480062007904,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 4.166394233703613,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726003490401396,
+ "grad_norm": 0.04666793718934059,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 4.124961853027344,
+ "step": 1420
+ },
+ {
+ "epoch": 19.739965095986037,
+ "grad_norm": 0.047407012432813644,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 4.1794328689575195,
+ "step": 1421
+ },
+ {
+ "epoch": 19.75392670157068,
+ "grad_norm": 0.04690873995423317,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 4.160114765167236,
+ "step": 1422
+ },
+ {
+ "epoch": 19.767888307155324,
+ "grad_norm": 0.04589436575770378,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 4.137969970703125,
+ "step": 1423
+ },
+ {
+ "epoch": 19.781849912739965,
+ "grad_norm": 0.04641961678862572,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 4.161533355712891,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79581151832461,
+ "grad_norm": 0.045799050480127335,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 4.187039375305176,
+ "step": 1425
+ },
+ {
+ "epoch": 19.80977312390925,
+ "grad_norm": 0.04881894588470459,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 4.158107757568359,
+ "step": 1426
+ },
+ {
+ "epoch": 19.823734729493893,
+ "grad_norm": 0.04506978392601013,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 4.143360137939453,
+ "step": 1427
+ },
+ {
+ "epoch": 19.837696335078533,
+ "grad_norm": 0.04504861310124397,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 4.174564361572266,
+ "step": 1428
+ },
+ {
+ "epoch": 19.851657940663177,
+ "grad_norm": 0.04639774188399315,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 4.162899971008301,
+ "step": 1429
+ },
+ {
+ "epoch": 19.865619546247817,
+ "grad_norm": 0.047444432973861694,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 4.143390655517578,
+ "step": 1430
+ },
+ {
+ "epoch": 19.87958115183246,
+ "grad_norm": 0.04461613669991493,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 4.171284198760986,
+ "step": 1431
+ },
+ {
+ "epoch": 19.8935427574171,
+ "grad_norm": 0.046383798122406006,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 4.14809513092041,
+ "step": 1432
+ },
+ {
+ "epoch": 19.907504363001745,
+ "grad_norm": 0.04642366245388985,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 4.178903579711914,
+ "step": 1433
+ },
+ {
+ "epoch": 19.921465968586386,
+ "grad_norm": 0.0451771542429924,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 4.171825408935547,
+ "step": 1434
+ },
+ {
+ "epoch": 19.93542757417103,
+ "grad_norm": 0.04767893627285957,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 4.169270038604736,
+ "step": 1435
+ },
+ {
+ "epoch": 19.949389179755673,
+ "grad_norm": 0.04673412814736366,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 4.175506591796875,
+ "step": 1436
+ },
+ {
+ "epoch": 19.963350785340314,
+ "grad_norm": 0.04465145245194435,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 4.149707794189453,
+ "step": 1437
+ },
+ {
+ "epoch": 19.977312390924958,
+ "grad_norm": 0.04734131321310997,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 4.166044235229492,
+ "step": 1438
+ },
+ {
+ "epoch": 19.991273996509598,
+ "grad_norm": 0.04706178605556488,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 4.152081489562988,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.031556785106658936,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 2.60109543800354,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013961605584644,
+ "grad_norm": 0.05152537673711777,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 4.084830284118652,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027923211169284,
+ "grad_norm": 0.05918274074792862,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 4.045302391052246,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041884816753928,
+ "grad_norm": 0.05093073844909668,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 4.070842742919922,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05584642233857,
+ "grad_norm": 0.05095985531806946,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 4.083273410797119,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069808027923212,
+ "grad_norm": 0.05502687394618988,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 4.072626113891602,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083769633507853,
+ "grad_norm": 0.05037364363670349,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 4.075478553771973,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097731239092496,
+ "grad_norm": 0.04983789846301079,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 4.066207408905029,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111692844677137,
+ "grad_norm": 0.049478255212306976,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 4.069969654083252,
+ "step": 1448
+ },
+ {
+ "epoch": 20.12565445026178,
+ "grad_norm": 0.051651593297719955,
+ "learning_rate": 0.000407857329622967,
+ "loss": 4.050713062286377,
+ "step": 1449
+ },
+ {
+ "epoch": 20.13961605584642,
+ "grad_norm": 0.05436555668711662,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 4.069849967956543,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153577661431065,
+ "grad_norm": 0.055190350860357285,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 4.081624984741211,
+ "step": 1451
+ },
+ {
+ "epoch": 20.167539267015705,
+ "grad_norm": 0.053246211260557175,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 4.076076507568359,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18150087260035,
+ "grad_norm": 0.051757775247097015,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 4.049139022827148,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195462478184993,
+ "grad_norm": 0.05064203217625618,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 4.068076133728027,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209424083769633,
+ "grad_norm": 0.05114085227251053,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 4.092901229858398,
+ "step": 1455
+ },
+ {
+ "epoch": 20.223385689354277,
+ "grad_norm": 0.047997843474149704,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 4.077084541320801,
+ "step": 1456
+ },
+ {
+ "epoch": 20.237347294938917,
+ "grad_norm": 0.04880451038479805,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 4.059187889099121,
+ "step": 1457
+ },
+ {
+ "epoch": 20.25130890052356,
+ "grad_norm": 0.05036284029483795,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 4.102115154266357,
+ "step": 1458
+ },
+ {
+ "epoch": 20.2652705061082,
+ "grad_norm": 0.04916979745030403,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 4.098356246948242,
+ "step": 1459
+ },
+ {
+ "epoch": 20.279232111692846,
+ "grad_norm": 0.04713984206318855,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 4.057218074798584,
+ "step": 1460
+ },
+ {
+ "epoch": 20.293193717277486,
+ "grad_norm": 0.04982447996735573,
+ "learning_rate": 0.000404858275823277,
+ "loss": 4.070363998413086,
+ "step": 1461
+ },
+ {
+ "epoch": 20.30715532286213,
+ "grad_norm": 0.0475616492331028,
+ "learning_rate": 0.000404607816444578,
+ "loss": 4.074314117431641,
+ "step": 1462
+ },
+ {
+ "epoch": 20.32111692844677,
+ "grad_norm": 0.046537093818187714,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 4.079501152038574,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335078534031414,
+ "grad_norm": 0.04889644682407379,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 4.0731425285339355,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349040139616054,
+ "grad_norm": 0.05095166340470314,
+ "learning_rate": 0.000403855947934478,
+ "loss": 4.08765983581543,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363001745200698,
+ "grad_norm": 0.049677830189466476,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 4.128052711486816,
+ "step": 1466
+ },
+ {
+ "epoch": 20.376963350785342,
+ "grad_norm": 0.05221942439675331,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 4.094407081604004,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390924956369982,
+ "grad_norm": 0.052042894065380096,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 4.088282585144043,
+ "step": 1468
+ },
+ {
+ "epoch": 20.404886561954626,
+ "grad_norm": 0.04822041094303131,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 4.1070990562438965,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418848167539267,
+ "grad_norm": 0.047207772731781006,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 4.102565765380859,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43280977312391,
+ "grad_norm": 0.049633774906396866,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 4.097216606140137,
+ "step": 1471
+ },
+ {
+ "epoch": 20.44677137870855,
+ "grad_norm": 0.05135185644030571,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 4.095193862915039,
+ "step": 1472
+ },
+ {
+ "epoch": 20.460732984293195,
+ "grad_norm": 0.05215826258063316,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 4.085207939147949,
+ "step": 1473
+ },
+ {
+ "epoch": 20.474694589877835,
+ "grad_norm": 0.050653353333473206,
+ "learning_rate": 0.000401595976318565,
+ "loss": 4.085142612457275,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48865619546248,
+ "grad_norm": 0.04795852303504944,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 4.1158270835876465,
+ "step": 1475
+ },
+ {
+ "epoch": 20.50261780104712,
+ "grad_norm": 0.047416940331459045,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 4.096989154815674,
+ "step": 1476
+ },
+ {
+ "epoch": 20.516579406631763,
+ "grad_norm": 0.049753669649362564,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 4.118251800537109,
+ "step": 1477
+ },
+ {
+ "epoch": 20.530541012216403,
+ "grad_norm": 0.04889055714011192,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 4.081287384033203,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544502617801047,
+ "grad_norm": 0.04857790097594261,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 4.084896564483643,
+ "step": 1479
+ },
+ {
+ "epoch": 20.55846422338569,
+ "grad_norm": 0.047250498086214066,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 4.113656997680664,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57242582897033,
+ "grad_norm": 0.048608481884002686,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 4.113809108734131,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586387434554975,
+ "grad_norm": 0.048189081251621246,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 4.113949775695801,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600349040139616,
+ "grad_norm": 0.049929242581129074,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 4.104221820831299,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61431064572426,
+ "grad_norm": 0.048482850193977356,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 4.091039180755615,
+ "step": 1484
+ },
+ {
+ "epoch": 20.6282722513089,
+ "grad_norm": 0.04810645803809166,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 4.130088806152344,
+ "step": 1485
+ },
+ {
+ "epoch": 20.642233856893544,
+ "grad_norm": 0.04745674133300781,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 4.106161117553711,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656195462478184,
+ "grad_norm": 0.048513609915971756,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 4.119222164154053,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670157068062828,
+ "grad_norm": 0.04506325349211693,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 4.115543842315674,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68411867364747,
+ "grad_norm": 0.04726741835474968,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 4.084345817565918,
+ "step": 1489
+ },
+ {
+ "epoch": 20.698080279232112,
+ "grad_norm": 0.04873545467853546,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 4.116580009460449,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712041884816752,
+ "grad_norm": 0.051252465695142746,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 4.129733085632324,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726003490401396,
+ "grad_norm": 0.0489567406475544,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 4.103477954864502,
+ "step": 1492
+ },
+ {
+ "epoch": 20.739965095986037,
+ "grad_norm": 0.04539056867361069,
+ "learning_rate": 0.000396803974909638,
+ "loss": 4.1327667236328125,
+ "step": 1493
+ },
+ {
+ "epoch": 20.75392670157068,
+ "grad_norm": 0.04954095557332039,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 4.084133625030518,
+ "step": 1494
+ },
+ {
+ "epoch": 20.767888307155324,
+ "grad_norm": 0.05041591823101044,
+ "learning_rate": 0.000396297942393265,
+ "loss": 4.12288761138916,
+ "step": 1495
+ },
+ {
+ "epoch": 20.781849912739965,
+ "grad_norm": 0.046549517661333084,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 4.1259918212890625,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79581151832461,
+ "grad_norm": 0.04704999923706055,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 4.102563858032227,
+ "step": 1497
+ },
+ {
+ "epoch": 20.80977312390925,
+ "grad_norm": 0.05025320500135422,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 4.135492324829102,
+ "step": 1498
+ },
+ {
+ "epoch": 20.823734729493893,
+ "grad_norm": 0.04559441655874252,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 4.118548393249512,
+ "step": 1499
+ },
+ {
+ "epoch": 20.837696335078533,
+ "grad_norm": 0.04913092032074928,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 4.129302024841309,
+ "step": 1500
+ },
+ {
+ "epoch": 20.851657940663177,
+ "grad_norm": 0.04846089705824852,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 4.166059970855713,
+ "step": 1501
+ },
+ {
+ "epoch": 20.865619546247817,
+ "grad_norm": 0.04710553213953972,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 4.104712009429932,
+ "step": 1502
+ },
+ {
+ "epoch": 20.87958115183246,
+ "grad_norm": 0.04896444454789162,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 4.140282154083252,
+ "step": 1503
+ },
+ {
+ "epoch": 20.8935427574171,
+ "grad_norm": 0.0469740591943264,
+ "learning_rate": 0.000394017098798633,
+ "loss": 4.125306129455566,
+ "step": 1504
+ },
+ {
+ "epoch": 20.907504363001745,
+ "grad_norm": 0.04638523980975151,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 4.133161544799805,
+ "step": 1505
+ },
+ {
+ "epoch": 20.921465968586386,
+ "grad_norm": 0.04814442619681358,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 4.093092918395996,
+ "step": 1506
+ },
+ {
+ "epoch": 20.93542757417103,
+ "grad_norm": 0.049911532551050186,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 4.1001715660095215,
+ "step": 1507
+ },
+ {
+ "epoch": 20.949389179755673,
+ "grad_norm": 0.0446670837700367,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 4.140350341796875,
+ "step": 1508
+ },
+ {
+ "epoch": 20.963350785340314,
+ "grad_norm": 0.04789397493004799,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 4.163155555725098,
+ "step": 1509
+ },
+ {
+ "epoch": 20.977312390924958,
+ "grad_norm": 0.050148315727710724,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 4.118380069732666,
+ "step": 1510
+ },
+ {
+ "epoch": 20.991273996509598,
+ "grad_norm": 0.047446198761463165,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 4.116507530212402,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.03415544703602791,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 2.5738282203674316,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.5933796167373657,
+ "eval_runtime": 60.4011,
+ "eval_samples_per_second": 40.43,
+ "eval_steps_per_second": 0.646,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013961605584644,
+ "grad_norm": 0.052276045083999634,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 4.0301923751831055,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027923211169284,
+ "grad_norm": 0.06514342874288559,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 4.031280517578125,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041884816753928,
+ "grad_norm": 0.05948134884238243,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 4.060680389404297,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05584642233857,
+ "grad_norm": 0.05251640826463699,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 4.0218400955200195,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069808027923212,
+ "grad_norm": 0.05426032841205597,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 4.026187896728516,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083769633507853,
+ "grad_norm": 0.052583567798137665,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 4.016767501831055,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097731239092496,
+ "grad_norm": 0.04966135695576668,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 4.043508529663086,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111692844677137,
+ "grad_norm": 0.053838398307561874,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 4.054832458496094,
+ "step": 1520
+ },
+ {
+ "epoch": 21.12565445026178,
+ "grad_norm": 0.05304184556007385,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 4.03280782699585,
+ "step": 1521
+ },
+ {
+ "epoch": 21.13961605584642,
+ "grad_norm": 0.05252664163708687,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 4.039841175079346,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153577661431065,
+ "grad_norm": 0.054593686014413834,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 4.048308372497559,
+ "step": 1523
+ },
+ {
+ "epoch": 21.167539267015705,
+ "grad_norm": 0.050722211599349976,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 4.051860332489014,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18150087260035,
+ "grad_norm": 0.05248095095157623,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 4.056846618652344,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195462478184993,
+ "grad_norm": 0.05166107788681984,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 4.045178413391113,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209424083769633,
+ "grad_norm": 0.04987792298197746,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 4.014827728271484,
+ "step": 1527
+ },
+ {
+ "epoch": 21.223385689354277,
+ "grad_norm": 0.05316167324781418,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 4.033143997192383,
+ "step": 1528
+ },
+ {
+ "epoch": 21.237347294938917,
+ "grad_norm": 0.05228257551789284,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 4.033842086791992,
+ "step": 1529
+ },
+ {
+ "epoch": 21.25130890052356,
+ "grad_norm": 0.049349382519721985,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 4.025982856750488,
+ "step": 1530
+ },
+ {
+ "epoch": 21.2652705061082,
+ "grad_norm": 0.055104926228523254,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 4.059151649475098,
+ "step": 1531
+ },
+ {
+ "epoch": 21.279232111692846,
+ "grad_norm": 0.052629340440034866,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 4.032320976257324,
+ "step": 1532
+ },
+ {
+ "epoch": 21.293193717277486,
+ "grad_norm": 0.05396415665745735,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 4.0425567626953125,
+ "step": 1533
+ },
+ {
+ "epoch": 21.30715532286213,
+ "grad_norm": 0.054417286068201065,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 4.03759241104126,
+ "step": 1534
+ },
+ {
+ "epoch": 21.32111692844677,
+ "grad_norm": 0.061248697340488434,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 4.0441083908081055,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335078534031414,
+ "grad_norm": 0.05934970825910568,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 4.050718307495117,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349040139616054,
+ "grad_norm": 0.052805181592702866,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 4.069460868835449,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363001745200698,
+ "grad_norm": 0.052820611745119095,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 4.063657283782959,
+ "step": 1538
+ },
+ {
+ "epoch": 21.376963350785342,
+ "grad_norm": 0.05043052136898041,
+ "learning_rate": 0.00038509205478744,
+ "loss": 4.029552459716797,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390924956369982,
+ "grad_norm": 0.04828709363937378,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 4.05656623840332,
+ "step": 1540
+ },
+ {
+ "epoch": 21.404886561954626,
+ "grad_norm": 0.049760788679122925,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 4.066338539123535,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418848167539267,
+ "grad_norm": 0.05037159472703934,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 4.06976842880249,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43280977312391,
+ "grad_norm": 0.05276317894458771,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 4.062975883483887,
+ "step": 1543
+ },
+ {
+ "epoch": 21.44677137870855,
+ "grad_norm": 0.05571872368454933,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 4.065622329711914,
+ "step": 1544
+ },
+ {
+ "epoch": 21.460732984293195,
+ "grad_norm": 0.05438469722867012,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 4.074963569641113,
+ "step": 1545
+ },
+ {
+ "epoch": 21.474694589877835,
+ "grad_norm": 0.053415946662425995,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 4.080055236816406,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48865619546248,
+ "grad_norm": 0.05216296389698982,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 4.091355800628662,
+ "step": 1547
+ },
+ {
+ "epoch": 21.50261780104712,
+ "grad_norm": 0.051212459802627563,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 4.058528900146484,
+ "step": 1548
+ },
+ {
+ "epoch": 21.516579406631763,
+ "grad_norm": 0.05033959075808525,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 4.064615726470947,
+ "step": 1549
+ },
+ {
+ "epoch": 21.530541012216403,
+ "grad_norm": 0.04665188118815422,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 4.059698104858398,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544502617801047,
+ "grad_norm": 0.050221361219882965,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 4.068614959716797,
+ "step": 1551
+ },
+ {
+ "epoch": 21.55846422338569,
+ "grad_norm": 0.05089551955461502,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 4.0833892822265625,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57242582897033,
+ "grad_norm": 0.051966603845357895,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 4.083354949951172,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586387434554975,
+ "grad_norm": 0.04981927201151848,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 4.083340644836426,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600349040139616,
+ "grad_norm": 0.05141765624284744,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 4.066590309143066,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61431064572426,
+ "grad_norm": 0.048985838890075684,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 4.088237762451172,
+ "step": 1556
+ },
+ {
+ "epoch": 21.6282722513089,
+ "grad_norm": 0.05013054236769676,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 4.081377983093262,
+ "step": 1557
+ },
+ {
+ "epoch": 21.642233856893544,
+ "grad_norm": 0.05170975625514984,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 4.054303169250488,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656195462478184,
+ "grad_norm": 0.048257142305374146,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 4.103672504425049,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670157068062828,
+ "grad_norm": 0.049942679703235626,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 4.029951572418213,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68411867364747,
+ "grad_norm": 0.05083315819501877,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 4.048798084259033,
+ "step": 1561
+ },
+ {
+ "epoch": 21.698080279232112,
+ "grad_norm": 0.04984128847718239,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 4.079655647277832,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712041884816752,
+ "grad_norm": 0.04985307902097702,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 4.071877956390381,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726003490401396,
+ "grad_norm": 0.051137227565050125,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 4.076539993286133,
+ "step": 1564
+ },
+ {
+ "epoch": 21.739965095986037,
+ "grad_norm": 0.04968905821442604,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 4.084498882293701,
+ "step": 1565
+ },
+ {
+ "epoch": 21.75392670157068,
+ "grad_norm": 0.049119509756565094,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 4.066773414611816,
+ "step": 1566
+ },
+ {
+ "epoch": 21.767888307155324,
+ "grad_norm": 0.05107425898313522,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 4.065309047698975,
+ "step": 1567
+ },
+ {
+ "epoch": 21.781849912739965,
+ "grad_norm": 0.049012504518032074,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 4.062110900878906,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79581151832461,
+ "grad_norm": 0.04972011595964432,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 4.0630903244018555,
+ "step": 1569
+ },
+ {
+ "epoch": 21.80977312390925,
+ "grad_norm": 0.04822700098156929,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 4.106127738952637,
+ "step": 1570
+ },
+ {
+ "epoch": 21.823734729493893,
+ "grad_norm": 0.04808267205953598,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 4.085200309753418,
+ "step": 1571
+ },
+ {
+ "epoch": 21.837696335078533,
+ "grad_norm": 0.048644762486219406,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 4.0856170654296875,
+ "step": 1572
+ },
+ {
+ "epoch": 21.851657940663177,
+ "grad_norm": 0.05232283100485802,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 4.126387119293213,
+ "step": 1573
+ },
+ {
+ "epoch": 21.865619546247817,
+ "grad_norm": 0.05095643922686577,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 4.108254432678223,
+ "step": 1574
+ },
+ {
+ "epoch": 21.87958115183246,
+ "grad_norm": 0.045972101390361786,
+ "learning_rate": 0.000375827577564042,
+ "loss": 4.058515548706055,
+ "step": 1575
+ },
+ {
+ "epoch": 21.8935427574171,
+ "grad_norm": 0.04974910616874695,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 4.089197158813477,
+ "step": 1576
+ },
+ {
+ "epoch": 21.907504363001745,
+ "grad_norm": 0.04890159144997597,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 4.1078410148620605,
+ "step": 1577
+ },
+ {
+ "epoch": 21.921465968586386,
+ "grad_norm": 0.04738525673747063,
+ "learning_rate": 0.000375051971546236,
+ "loss": 4.1219282150268555,
+ "step": 1578
+ },
+ {
+ "epoch": 21.93542757417103,
+ "grad_norm": 0.04713096842169762,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 4.10295295715332,
+ "step": 1579
+ },
+ {
+ "epoch": 21.949389179755673,
+ "grad_norm": 0.04834369197487831,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 4.102747917175293,
+ "step": 1580
+ },
+ {
+ "epoch": 21.963350785340314,
+ "grad_norm": 0.047073688358068466,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 4.08024787902832,
+ "step": 1581
+ },
+ {
+ "epoch": 21.977312390924958,
+ "grad_norm": 0.04568164795637131,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 4.078930854797363,
+ "step": 1582
+ },
+ {
+ "epoch": 21.991273996509598,
+ "grad_norm": 0.04822581633925438,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 4.089601516723633,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.03507270663976669,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 2.5586814880371094,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.595272958278656,
+ "eval_runtime": 60.7797,
+ "eval_samples_per_second": 40.178,
+ "eval_steps_per_second": 0.642,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013961605584644,
+ "grad_norm": 0.05232533439993858,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 4.011311054229736,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027923211169284,
+ "grad_norm": 0.058515798300504684,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 4.0049638748168945,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041884816753928,
+ "grad_norm": 0.055082619190216064,
+ "learning_rate": 0.000372722041257003,
+ "loss": 3.9904184341430664,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05584642233857,
+ "grad_norm": 0.05213981494307518,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 4.003454685211182,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069808027923212,
+ "grad_norm": 0.051619648933410645,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 3.9900388717651367,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083769633507853,
+ "grad_norm": 0.055212799459695816,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 3.973494052886963,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097731239092496,
+ "grad_norm": 0.05262092128396034,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 4.00933837890625,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111692844677137,
+ "grad_norm": 0.05012505501508713,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 4.01861047744751,
+ "step": 1592
+ },
+ {
+ "epoch": 22.12565445026178,
+ "grad_norm": 0.049657098948955536,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 4.011143684387207,
+ "step": 1593
+ },
+ {
+ "epoch": 22.13961605584642,
+ "grad_norm": 0.05187768116593361,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 3.9922969341278076,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153577661431065,
+ "grad_norm": 0.05053112655878067,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 4.007021903991699,
+ "step": 1595
+ },
+ {
+ "epoch": 22.167539267015705,
+ "grad_norm": 0.048542529344558716,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 3.9966416358947754,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18150087260035,
+ "grad_norm": 0.050972841680049896,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 3.9803225994110107,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195462478184993,
+ "grad_norm": 0.05001165345311165,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 4.0140509605407715,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209424083769633,
+ "grad_norm": 0.05100536718964577,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 3.9906058311462402,
+ "step": 1599
+ },
+ {
+ "epoch": 22.223385689354277,
+ "grad_norm": 0.052977193146944046,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 4.012383460998535,
+ "step": 1600
+ },
+ {
+ "epoch": 22.237347294938917,
+ "grad_norm": 0.051219623535871506,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 3.982931613922119,
+ "step": 1601
+ },
+ {
+ "epoch": 22.25130890052356,
+ "grad_norm": 0.05275029316544533,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 4.032669544219971,
+ "step": 1602
+ },
+ {
+ "epoch": 22.2652705061082,
+ "grad_norm": 0.04781945049762726,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 3.9933102130889893,
+ "step": 1603
+ },
+ {
+ "epoch": 22.279232111692846,
+ "grad_norm": 0.054087597876787186,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 4.0075201988220215,
+ "step": 1604
+ },
+ {
+ "epoch": 22.293193717277486,
+ "grad_norm": 0.05485915020108223,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 4.004299640655518,
+ "step": 1605
+ },
+ {
+ "epoch": 22.30715532286213,
+ "grad_norm": 0.05148882791399956,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 4.044161796569824,
+ "step": 1606
+ },
+ {
+ "epoch": 22.32111692844677,
+ "grad_norm": 0.051912397146224976,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 3.9777069091796875,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335078534031414,
+ "grad_norm": 0.052630942314863205,
+ "learning_rate": 0.000367268105959126,
+ "loss": 4.029580116271973,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349040139616054,
+ "grad_norm": 0.0512809231877327,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 4.022092342376709,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363001745200698,
+ "grad_norm": 0.05094102770090103,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 4.018102169036865,
+ "step": 1610
+ },
+ {
+ "epoch": 22.376963350785342,
+ "grad_norm": 0.04972759634256363,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 4.027868270874023,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390924956369982,
+ "grad_norm": 0.05072534456849098,
+ "learning_rate": 0.000366226612810024,
+ "loss": 3.995988368988037,
+ "step": 1612
+ },
+ {
+ "epoch": 22.404886561954626,
+ "grad_norm": 0.050623781979084015,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 4.019711494445801,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418848167539267,
+ "grad_norm": 0.05035007372498512,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 4.007514953613281,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43280977312391,
+ "grad_norm": 0.054426416754722595,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 4.0167131423950195,
+ "step": 1615
+ },
+ {
+ "epoch": 22.44677137870855,
+ "grad_norm": 0.05365690588951111,
+ "learning_rate": 0.000365184304613104,
+ "loss": 4.0313920974731445,
+ "step": 1616
+ },
+ {
+ "epoch": 22.460732984293195,
+ "grad_norm": 0.0509161539375782,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 4.022063255310059,
+ "step": 1617
+ },
+ {
+ "epoch": 22.474694589877835,
+ "grad_norm": 0.05309660732746124,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 4.017317771911621,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48865619546248,
+ "grad_norm": 0.0488961897790432,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 4.028288841247559,
+ "step": 1619
+ },
+ {
+ "epoch": 22.50261780104712,
+ "grad_norm": 0.04982095956802368,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 4.007115840911865,
+ "step": 1620
+ },
+ {
+ "epoch": 22.516579406631763,
+ "grad_norm": 0.05136589705944061,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 4.028022766113281,
+ "step": 1621
+ },
+ {
+ "epoch": 22.530541012216403,
+ "grad_norm": 0.04865279048681259,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 4.015928268432617,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544502617801047,
+ "grad_norm": 0.05131891369819641,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 4.041997909545898,
+ "step": 1623
+ },
+ {
+ "epoch": 22.55846422338569,
+ "grad_norm": 0.05063068866729736,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 4.042117118835449,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57242582897033,
+ "grad_norm": 0.049348752945661545,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 4.012609481811523,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586387434554975,
+ "grad_norm": 0.05123874917626381,
+ "learning_rate": 0.000362575056487632,
+ "loss": 4.03358793258667,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600349040139616,
+ "grad_norm": 0.04921206086874008,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 4.039857864379883,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61431064572426,
+ "grad_norm": 0.05011922866106033,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 4.015835762023926,
+ "step": 1628
+ },
+ {
+ "epoch": 22.6282722513089,
+ "grad_norm": 0.049950532615184784,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 4.068299293518066,
+ "step": 1629
+ },
+ {
+ "epoch": 22.642233856893544,
+ "grad_norm": 0.04962172359228134,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 4.03205680847168,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656195462478184,
+ "grad_norm": 0.05059773847460747,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 4.065027236938477,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670157068062828,
+ "grad_norm": 0.04986768960952759,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 4.01540470123291,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68411867364747,
+ "grad_norm": 0.050884976983070374,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 4.060809135437012,
+ "step": 1633
+ },
+ {
+ "epoch": 22.698080279232112,
+ "grad_norm": 0.052478861063718796,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 4.088409423828125,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712041884816752,
+ "grad_norm": 0.05118529871106148,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 4.071645736694336,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726003490401396,
+ "grad_norm": 0.05059075728058815,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 4.028438568115234,
+ "step": 1636
+ },
+ {
+ "epoch": 22.739965095986037,
+ "grad_norm": 0.049222562462091446,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 4.02166223526001,
+ "step": 1637
+ },
+ {
+ "epoch": 22.75392670157068,
+ "grad_norm": 0.049673549830913544,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 4.0511860847473145,
+ "step": 1638
+ },
+ {
+ "epoch": 22.767888307155324,
+ "grad_norm": 0.04954935610294342,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 4.036063194274902,
+ "step": 1639
+ },
+ {
+ "epoch": 22.781849912739965,
+ "grad_norm": 0.05110877379775047,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 4.081101417541504,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79581151832461,
+ "grad_norm": 0.049009695649147034,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 4.042314529418945,
+ "step": 1641
+ },
+ {
+ "epoch": 22.80977312390925,
+ "grad_norm": 0.0496281161904335,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 4.028315544128418,
+ "step": 1642
+ },
+ {
+ "epoch": 22.823734729493893,
+ "grad_norm": 0.04931195452809334,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 4.018457412719727,
+ "step": 1643
+ },
+ {
+ "epoch": 22.837696335078533,
+ "grad_norm": 0.05006880313158035,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 4.0709991455078125,
+ "step": 1644
+ },
+ {
+ "epoch": 22.851657940663177,
+ "grad_norm": 0.04956946149468422,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 4.065585136413574,
+ "step": 1645
+ },
+ {
+ "epoch": 22.865619546247817,
+ "grad_norm": 0.048182398080825806,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 4.0522589683532715,
+ "step": 1646
+ },
+ {
+ "epoch": 22.87958115183246,
+ "grad_norm": 0.04964807257056236,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 4.06989860534668,
+ "step": 1647
+ },
+ {
+ "epoch": 22.8935427574171,
+ "grad_norm": 0.049299128353595734,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 4.082858085632324,
+ "step": 1648
+ },
+ {
+ "epoch": 22.907504363001745,
+ "grad_norm": 0.04893320053815842,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 4.047144889831543,
+ "step": 1649
+ },
+ {
+ "epoch": 22.921465968586386,
+ "grad_norm": 0.04772866889834404,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 4.043060302734375,
+ "step": 1650
+ },
+ {
+ "epoch": 22.93542757417103,
+ "grad_norm": 0.049417730420827866,
+ "learning_rate": 0.000356031397755195,
+ "loss": 4.031186580657959,
+ "step": 1651
+ },
+ {
+ "epoch": 22.949389179755673,
+ "grad_norm": 0.049944665282964706,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 4.074260711669922,
+ "step": 1652
+ },
+ {
+ "epoch": 22.963350785340314,
+ "grad_norm": 0.04856271669268608,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 4.054103851318359,
+ "step": 1653
+ },
+ {
+ "epoch": 22.977312390924958,
+ "grad_norm": 0.0470513179898262,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 4.041568756103516,
+ "step": 1654
+ },
+ {
+ "epoch": 22.991273996509598,
+ "grad_norm": 0.049018729478120804,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 4.066161632537842,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.034226156771183014,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 2.5147223472595215,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.5970026850700378,
+ "eval_runtime": 60.0787,
+ "eval_samples_per_second": 40.647,
+ "eval_steps_per_second": 0.649,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013961605584644,
+ "grad_norm": 0.05091085657477379,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 3.95741868019104,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027923211169284,
+ "grad_norm": 0.05874285474419594,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 3.966327667236328,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041884816753928,
+ "grad_norm": 0.05394266918301582,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.937232732772827,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05584642233857,
+ "grad_norm": 0.05120540037751198,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 3.9637370109558105,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069808027923212,
+ "grad_norm": 0.05310065299272537,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 3.998631477355957,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083769633507853,
+ "grad_norm": 0.05127466097474098,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.949101448059082,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097731239092496,
+ "grad_norm": 0.05201594531536102,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 3.986267566680908,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111692844677137,
+ "grad_norm": 0.051927369087934494,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 3.9604103565216064,
+ "step": 1664
+ },
+ {
+ "epoch": 23.12565445026178,
+ "grad_norm": 0.05125141143798828,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.9339547157287598,
+ "step": 1665
+ },
+ {
+ "epoch": 23.13961605584642,
+ "grad_norm": 0.05241088196635246,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 3.971261978149414,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153577661431065,
+ "grad_norm": 0.050563592463731766,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.9450082778930664,
+ "step": 1667
+ },
+ {
+ "epoch": 23.167539267015705,
+ "grad_norm": 0.05204763635993004,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 3.9696707725524902,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18150087260035,
+ "grad_norm": 0.05518610402941704,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 3.9816057682037354,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195462478184993,
+ "grad_norm": 0.05336211621761322,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.932252883911133,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209424083769633,
+ "grad_norm": 0.052204642444849014,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 3.9818131923675537,
+ "step": 1671
+ },
+ {
+ "epoch": 23.223385689354277,
+ "grad_norm": 0.05299545079469681,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.9488978385925293,
+ "step": 1672
+ },
+ {
+ "epoch": 23.237347294938917,
+ "grad_norm": 0.05143733322620392,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 4.00959587097168,
+ "step": 1673
+ },
+ {
+ "epoch": 23.25130890052356,
+ "grad_norm": 0.05186324194073677,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.9529542922973633,
+ "step": 1674
+ },
+ {
+ "epoch": 23.2652705061082,
+ "grad_norm": 0.05334410071372986,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 3.96693754196167,
+ "step": 1675
+ },
+ {
+ "epoch": 23.279232111692846,
+ "grad_norm": 0.05345796048641205,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 3.9674174785614014,
+ "step": 1676
+ },
+ {
+ "epoch": 23.293193717277486,
+ "grad_norm": 0.050988901406526566,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 3.9698638916015625,
+ "step": 1677
+ },
+ {
+ "epoch": 23.30715532286213,
+ "grad_norm": 0.056467678397893906,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 3.9645955562591553,
+ "step": 1678
+ },
+ {
+ "epoch": 23.32111692844677,
+ "grad_norm": 0.05554647371172905,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 4.0036091804504395,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335078534031414,
+ "grad_norm": 0.05607818067073822,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 3.973773717880249,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349040139616054,
+ "grad_norm": 0.05734477937221527,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 3.987692356109619,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363001745200698,
+ "grad_norm": 0.052175361663103104,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 3.9761803150177,
+ "step": 1682
+ },
+ {
+ "epoch": 23.376963350785342,
+ "grad_norm": 0.053368814289569855,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 3.965020179748535,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390924956369982,
+ "grad_norm": 0.05505364388227463,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 3.987494468688965,
+ "step": 1684
+ },
+ {
+ "epoch": 23.404886561954626,
+ "grad_norm": 0.0524514764547348,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 3.9624319076538086,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418848167539267,
+ "grad_norm": 0.05681568384170532,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 4.008373737335205,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43280977312391,
+ "grad_norm": 0.05344909429550171,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 3.9840354919433594,
+ "step": 1687
+ },
+ {
+ "epoch": 23.44677137870855,
+ "grad_norm": 0.052657466381788254,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 3.992262363433838,
+ "step": 1688
+ },
+ {
+ "epoch": 23.460732984293195,
+ "grad_norm": 0.05526747554540634,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 3.97529935836792,
+ "step": 1689
+ },
+ {
+ "epoch": 23.474694589877835,
+ "grad_norm": 0.05492551624774933,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 4.004708766937256,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48865619546248,
+ "grad_norm": 0.053742099553346634,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 3.994016647338867,
+ "step": 1691
+ },
+ {
+ "epoch": 23.50261780104712,
+ "grad_norm": 0.05363722890615463,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 3.9942378997802734,
+ "step": 1692
+ },
+ {
+ "epoch": 23.516579406631763,
+ "grad_norm": 0.05167972669005394,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 3.998253345489502,
+ "step": 1693
+ },
+ {
+ "epoch": 23.530541012216403,
+ "grad_norm": 0.0518590584397316,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 3.9850645065307617,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544502617801047,
+ "grad_norm": 0.05521136894822121,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 4.02669620513916,
+ "step": 1695
+ },
+ {
+ "epoch": 23.55846422338569,
+ "grad_norm": 0.054986029863357544,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 4.022308826446533,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57242582897033,
+ "grad_norm": 0.05405082181096077,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 4.021487236022949,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586387434554975,
+ "grad_norm": 0.051257841289043427,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 3.9724135398864746,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600349040139616,
+ "grad_norm": 0.05109119415283203,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 3.978914499282837,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61431064572426,
+ "grad_norm": 0.05303420498967171,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 4.027509689331055,
+ "step": 1700
+ },
+ {
+ "epoch": 23.6282722513089,
+ "grad_norm": 0.055543459951877594,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 3.9750852584838867,
+ "step": 1701
+ },
+ {
+ "epoch": 23.642233856893544,
+ "grad_norm": 0.05068540200591087,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 4.029600143432617,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656195462478184,
+ "grad_norm": 0.04991193488240242,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 3.9901509284973145,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670157068062828,
+ "grad_norm": 0.05403713881969452,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 4.017679214477539,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68411867364747,
+ "grad_norm": 0.05203867331147194,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 4.008406639099121,
+ "step": 1705
+ },
+ {
+ "epoch": 23.698080279232112,
+ "grad_norm": 0.05106597766280174,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 4.009983062744141,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712041884816752,
+ "grad_norm": 0.04976251348853111,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 4.007230281829834,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726003490401396,
+ "grad_norm": 0.05199650675058365,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 3.9902141094207764,
+ "step": 1708
+ },
+ {
+ "epoch": 23.739965095986037,
+ "grad_norm": 0.05285118520259857,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 3.9966492652893066,
+ "step": 1709
+ },
+ {
+ "epoch": 23.75392670157068,
+ "grad_norm": 0.051937371492385864,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 3.962397336959839,
+ "step": 1710
+ },
+ {
+ "epoch": 23.767888307155324,
+ "grad_norm": 0.05179062858223915,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 4.027141571044922,
+ "step": 1711
+ },
+ {
+ "epoch": 23.781849912739965,
+ "grad_norm": 0.049667056649923325,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 4.024991512298584,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79581151832461,
+ "grad_norm": 0.05204658955335617,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 4.002647876739502,
+ "step": 1713
+ },
+ {
+ "epoch": 23.80977312390925,
+ "grad_norm": 0.05069981515407562,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 4.035422325134277,
+ "step": 1714
+ },
+ {
+ "epoch": 23.823734729493893,
+ "grad_norm": 0.05044785887002945,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 4.042465686798096,
+ "step": 1715
+ },
+ {
+ "epoch": 23.837696335078533,
+ "grad_norm": 0.053949885070323944,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 4.002717971801758,
+ "step": 1716
+ },
+ {
+ "epoch": 23.851657940663177,
+ "grad_norm": 0.05139258876442909,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 4.026609897613525,
+ "step": 1717
+ },
+ {
+ "epoch": 23.865619546247817,
+ "grad_norm": 0.05316372960805893,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 4.019758224487305,
+ "step": 1718
+ },
+ {
+ "epoch": 23.87958115183246,
+ "grad_norm": 0.050791382789611816,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 3.9859633445739746,
+ "step": 1719
+ },
+ {
+ "epoch": 23.8935427574171,
+ "grad_norm": 0.05124847590923309,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 4.038055419921875,
+ "step": 1720
+ },
+ {
+ "epoch": 23.907504363001745,
+ "grad_norm": 0.0495096817612648,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 4.02764368057251,
+ "step": 1721
+ },
+ {
+ "epoch": 23.921465968586386,
+ "grad_norm": 0.05205567181110382,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 4.003867149353027,
+ "step": 1722
+ },
+ {
+ "epoch": 23.93542757417103,
+ "grad_norm": 0.05233347788453102,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 4.024157524108887,
+ "step": 1723
+ },
+ {
+ "epoch": 23.949389179755673,
+ "grad_norm": 0.05126349255442619,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 4.05349063873291,
+ "step": 1724
+ },
+ {
+ "epoch": 23.963350785340314,
+ "grad_norm": 0.050006967037916183,
+ "learning_rate": 0.000336518346307424,
+ "loss": 4.024704933166504,
+ "step": 1725
+ },
+ {
+ "epoch": 23.977312390924958,
+ "grad_norm": 0.052343279123306274,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 4.008549213409424,
+ "step": 1726
+ },
+ {
+ "epoch": 23.991273996509598,
+ "grad_norm": 0.05126652494072914,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 4.002162933349609,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.03560106083750725,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 2.511049270629883,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.5985628962516785,
+ "eval_runtime": 59.0712,
+ "eval_samples_per_second": 41.34,
+ "eval_steps_per_second": 0.66,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013961605584644,
+ "grad_norm": 0.05205906555056572,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.936516284942627,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027923211169284,
+ "grad_norm": 0.056368257850408554,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.95200777053833,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041884816753928,
+ "grad_norm": 0.05075085908174515,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.940810441970825,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05584642233857,
+ "grad_norm": 0.05119713768362999,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.90130615234375,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069808027923212,
+ "grad_norm": 0.05254847928881645,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.9491095542907715,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083769633507853,
+ "grad_norm": 0.053025662899017334,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.9371917247772217,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097731239092496,
+ "grad_norm": 0.05402665585279465,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 3.9651317596435547,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111692844677137,
+ "grad_norm": 0.0519571453332901,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.9189062118530273,
+ "step": 1736
+ },
+ {
+ "epoch": 24.12565445026178,
+ "grad_norm": 0.05436019226908684,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.948184013366699,
+ "step": 1737
+ },
+ {
+ "epoch": 24.13961605584642,
+ "grad_norm": 0.056572459638118744,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.9076738357543945,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153577661431065,
+ "grad_norm": 0.05553183704614639,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.933659076690674,
+ "step": 1739
+ },
+ {
+ "epoch": 24.167539267015705,
+ "grad_norm": 0.05554374307394028,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.941751480102539,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18150087260035,
+ "grad_norm": 0.05451614409685135,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.928593635559082,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195462478184993,
+ "grad_norm": 0.055458780378103256,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 3.961820602416992,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209424083769633,
+ "grad_norm": 0.054980456829071045,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 3.9736528396606445,
+ "step": 1743
+ },
+ {
+ "epoch": 24.223385689354277,
+ "grad_norm": 0.05538604408502579,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.935093641281128,
+ "step": 1744
+ },
+ {
+ "epoch": 24.237347294938917,
+ "grad_norm": 0.05580678582191467,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.9240689277648926,
+ "step": 1745
+ },
+ {
+ "epoch": 24.25130890052356,
+ "grad_norm": 0.052294690161943436,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.937718152999878,
+ "step": 1746
+ },
+ {
+ "epoch": 24.2652705061082,
+ "grad_norm": 0.052018795162439346,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.942049503326416,
+ "step": 1747
+ },
+ {
+ "epoch": 24.279232111692846,
+ "grad_norm": 0.0550091527402401,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.9215564727783203,
+ "step": 1748
+ },
+ {
+ "epoch": 24.293193717277486,
+ "grad_norm": 0.05447716638445854,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.9154953956604004,
+ "step": 1749
+ },
+ {
+ "epoch": 24.30715532286213,
+ "grad_norm": 0.05300029739737511,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 3.9642622470855713,
+ "step": 1750
+ },
+ {
+ "epoch": 24.32111692844677,
+ "grad_norm": 0.05312013998627663,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.938897132873535,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335078534031414,
+ "grad_norm": 0.053737837821245193,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.94972562789917,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349040139616054,
+ "grad_norm": 0.05087469890713692,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.9301164150238037,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363001745200698,
+ "grad_norm": 0.05179893225431442,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.9393205642700195,
+ "step": 1754
+ },
+ {
+ "epoch": 24.376963350785342,
+ "grad_norm": 0.05277438461780548,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 3.960968494415283,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390924956369982,
+ "grad_norm": 0.05090474337339401,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 3.9516637325286865,
+ "step": 1756
+ },
+ {
+ "epoch": 24.404886561954626,
+ "grad_norm": 0.052786506712436676,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.9332518577575684,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418848167539267,
+ "grad_norm": 0.052095942199230194,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.9108662605285645,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43280977312391,
+ "grad_norm": 0.05310450494289398,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 3.9746041297912598,
+ "step": 1759
+ },
+ {
+ "epoch": 24.44677137870855,
+ "grad_norm": 0.05078566074371338,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.93316650390625,
+ "step": 1760
+ },
+ {
+ "epoch": 24.460732984293195,
+ "grad_norm": 0.0519949309527874,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.9489870071411133,
+ "step": 1761
+ },
+ {
+ "epoch": 24.474694589877835,
+ "grad_norm": 0.05239349976181984,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 3.955118417739868,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48865619546248,
+ "grad_norm": 0.05193360522389412,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 3.9707555770874023,
+ "step": 1763
+ },
+ {
+ "epoch": 24.50261780104712,
+ "grad_norm": 0.05397362262010574,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 3.973053455352783,
+ "step": 1764
+ },
+ {
+ "epoch": 24.516579406631763,
+ "grad_norm": 0.05379117652773857,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 3.957289695739746,
+ "step": 1765
+ },
+ {
+ "epoch": 24.530541012216403,
+ "grad_norm": 0.0526256300508976,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 3.9595577716827393,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544502617801047,
+ "grad_norm": 0.05377693101763725,
+ "learning_rate": 0.000325372061241796,
+ "loss": 3.97849178314209,
+ "step": 1767
+ },
+ {
+ "epoch": 24.55846422338569,
+ "grad_norm": 0.05182809755206108,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 3.9714043140411377,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57242582897033,
+ "grad_norm": 0.05364793911576271,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 3.9582200050354004,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586387434554975,
+ "grad_norm": 0.051558300852775574,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 3.9747936725616455,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600349040139616,
+ "grad_norm": 0.053487639874219894,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 3.9556500911712646,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61431064572426,
+ "grad_norm": 0.05146970972418785,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.94858455657959,
+ "step": 1772
+ },
+ {
+ "epoch": 24.6282722513089,
+ "grad_norm": 0.05099836364388466,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 3.9716954231262207,
+ "step": 1773
+ },
+ {
+ "epoch": 24.642233856893544,
+ "grad_norm": 0.054634734988212585,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 3.974881887435913,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656195462478184,
+ "grad_norm": 0.054601747542619705,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 3.9746241569519043,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670157068062828,
+ "grad_norm": 0.052188221365213394,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 3.957106113433838,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68411867364747,
+ "grad_norm": 0.055673446506261826,
+ "learning_rate": 0.000322712903929477,
+ "loss": 3.9803967475891113,
+ "step": 1777
+ },
+ {
+ "epoch": 24.698080279232112,
+ "grad_norm": 0.05397328734397888,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 3.967238426208496,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712041884816752,
+ "grad_norm": 0.05384642630815506,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 3.949392795562744,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726003490401396,
+ "grad_norm": 0.05452995374798775,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.9432482719421387,
+ "step": 1780
+ },
+ {
+ "epoch": 24.739965095986037,
+ "grad_norm": 0.052430473268032074,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 3.9944653511047363,
+ "step": 1781
+ },
+ {
+ "epoch": 24.75392670157068,
+ "grad_norm": 0.05514240264892578,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 3.97713565826416,
+ "step": 1782
+ },
+ {
+ "epoch": 24.767888307155324,
+ "grad_norm": 0.05189919099211693,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 3.983150005340576,
+ "step": 1783
+ },
+ {
+ "epoch": 24.781849912739965,
+ "grad_norm": 0.05590776726603508,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 3.9548301696777344,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79581151832461,
+ "grad_norm": 0.05564190074801445,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 3.9725170135498047,
+ "step": 1785
+ },
+ {
+ "epoch": 24.80977312390925,
+ "grad_norm": 0.05121326446533203,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 3.9722347259521484,
+ "step": 1786
+ },
+ {
+ "epoch": 24.823734729493893,
+ "grad_norm": 0.05444202199578285,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 3.9634029865264893,
+ "step": 1787
+ },
+ {
+ "epoch": 24.837696335078533,
+ "grad_norm": 0.05242442339658737,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 3.973270893096924,
+ "step": 1788
+ },
+ {
+ "epoch": 24.851657940663177,
+ "grad_norm": 0.05497463420033455,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 3.9851887226104736,
+ "step": 1789
+ },
+ {
+ "epoch": 24.865619546247817,
+ "grad_norm": 0.05482025071978569,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 4.001110076904297,
+ "step": 1790
+ },
+ {
+ "epoch": 24.87958115183246,
+ "grad_norm": 0.054101862013339996,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 3.9782891273498535,
+ "step": 1791
+ },
+ {
+ "epoch": 24.8935427574171,
+ "grad_norm": 0.05538944527506828,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 3.973541498184204,
+ "step": 1792
+ },
+ {
+ "epoch": 24.907504363001745,
+ "grad_norm": 0.0502893440425396,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 3.9875426292419434,
+ "step": 1793
+ },
+ {
+ "epoch": 24.921465968586386,
+ "grad_norm": 0.05576152354478836,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.947035551071167,
+ "step": 1794
+ },
+ {
+ "epoch": 24.93542757417103,
+ "grad_norm": 0.05356278270483017,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 3.9683666229248047,
+ "step": 1795
+ },
+ {
+ "epoch": 24.949389179755673,
+ "grad_norm": 0.05377693101763725,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 3.9973440170288086,
+ "step": 1796
+ },
+ {
+ "epoch": 24.963350785340314,
+ "grad_norm": 0.05334382876753807,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 3.9843404293060303,
+ "step": 1797
+ },
+ {
+ "epoch": 24.977312390924958,
+ "grad_norm": 0.05514024198055267,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 3.9920568466186523,
+ "step": 1798
+ },
+ {
+ "epoch": 24.991273996509598,
+ "grad_norm": 0.05631054937839508,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 4.005859375,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.03653077408671379,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 2.503864049911499,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 0.5995591878890991,
+ "eval_runtime": 59.5213,
+ "eval_samples_per_second": 41.027,
+ "eval_steps_per_second": 0.655,
+ "step": 1800
+ },
+ {
+ "epoch": 25.013961605584644,
+ "grad_norm": 0.0576021745800972,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 3.894834518432617,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027923211169284,
+ "grad_norm": 0.05842140316963196,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 3.8879270553588867,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041884816753928,
+ "grad_norm": 0.051821205765008926,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 3.9011340141296387,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05584642233857,
+ "grad_norm": 0.053872160613536835,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 3.877183437347412,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069808027923212,
+ "grad_norm": 0.05487574636936188,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 3.9126815795898438,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083769633507853,
+ "grad_norm": 0.054469428956508636,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 3.894768238067627,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097731239092496,
+ "grad_norm": 0.05124446377158165,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 3.916809320449829,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111692844677137,
+ "grad_norm": 0.05193263292312622,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 3.917140483856201,
+ "step": 1808
+ },
+ {
+ "epoch": 25.12565445026178,
+ "grad_norm": 0.0519125871360302,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 3.903704881668091,
+ "step": 1809
+ },
+ {
+ "epoch": 25.13961605584642,
+ "grad_norm": 0.052205223590135574,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 3.892639636993408,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153577661431065,
+ "grad_norm": 0.052036914974451065,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 3.895559787750244,
+ "step": 1811
+ },
+ {
+ "epoch": 25.167539267015705,
+ "grad_norm": 0.054616156965494156,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 3.8976142406463623,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18150087260035,
+ "grad_norm": 0.05603514611721039,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 3.914311170578003,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195462478184993,
+ "grad_norm": 0.05170242115855217,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 3.9141643047332764,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209424083769633,
+ "grad_norm": 0.05476640164852142,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 3.8952436447143555,
+ "step": 1815
+ },
+ {
+ "epoch": 25.223385689354277,
+ "grad_norm": 0.055114105343818665,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 3.872978687286377,
+ "step": 1816
+ },
+ {
+ "epoch": 25.237347294938917,
+ "grad_norm": 0.051593828946352005,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 3.893911600112915,
+ "step": 1817
+ },
+ {
+ "epoch": 25.25130890052356,
+ "grad_norm": 0.05363611876964569,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 3.9062986373901367,
+ "step": 1818
+ },
+ {
+ "epoch": 25.2652705061082,
+ "grad_norm": 0.054656051099300385,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 3.9145383834838867,
+ "step": 1819
+ },
+ {
+ "epoch": 25.279232111692846,
+ "grad_norm": 0.05251947045326233,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 3.912965774536133,
+ "step": 1820
+ },
+ {
+ "epoch": 25.293193717277486,
+ "grad_norm": 0.05575991049408913,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 3.901510238647461,
+ "step": 1821
+ },
+ {
+ "epoch": 25.30715532286213,
+ "grad_norm": 0.05540676787495613,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 3.8659887313842773,
+ "step": 1822
+ },
+ {
+ "epoch": 25.32111692844677,
+ "grad_norm": 0.0577131025493145,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 3.884077548980713,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335078534031414,
+ "grad_norm": 0.055947501212358475,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 3.9274377822875977,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349040139616054,
+ "grad_norm": 0.05328478664159775,
+ "learning_rate": 0.000309930621834911,
+ "loss": 3.8783111572265625,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363001745200698,
+ "grad_norm": 0.057696059346199036,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 3.903721570968628,
+ "step": 1826
+ },
+ {
+ "epoch": 25.376963350785342,
+ "grad_norm": 0.05717543140053749,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 3.9259400367736816,
+ "step": 1827
+ },
+ {
+ "epoch": 25.390924956369982,
+ "grad_norm": 0.05328287556767464,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 3.9450511932373047,
+ "step": 1828
+ },
+ {
+ "epoch": 25.404886561954626,
+ "grad_norm": 0.05414140224456787,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 3.9231765270233154,
+ "step": 1829
+ },
+ {
+ "epoch": 25.418848167539267,
+ "grad_norm": 0.05535200983285904,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 3.8928630352020264,
+ "step": 1830
+ },
+ {
+ "epoch": 25.43280977312391,
+ "grad_norm": 0.05468215420842171,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 3.933401584625244,
+ "step": 1831
+ },
+ {
+ "epoch": 25.44677137870855,
+ "grad_norm": 0.056624677032232285,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 3.922624111175537,
+ "step": 1832
+ },
+ {
+ "epoch": 25.460732984293195,
+ "grad_norm": 0.0545354001224041,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 3.916515827178955,
+ "step": 1833
+ },
+ {
+ "epoch": 25.474694589877835,
+ "grad_norm": 0.053979165852069855,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 3.891812324523926,
+ "step": 1834
+ },
+ {
+ "epoch": 25.48865619546248,
+ "grad_norm": 0.05619875714182854,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 3.959524631500244,
+ "step": 1835
+ },
+ {
+ "epoch": 25.50261780104712,
+ "grad_norm": 0.053561773151159286,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 3.9109320640563965,
+ "step": 1836
+ },
+ {
+ "epoch": 25.516579406631763,
+ "grad_norm": 0.05566735938191414,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 3.910492420196533,
+ "step": 1837
+ },
+ {
+ "epoch": 25.530541012216403,
+ "grad_norm": 0.05757030472159386,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 3.9716763496398926,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544502617801047,
+ "grad_norm": 0.05121897906064987,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 3.9114646911621094,
+ "step": 1839
+ },
+ {
+ "epoch": 25.55846422338569,
+ "grad_norm": 0.05664774775505066,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 3.9465560913085938,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57242582897033,
+ "grad_norm": 0.0569215826690197,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 3.908374309539795,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586387434554975,
+ "grad_norm": 0.05345458909869194,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 3.9336414337158203,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600349040139616,
+ "grad_norm": 0.0565933920443058,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 3.9478759765625,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61431064572426,
+ "grad_norm": 0.05887870490550995,
+ "learning_rate": 0.000304866093757771,
+ "loss": 3.9018349647521973,
+ "step": 1844
+ },
+ {
+ "epoch": 25.6282722513089,
+ "grad_norm": 0.05515308678150177,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 3.9446825981140137,
+ "step": 1845
+ },
+ {
+ "epoch": 25.642233856893544,
+ "grad_norm": 0.058602213859558105,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 3.9497885704040527,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656195462478184,
+ "grad_norm": 0.053806282579898834,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 3.9247193336486816,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670157068062828,
+ "grad_norm": 0.05495952442288399,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 3.912296772003174,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68411867364747,
+ "grad_norm": 0.05479473993182182,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 3.9471306800842285,
+ "step": 1849
+ },
+ {
+ "epoch": 25.698080279232112,
+ "grad_norm": 0.053319111466407776,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 3.955942153930664,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712041884816752,
+ "grad_norm": 0.05832696333527565,
+ "learning_rate": 0.000303,
+ "loss": 3.959169626235962,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726003490401396,
+ "grad_norm": 0.057019371539354324,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 3.9682188034057617,
+ "step": 1852
+ },
+ {
+ "epoch": 25.739965095986037,
+ "grad_norm": 0.05099530518054962,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 3.921377658843994,
+ "step": 1853
+ },
+ {
+ "epoch": 25.75392670157068,
+ "grad_norm": 0.05466439574956894,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 3.946341037750244,
+ "step": 1854
+ },
+ {
+ "epoch": 25.767888307155324,
+ "grad_norm": 0.05742308124899864,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 3.948031425476074,
+ "step": 1855
+ },
+ {
+ "epoch": 25.781849912739965,
+ "grad_norm": 0.057581741362810135,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 3.91162109375,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79581151832461,
+ "grad_norm": 0.05520833283662796,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 3.937495231628418,
+ "step": 1857
+ },
+ {
+ "epoch": 25.80977312390925,
+ "grad_norm": 0.05763327330350876,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 3.9565534591674805,
+ "step": 1858
+ },
+ {
+ "epoch": 25.823734729493893,
+ "grad_norm": 0.0544707253575325,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 3.965190887451172,
+ "step": 1859
+ },
+ {
+ "epoch": 25.837696335078533,
+ "grad_norm": 0.05445104092359543,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 3.938063859939575,
+ "step": 1860
+ },
+ {
+ "epoch": 25.851657940663177,
+ "grad_norm": 0.05505113676190376,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 3.928392171859741,
+ "step": 1861
+ },
+ {
+ "epoch": 25.865619546247817,
+ "grad_norm": 0.05574212223291397,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 3.959397077560425,
+ "step": 1862
+ },
+ {
+ "epoch": 25.87958115183246,
+ "grad_norm": 0.053950466215610504,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 3.9532437324523926,
+ "step": 1863
+ },
+ {
+ "epoch": 25.8935427574171,
+ "grad_norm": 0.05495546758174896,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 3.946746826171875,
+ "step": 1864
+ },
+ {
+ "epoch": 25.907504363001745,
+ "grad_norm": 0.05328625068068504,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 3.9494776725769043,
+ "step": 1865
+ },
+ {
+ "epoch": 25.921465968586386,
+ "grad_norm": 0.05593104660511017,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 3.959989547729492,
+ "step": 1866
+ },
+ {
+ "epoch": 25.93542757417103,
+ "grad_norm": 0.05442969873547554,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 3.952371597290039,
+ "step": 1867
+ },
+ {
+ "epoch": 25.949389179755673,
+ "grad_norm": 0.05446415767073631,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 3.9538888931274414,
+ "step": 1868
+ },
+ {
+ "epoch": 25.963350785340314,
+ "grad_norm": 0.054761797189712524,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 3.9386215209960938,
+ "step": 1869
+ },
+ {
+ "epoch": 25.977312390924958,
+ "grad_norm": 0.054201845079660416,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 3.9252963066101074,
+ "step": 1870
+ },
+ {
+ "epoch": 25.991273996509598,
+ "grad_norm": 0.05364517495036125,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 3.9733197689056396,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.038914378732442856,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 2.4261250495910645,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 0.6009622812271118,
+ "eval_runtime": 60.6864,
+ "eval_samples_per_second": 40.24,
+ "eval_steps_per_second": 0.643,
+ "step": 1872
+ },
+ {
+ "epoch": 26.013961605584644,
+ "grad_norm": 0.05345558747649193,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 3.8902673721313477,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027923211169284,
+ "grad_norm": 0.05672457069158554,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 3.8676247596740723,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041884816753928,
+ "grad_norm": 0.051981374621391296,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 3.872048854827881,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05584642233857,
+ "grad_norm": 0.05262405797839165,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 3.8582329750061035,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069808027923212,
+ "grad_norm": 0.054473262280225754,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 3.8352112770080566,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083769633507853,
+ "grad_norm": 0.053167302161455154,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 3.881760835647583,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097731239092496,
+ "grad_norm": 0.053720805794000626,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 3.891775608062744,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111692844677137,
+ "grad_norm": 0.055102936923503876,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 3.8872156143188477,
+ "step": 1880
+ },
+ {
+ "epoch": 26.12565445026178,
+ "grad_norm": 0.05108853057026863,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 3.8625407218933105,
+ "step": 1881
+ },
+ {
+ "epoch": 26.13961605584642,
+ "grad_norm": 0.05526706948876381,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 3.886127233505249,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153577661431065,
+ "grad_norm": 0.05458226427435875,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 3.8722105026245117,
+ "step": 1883
+ },
+ {
+ "epoch": 26.167539267015705,
+ "grad_norm": 0.05342113599181175,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 3.853677749633789,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18150087260035,
+ "grad_norm": 0.053665641695261,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 3.849315881729126,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195462478184993,
+ "grad_norm": 0.054922688752412796,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 3.8599777221679688,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209424083769633,
+ "grad_norm": 0.05412537604570389,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 3.8767571449279785,
+ "step": 1887
+ },
+ {
+ "epoch": 26.223385689354277,
+ "grad_norm": 0.05467838793992996,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 3.8436965942382812,
+ "step": 1888
+ },
+ {
+ "epoch": 26.237347294938917,
+ "grad_norm": 0.054886411875486374,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 3.8796300888061523,
+ "step": 1889
+ },
+ {
+ "epoch": 26.25130890052356,
+ "grad_norm": 0.05437029153108597,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 3.858133316040039,
+ "step": 1890
+ },
+ {
+ "epoch": 26.2652705061082,
+ "grad_norm": 0.05273836478590965,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 3.839015483856201,
+ "step": 1891
+ },
+ {
+ "epoch": 26.279232111692846,
+ "grad_norm": 0.0546451173722744,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 3.8723373413085938,
+ "step": 1892
+ },
+ {
+ "epoch": 26.293193717277486,
+ "grad_norm": 0.05240490287542343,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 3.877549886703491,
+ "step": 1893
+ },
+ {
+ "epoch": 26.30715532286213,
+ "grad_norm": 0.05420667678117752,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 3.8643171787261963,
+ "step": 1894
+ },
+ {
+ "epoch": 26.32111692844677,
+ "grad_norm": 0.053440142422914505,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 3.8681447505950928,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335078534031414,
+ "grad_norm": 0.052396632730960846,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 3.8691887855529785,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349040139616054,
+ "grad_norm": 0.05318925902247429,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 3.897003650665283,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363001745200698,
+ "grad_norm": 0.05328533053398132,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 3.9079437255859375,
+ "step": 1898
+ },
+ {
+ "epoch": 26.376963350785342,
+ "grad_norm": 0.05196928232908249,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 3.8392016887664795,
+ "step": 1899
+ },
+ {
+ "epoch": 26.390924956369982,
+ "grad_norm": 0.053326766937971115,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 3.8930301666259766,
+ "step": 1900
+ },
+ {
+ "epoch": 26.404886561954626,
+ "grad_norm": 0.052926234900951385,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 3.866997718811035,
+ "step": 1901
+ },
+ {
+ "epoch": 26.418848167539267,
+ "grad_norm": 0.05289287120103836,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 3.896690607070923,
+ "step": 1902
+ },
+ {
+ "epoch": 26.43280977312391,
+ "grad_norm": 0.052162300795316696,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 3.8642401695251465,
+ "step": 1903
+ },
+ {
+ "epoch": 26.44677137870855,
+ "grad_norm": 0.05495629832148552,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 3.8982696533203125,
+ "step": 1904
+ },
+ {
+ "epoch": 26.460732984293195,
+ "grad_norm": 0.055102866142988205,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 3.8725972175598145,
+ "step": 1905
+ },
+ {
+ "epoch": 26.474694589877835,
+ "grad_norm": 0.05537770688533783,
+ "learning_rate": 0.000288343693342466,
+ "loss": 3.9050111770629883,
+ "step": 1906
+ },
+ {
+ "epoch": 26.48865619546248,
+ "grad_norm": 0.05490661785006523,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 3.90586256980896,
+ "step": 1907
+ },
+ {
+ "epoch": 26.50261780104712,
+ "grad_norm": 0.054393451660871506,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 3.8841705322265625,
+ "step": 1908
+ },
+ {
+ "epoch": 26.516579406631763,
+ "grad_norm": 0.0553058385848999,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 3.883500576019287,
+ "step": 1909
+ },
+ {
+ "epoch": 26.530541012216403,
+ "grad_norm": 0.053976885974407196,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 3.894948959350586,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544502617801047,
+ "grad_norm": 0.05290692299604416,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 3.8880579471588135,
+ "step": 1911
+ },
+ {
+ "epoch": 26.55846422338569,
+ "grad_norm": 0.05570613220334053,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 3.9035487174987793,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57242582897033,
+ "grad_norm": 0.054136309772729874,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 3.9137563705444336,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586387434554975,
+ "grad_norm": 0.05328850820660591,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 3.87838077545166,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600349040139616,
+ "grad_norm": 0.05169905349612236,
+ "learning_rate": 0.000285947841605349,
+ "loss": 3.882963180541992,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61431064572426,
+ "grad_norm": 0.05304642394185066,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 3.9250073432922363,
+ "step": 1916
+ },
+ {
+ "epoch": 26.6282722513089,
+ "grad_norm": 0.051766350865364075,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 3.8915140628814697,
+ "step": 1917
+ },
+ {
+ "epoch": 26.642233856893544,
+ "grad_norm": 0.054485417902469635,
+ "learning_rate": 0.000285149463934261,
+ "loss": 3.8973851203918457,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656195462478184,
+ "grad_norm": 0.05437815189361572,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 3.9054274559020996,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670157068062828,
+ "grad_norm": 0.05402572825551033,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 3.9060683250427246,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68411867364747,
+ "grad_norm": 0.05332810431718826,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 3.9053144454956055,
+ "step": 1921
+ },
+ {
+ "epoch": 26.698080279232112,
+ "grad_norm": 0.0557575598359108,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 3.9026694297790527,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712041884816752,
+ "grad_norm": 0.05650033429265022,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 3.8795413970947266,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726003490401396,
+ "grad_norm": 0.054456304758787155,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 3.932950496673584,
+ "step": 1924
+ },
+ {
+ "epoch": 26.739965095986037,
+ "grad_norm": 0.05676668882369995,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 3.878704071044922,
+ "step": 1925
+ },
+ {
+ "epoch": 26.75392670157068,
+ "grad_norm": 0.05875959247350693,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 3.920717239379883,
+ "step": 1926
+ },
+ {
+ "epoch": 26.767888307155324,
+ "grad_norm": 0.054490379989147186,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 3.912592887878418,
+ "step": 1927
+ },
+ {
+ "epoch": 26.781849912739965,
+ "grad_norm": 0.055268481373786926,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 3.900418758392334,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79581151832461,
+ "grad_norm": 0.05617086961865425,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 3.899095296859741,
+ "step": 1929
+ },
+ {
+ "epoch": 26.80977312390925,
+ "grad_norm": 0.05310523882508278,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 3.917501449584961,
+ "step": 1930
+ },
+ {
+ "epoch": 26.823734729493893,
+ "grad_norm": 0.05476578325033188,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 3.867847442626953,
+ "step": 1931
+ },
+ {
+ "epoch": 26.837696335078533,
+ "grad_norm": 0.05456431582570076,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 3.8760366439819336,
+ "step": 1932
+ },
+ {
+ "epoch": 26.851657940663177,
+ "grad_norm": 0.055269334465265274,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 3.910797119140625,
+ "step": 1933
+ },
+ {
+ "epoch": 26.865619546247817,
+ "grad_norm": 0.054999858140945435,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 3.946951389312744,
+ "step": 1934
+ },
+ {
+ "epoch": 26.87958115183246,
+ "grad_norm": 0.05520344525575638,
+ "learning_rate": 0.000280627938758204,
+ "loss": 3.9161229133605957,
+ "step": 1935
+ },
+ {
+ "epoch": 26.8935427574171,
+ "grad_norm": 0.05451727285981178,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 3.908015727996826,
+ "step": 1936
+ },
+ {
+ "epoch": 26.907504363001745,
+ "grad_norm": 0.05561735853552818,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 3.944608449935913,
+ "step": 1937
+ },
+ {
+ "epoch": 26.921465968586386,
+ "grad_norm": 0.054743457585573196,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 3.906559467315674,
+ "step": 1938
+ },
+ {
+ "epoch": 26.93542757417103,
+ "grad_norm": 0.05497661978006363,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 3.900981903076172,
+ "step": 1939
+ },
+ {
+ "epoch": 26.949389179755673,
+ "grad_norm": 0.05452290549874306,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 3.910825252532959,
+ "step": 1940
+ },
+ {
+ "epoch": 26.963350785340314,
+ "grad_norm": 0.05409001186490059,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 3.921509265899658,
+ "step": 1941
+ },
+ {
+ "epoch": 26.977312390924958,
+ "grad_norm": 0.05467002093791962,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 3.9039864540100098,
+ "step": 1942
+ },
+ {
+ "epoch": 26.991273996509598,
+ "grad_norm": 0.05442725867033005,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 3.9153316020965576,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.039271868765354156,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 2.451029062271118,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 0.6021912097930908,
+ "eval_runtime": 61.3232,
+ "eval_samples_per_second": 39.822,
+ "eval_steps_per_second": 0.636,
+ "step": 1944
+ },
+ {
+ "epoch": 27.013961605584644,
+ "grad_norm": 0.05251244455575943,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 3.8143270015716553,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027923211169284,
+ "grad_norm": 0.056699808686971664,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 3.8331453800201416,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041884816753928,
+ "grad_norm": 0.055446673184633255,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 3.833841323852539,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05584642233857,
+ "grad_norm": 0.053186994045972824,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 3.82643985748291,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069808027923212,
+ "grad_norm": 0.05483868718147278,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 3.8209316730499268,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083769633507853,
+ "grad_norm": 0.05387664586305618,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 3.8402364253997803,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097731239092496,
+ "grad_norm": 0.05422234162688255,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 3.8125908374786377,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111692844677137,
+ "grad_norm": 0.05522442236542702,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 3.830599784851074,
+ "step": 1952
+ },
+ {
+ "epoch": 27.12565445026178,
+ "grad_norm": 0.05163128301501274,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 3.833247184753418,
+ "step": 1953
+ },
+ {
+ "epoch": 27.13961605584642,
+ "grad_norm": 0.053523529320955276,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 3.812887191772461,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153577661431065,
+ "grad_norm": 0.05362863466143608,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 3.8325610160827637,
+ "step": 1955
+ },
+ {
+ "epoch": 27.167539267015705,
+ "grad_norm": 0.054255858063697815,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 3.807152271270752,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18150087260035,
+ "grad_norm": 0.05259740352630615,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 3.843714475631714,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195462478184993,
+ "grad_norm": 0.05457160621881485,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 3.847249984741211,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209424083769633,
+ "grad_norm": 0.05517619848251343,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 3.864114284515381,
+ "step": 1959
+ },
+ {
+ "epoch": 27.223385689354277,
+ "grad_norm": 0.052946750074625015,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 3.8219656944274902,
+ "step": 1960
+ },
+ {
+ "epoch": 27.237347294938917,
+ "grad_norm": 0.053545452654361725,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 3.8502726554870605,
+ "step": 1961
+ },
+ {
+ "epoch": 27.25130890052356,
+ "grad_norm": 0.054657287895679474,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 3.837273120880127,
+ "step": 1962
+ },
+ {
+ "epoch": 27.2652705061082,
+ "grad_norm": 0.05385086312890053,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 3.837573528289795,
+ "step": 1963
+ },
+ {
+ "epoch": 27.279232111692846,
+ "grad_norm": 0.05512048304080963,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 3.842432737350464,
+ "step": 1964
+ },
+ {
+ "epoch": 27.293193717277486,
+ "grad_norm": 0.05286296457052231,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 3.845123767852783,
+ "step": 1965
+ },
+ {
+ "epoch": 27.30715532286213,
+ "grad_norm": 0.05396866053342819,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 3.836874485015869,
+ "step": 1966
+ },
+ {
+ "epoch": 27.32111692844677,
+ "grad_norm": 0.053341615945100784,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 3.8860960006713867,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335078534031414,
+ "grad_norm": 0.0531761609017849,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 3.828686237335205,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349040139616054,
+ "grad_norm": 0.054875023663043976,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 3.83782696723938,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363001745200698,
+ "grad_norm": 0.057710494846105576,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 3.8841819763183594,
+ "step": 1970
+ },
+ {
+ "epoch": 27.376963350785342,
+ "grad_norm": 0.05443769320845604,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 3.856564521789551,
+ "step": 1971
+ },
+ {
+ "epoch": 27.390924956369982,
+ "grad_norm": 0.05571551248431206,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 3.847541570663452,
+ "step": 1972
+ },
+ {
+ "epoch": 27.404886561954626,
+ "grad_norm": 0.05770037695765495,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 3.843390941619873,
+ "step": 1973
+ },
+ {
+ "epoch": 27.418848167539267,
+ "grad_norm": 0.05414504557847977,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 3.8786306381225586,
+ "step": 1974
+ },
+ {
+ "epoch": 27.43280977312391,
+ "grad_norm": 0.058058444410562515,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 3.8680999279022217,
+ "step": 1975
+ },
+ {
+ "epoch": 27.44677137870855,
+ "grad_norm": 0.056312717497348785,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 3.858879566192627,
+ "step": 1976
+ },
+ {
+ "epoch": 27.460732984293195,
+ "grad_norm": 0.055812690407037735,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 3.870741844177246,
+ "step": 1977
+ },
+ {
+ "epoch": 27.474694589877835,
+ "grad_norm": 0.057005349546670914,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 3.8850440979003906,
+ "step": 1978
+ },
+ {
+ "epoch": 27.48865619546248,
+ "grad_norm": 0.055214181542396545,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 3.8485918045043945,
+ "step": 1979
+ },
+ {
+ "epoch": 27.50261780104712,
+ "grad_norm": 0.054300058633089066,
+ "learning_rate": 0.00026868712586269,
+ "loss": 3.8759729862213135,
+ "step": 1980
+ },
+ {
+ "epoch": 27.516579406631763,
+ "grad_norm": 0.05370737239718437,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 3.8498623371124268,
+ "step": 1981
+ },
+ {
+ "epoch": 27.530541012216403,
+ "grad_norm": 0.05481775104999542,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 3.8408432006835938,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544502617801047,
+ "grad_norm": 0.05360998213291168,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 3.8800129890441895,
+ "step": 1983
+ },
+ {
+ "epoch": 27.55846422338569,
+ "grad_norm": 0.05510537698864937,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 3.8329358100891113,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57242582897033,
+ "grad_norm": 0.057950302958488464,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 3.870562791824341,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586387434554975,
+ "grad_norm": 0.0546487458050251,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 3.85231351852417,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600349040139616,
+ "grad_norm": 0.056359753012657166,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 3.8782615661621094,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61431064572426,
+ "grad_norm": 0.05409953370690346,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 3.8209736347198486,
+ "step": 1988
+ },
+ {
+ "epoch": 27.6282722513089,
+ "grad_norm": 0.05549091845750809,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 3.9041504859924316,
+ "step": 1989
+ },
+ {
+ "epoch": 27.642233856893544,
+ "grad_norm": 0.05437365546822548,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 3.8530654907226562,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656195462478184,
+ "grad_norm": 0.055456943809986115,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 3.862886428833008,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670157068062828,
+ "grad_norm": 0.05614335089921951,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 3.867595672607422,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68411867364747,
+ "grad_norm": 0.05374857038259506,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 3.8731391429901123,
+ "step": 1993
+ },
+ {
+ "epoch": 27.698080279232112,
+ "grad_norm": 0.05544714257121086,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 3.837608575820923,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712041884816752,
+ "grad_norm": 0.05495234578847885,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 3.8587915897369385,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726003490401396,
+ "grad_norm": 0.05354582890868187,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 3.8618052005767822,
+ "step": 1996
+ },
+ {
+ "epoch": 27.739965095986037,
+ "grad_norm": 0.05556262657046318,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 3.8630597591400146,
+ "step": 1997
+ },
+ {
+ "epoch": 27.75392670157068,
+ "grad_norm": 0.05576452240347862,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 3.8570914268493652,
+ "step": 1998
+ },
+ {
+ "epoch": 27.767888307155324,
+ "grad_norm": 0.053394027054309845,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 3.868783950805664,
+ "step": 1999
+ },
+ {
+ "epoch": 27.781849912739965,
+ "grad_norm": 0.05506644397974014,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 3.871555805206299,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79581151832461,
+ "grad_norm": 0.054367583245038986,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 3.8487513065338135,
+ "step": 2001
+ },
+ {
+ "epoch": 27.80977312390925,
+ "grad_norm": 0.05482787266373634,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 3.88199520111084,
+ "step": 2002
+ },
+ {
+ "epoch": 27.823734729493893,
+ "grad_norm": 0.05416601523756981,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 3.8876938819885254,
+ "step": 2003
+ },
+ {
+ "epoch": 27.837696335078533,
+ "grad_norm": 0.05612047761678696,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 3.8772940635681152,
+ "step": 2004
+ },
+ {
+ "epoch": 27.851657940663177,
+ "grad_norm": 0.05662519112229347,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 3.8429012298583984,
+ "step": 2005
+ },
+ {
+ "epoch": 27.865619546247817,
+ "grad_norm": 0.055557794868946075,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 3.8712010383605957,
+ "step": 2006
+ },
+ {
+ "epoch": 27.87958115183246,
+ "grad_norm": 0.054365914314985275,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 3.884854793548584,
+ "step": 2007
+ },
+ {
+ "epoch": 27.8935427574171,
+ "grad_norm": 0.05622334033250809,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 3.853909492492676,
+ "step": 2008
+ },
+ {
+ "epoch": 27.907504363001745,
+ "grad_norm": 0.05513857677578926,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 3.888460159301758,
+ "step": 2009
+ },
+ {
+ "epoch": 27.921465968586386,
+ "grad_norm": 0.05537945404648781,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 3.891343593597412,
+ "step": 2010
+ },
+ {
+ "epoch": 27.93542757417103,
+ "grad_norm": 0.05649508908390999,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 3.8807168006896973,
+ "step": 2011
+ },
+ {
+ "epoch": 27.949389179755673,
+ "grad_norm": 0.054261550307273865,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 3.8827662467956543,
+ "step": 2012
+ },
+ {
+ "epoch": 27.963350785340314,
+ "grad_norm": 0.05615224689245224,
+ "learning_rate": 0.00025996500713765,
+ "loss": 3.8720929622650146,
+ "step": 2013
+ },
+ {
+ "epoch": 27.977312390924958,
+ "grad_norm": 0.05571882426738739,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 3.8978285789489746,
+ "step": 2014
+ },
+ {
+ "epoch": 27.991273996509598,
+ "grad_norm": 0.05477411672472954,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 3.875284194946289,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.03992476686835289,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 2.404466152191162,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 0.6039577126502991,
+ "eval_runtime": 59.8223,
+ "eval_samples_per_second": 40.821,
+ "eval_steps_per_second": 0.652,
+ "step": 2016
+ },
+ {
+ "epoch": 28.013961605584644,
+ "grad_norm": 0.05395985022187233,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 3.78012752532959,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027923211169284,
+ "grad_norm": 0.05445432290434837,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 3.794297218322754,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041884816753928,
+ "grad_norm": 0.05048619955778122,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 3.8059635162353516,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05584642233857,
+ "grad_norm": 0.054649483412504196,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 3.8080105781555176,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069808027923212,
+ "grad_norm": 0.054571982473134995,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 3.8171639442443848,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083769633507853,
+ "grad_norm": 0.05476690083742142,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 3.7937488555908203,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097731239092496,
+ "grad_norm": 0.05586248263716698,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 3.804347515106201,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111692844677137,
+ "grad_norm": 0.05466577783226967,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 3.8114442825317383,
+ "step": 2024
+ },
+ {
+ "epoch": 28.12565445026178,
+ "grad_norm": 0.05342315137386322,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 3.8107991218566895,
+ "step": 2025
+ },
+ {
+ "epoch": 28.13961605584642,
+ "grad_norm": 0.05318617820739746,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 3.8390064239501953,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153577661431065,
+ "grad_norm": 0.05371803045272827,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 3.8021721839904785,
+ "step": 2027
+ },
+ {
+ "epoch": 28.167539267015705,
+ "grad_norm": 0.05319817736744881,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 3.844275712966919,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18150087260035,
+ "grad_norm": 0.05202893167734146,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 3.8008670806884766,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195462478184993,
+ "grad_norm": 0.054026514291763306,
+ "learning_rate": 0.000255486047794226,
+ "loss": 3.81990909576416,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209424083769633,
+ "grad_norm": 0.05195610597729683,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 3.8142457008361816,
+ "step": 2031
+ },
+ {
+ "epoch": 28.223385689354277,
+ "grad_norm": 0.05212428793311119,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 3.8139138221740723,
+ "step": 2032
+ },
+ {
+ "epoch": 28.237347294938917,
+ "grad_norm": 0.05503816530108452,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 3.8072667121887207,
+ "step": 2033
+ },
+ {
+ "epoch": 28.25130890052356,
+ "grad_norm": 0.05305452272295952,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 3.8146510124206543,
+ "step": 2034
+ },
+ {
+ "epoch": 28.2652705061082,
+ "grad_norm": 0.05301898717880249,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 3.7936806678771973,
+ "step": 2035
+ },
+ {
+ "epoch": 28.279232111692846,
+ "grad_norm": 0.0542520172894001,
+ "learning_rate": 0.000253907826333243,
+ "loss": 3.7920427322387695,
+ "step": 2036
+ },
+ {
+ "epoch": 28.293193717277486,
+ "grad_norm": 0.05349896103143692,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 3.819732189178467,
+ "step": 2037
+ },
+ {
+ "epoch": 28.30715532286213,
+ "grad_norm": 0.05463109537959099,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 3.804321765899658,
+ "step": 2038
+ },
+ {
+ "epoch": 28.32111692844677,
+ "grad_norm": 0.05445708706974983,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 3.8128511905670166,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335078534031414,
+ "grad_norm": 0.05655837804079056,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 3.798435688018799,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349040139616054,
+ "grad_norm": 0.05543001368641853,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 3.779137134552002,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363001745200698,
+ "grad_norm": 0.05458657816052437,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 3.833167552947998,
+ "step": 2042
+ },
+ {
+ "epoch": 28.376963350785342,
+ "grad_norm": 0.056820500642061234,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 3.8207151889801025,
+ "step": 2043
+ },
+ {
+ "epoch": 28.390924956369982,
+ "grad_norm": 0.05464114993810654,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 3.788215160369873,
+ "step": 2044
+ },
+ {
+ "epoch": 28.404886561954626,
+ "grad_norm": 0.053662411868572235,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 3.8059287071228027,
+ "step": 2045
+ },
+ {
+ "epoch": 28.418848167539267,
+ "grad_norm": 0.055799756199121475,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 3.8050360679626465,
+ "step": 2046
+ },
+ {
+ "epoch": 28.43280977312391,
+ "grad_norm": 0.05412062630057335,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 3.8110790252685547,
+ "step": 2047
+ },
+ {
+ "epoch": 28.44677137870855,
+ "grad_norm": 0.05672532692551613,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 3.841888904571533,
+ "step": 2048
+ },
+ {
+ "epoch": 28.460732984293195,
+ "grad_norm": 0.05622367933392525,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 3.8200669288635254,
+ "step": 2049
+ },
+ {
+ "epoch": 28.474694589877835,
+ "grad_norm": 0.0544997975230217,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 3.8031818866729736,
+ "step": 2050
+ },
+ {
+ "epoch": 28.48865619546248,
+ "grad_norm": 0.058867089450359344,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 3.856843948364258,
+ "step": 2051
+ },
+ {
+ "epoch": 28.50261780104712,
+ "grad_norm": 0.055324845016002655,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 3.807135820388794,
+ "step": 2052
+ },
+ {
+ "epoch": 28.516579406631763,
+ "grad_norm": 0.055414266884326935,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 3.8380215167999268,
+ "step": 2053
+ },
+ {
+ "epoch": 28.530541012216403,
+ "grad_norm": 0.05787660554051399,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 3.831721782684326,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544502617801047,
+ "grad_norm": 0.05551844462752342,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 3.8572306632995605,
+ "step": 2055
+ },
+ {
+ "epoch": 28.55846422338569,
+ "grad_norm": 0.05452537164092064,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 3.840963840484619,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57242582897033,
+ "grad_norm": 0.05530945956707001,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 3.840442419052124,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586387434554975,
+ "grad_norm": 0.05631886050105095,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 3.7851438522338867,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600349040139616,
+ "grad_norm": 0.05489915981888771,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 3.7870893478393555,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61431064572426,
+ "grad_norm": 0.055331867188215256,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 3.8441572189331055,
+ "step": 2060
+ },
+ {
+ "epoch": 28.6282722513089,
+ "grad_norm": 0.05580870062112808,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 3.834148645401001,
+ "step": 2061
+ },
+ {
+ "epoch": 28.642233856893544,
+ "grad_norm": 0.05681769549846649,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 3.8349759578704834,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656195462478184,
+ "grad_norm": 0.05484442040324211,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 3.798792600631714,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670157068062828,
+ "grad_norm": 0.05657925084233284,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 3.853915214538574,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68411867364747,
+ "grad_norm": 0.05419609695672989,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 3.8423995971679688,
+ "step": 2065
+ },
+ {
+ "epoch": 28.698080279232112,
+ "grad_norm": 0.05700303241610527,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 3.822547674179077,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712041884816752,
+ "grad_norm": 0.055846985429525375,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 3.830044984817505,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726003490401396,
+ "grad_norm": 0.05754954740405083,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 3.868316173553467,
+ "step": 2068
+ },
+ {
+ "epoch": 28.739965095986037,
+ "grad_norm": 0.056265417486429214,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 3.8297243118286133,
+ "step": 2069
+ },
+ {
+ "epoch": 28.75392670157068,
+ "grad_norm": 0.05429232493042946,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 3.872836112976074,
+ "step": 2070
+ },
+ {
+ "epoch": 28.767888307155324,
+ "grad_norm": 0.05589764565229416,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 3.838311195373535,
+ "step": 2071
+ },
+ {
+ "epoch": 28.781849912739965,
+ "grad_norm": 0.05460003390908241,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 3.8316597938537598,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79581151832461,
+ "grad_norm": 0.054542504251003265,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 3.816542625427246,
+ "step": 2073
+ },
+ {
+ "epoch": 28.80977312390925,
+ "grad_norm": 0.0549996979534626,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 3.796245574951172,
+ "step": 2074
+ },
+ {
+ "epoch": 28.823734729493893,
+ "grad_norm": 0.05521940439939499,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 3.8112473487854004,
+ "step": 2075
+ },
+ {
+ "epoch": 28.837696335078533,
+ "grad_norm": 0.05690402165055275,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 3.8242125511169434,
+ "step": 2076
+ },
+ {
+ "epoch": 28.851657940663177,
+ "grad_norm": 0.055383820086717606,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 3.84938383102417,
+ "step": 2077
+ },
+ {
+ "epoch": 28.865619546247817,
+ "grad_norm": 0.05612736567854881,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 3.8383970260620117,
+ "step": 2078
+ },
+ {
+ "epoch": 28.87958115183246,
+ "grad_norm": 0.05723752826452255,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 3.828171730041504,
+ "step": 2079
+ },
+ {
+ "epoch": 28.8935427574171,
+ "grad_norm": 0.0576927475631237,
+ "learning_rate": 0.000242380656502223,
+ "loss": 3.8756160736083984,
+ "step": 2080
+ },
+ {
+ "epoch": 28.907504363001745,
+ "grad_norm": 0.05609026178717613,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 3.8703267574310303,
+ "step": 2081
+ },
+ {
+ "epoch": 28.921465968586386,
+ "grad_norm": 0.057038433849811554,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 3.8566956520080566,
+ "step": 2082
+ },
+ {
+ "epoch": 28.93542757417103,
+ "grad_norm": 0.05683749541640282,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 3.8409035205841064,
+ "step": 2083
+ },
+ {
+ "epoch": 28.949389179755673,
+ "grad_norm": 0.05682830885052681,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 3.8435707092285156,
+ "step": 2084
+ },
+ {
+ "epoch": 28.963350785340314,
+ "grad_norm": 0.05645477771759033,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 3.8142364025115967,
+ "step": 2085
+ },
+ {
+ "epoch": 28.977312390924958,
+ "grad_norm": 0.05519694834947586,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 3.830415725708008,
+ "step": 2086
+ },
+ {
+ "epoch": 28.991273996509598,
+ "grad_norm": 0.05785970389842987,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 3.8413705825805664,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.04018386825919151,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 2.3790979385375977,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 0.6047304272651672,
+ "eval_runtime": 61.1049,
+ "eval_samples_per_second": 39.964,
+ "eval_steps_per_second": 0.638,
+ "step": 2088
+ },
+ {
+ "epoch": 29.013961605584644,
+ "grad_norm": 0.058019861578941345,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 3.793783664703369,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027923211169284,
+ "grad_norm": 0.058003105223178864,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 3.783708095550537,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041884816753928,
+ "grad_norm": 0.05485067516565323,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 3.7829298973083496,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05584642233857,
+ "grad_norm": 0.056009069085121155,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 3.7747392654418945,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069808027923212,
+ "grad_norm": 0.05480542406439781,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 3.774707317352295,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083769633507853,
+ "grad_norm": 0.05384726822376251,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 3.75443959236145,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097731239092496,
+ "grad_norm": 0.053818102926015854,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 3.7790050506591797,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111692844677137,
+ "grad_norm": 0.05595344305038452,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 3.74898624420166,
+ "step": 2096
+ },
+ {
+ "epoch": 29.12565445026178,
+ "grad_norm": 0.05394545570015907,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 3.7751331329345703,
+ "step": 2097
+ },
+ {
+ "epoch": 29.13961605584642,
+ "grad_norm": 0.05346408113837242,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 3.7671167850494385,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153577661431065,
+ "grad_norm": 0.0541236512362957,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 3.749098062515259,
+ "step": 2099
+ },
+ {
+ "epoch": 29.167539267015705,
+ "grad_norm": 0.05369208753108978,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 3.7778515815734863,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18150087260035,
+ "grad_norm": 0.055362965911626816,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 3.747697591781616,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195462478184993,
+ "grad_norm": 0.05389911308884621,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 3.7759058475494385,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209424083769633,
+ "grad_norm": 0.05315527692437172,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 3.7832705974578857,
+ "step": 2103
+ },
+ {
+ "epoch": 29.223385689354277,
+ "grad_norm": 0.05377589911222458,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 3.7743210792541504,
+ "step": 2104
+ },
+ {
+ "epoch": 29.237347294938917,
+ "grad_norm": 0.05307800695300102,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 3.7654497623443604,
+ "step": 2105
+ },
+ {
+ "epoch": 29.25130890052356,
+ "grad_norm": 0.056719157844781876,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 3.794179916381836,
+ "step": 2106
+ },
+ {
+ "epoch": 29.2652705061082,
+ "grad_norm": 0.053281862288713455,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 3.7620739936828613,
+ "step": 2107
+ },
+ {
+ "epoch": 29.279232111692846,
+ "grad_norm": 0.05470574274659157,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 3.788421869277954,
+ "step": 2108
+ },
+ {
+ "epoch": 29.293193717277486,
+ "grad_norm": 0.05293853580951691,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 3.7655067443847656,
+ "step": 2109
+ },
+ {
+ "epoch": 29.30715532286213,
+ "grad_norm": 0.05524440482258797,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 3.7701616287231445,
+ "step": 2110
+ },
+ {
+ "epoch": 29.32111692844677,
+ "grad_norm": 0.05508396402001381,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 3.7993319034576416,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335078534031414,
+ "grad_norm": 0.05559603124856949,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 3.7799668312072754,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349040139616054,
+ "grad_norm": 0.05530122295022011,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 3.751032829284668,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363001745200698,
+ "grad_norm": 0.05408896878361702,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 3.7761611938476562,
+ "step": 2114
+ },
+ {
+ "epoch": 29.376963350785342,
+ "grad_norm": 0.05530737340450287,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 3.745713233947754,
+ "step": 2115
+ },
+ {
+ "epoch": 29.390924956369982,
+ "grad_norm": 0.05424542352557182,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 3.8041834831237793,
+ "step": 2116
+ },
+ {
+ "epoch": 29.404886561954626,
+ "grad_norm": 0.054800380021333694,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 3.7995352745056152,
+ "step": 2117
+ },
+ {
+ "epoch": 29.418848167539267,
+ "grad_norm": 0.05685900151729584,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 3.7970776557922363,
+ "step": 2118
+ },
+ {
+ "epoch": 29.43280977312391,
+ "grad_norm": 0.05555475875735283,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 3.7950801849365234,
+ "step": 2119
+ },
+ {
+ "epoch": 29.44677137870855,
+ "grad_norm": 0.05670571327209473,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 3.7967746257781982,
+ "step": 2120
+ },
+ {
+ "epoch": 29.460732984293195,
+ "grad_norm": 0.05598503723740578,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 3.809349298477173,
+ "step": 2121
+ },
+ {
+ "epoch": 29.474694589877835,
+ "grad_norm": 0.05501085892319679,
+ "learning_rate": 0.000231465389734324,
+ "loss": 3.803968906402588,
+ "step": 2122
+ },
+ {
+ "epoch": 29.48865619546248,
+ "grad_norm": 0.05519728362560272,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 3.7615163326263428,
+ "step": 2123
+ },
+ {
+ "epoch": 29.50261780104712,
+ "grad_norm": 0.055866893380880356,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 3.7939207553863525,
+ "step": 2124
+ },
+ {
+ "epoch": 29.516579406631763,
+ "grad_norm": 0.05460986867547035,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 3.7740321159362793,
+ "step": 2125
+ },
+ {
+ "epoch": 29.530541012216403,
+ "grad_norm": 0.055752526968717575,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 3.82424259185791,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544502617801047,
+ "grad_norm": 0.056792002171278,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 3.8006062507629395,
+ "step": 2127
+ },
+ {
+ "epoch": 29.55846422338569,
+ "grad_norm": 0.054986681789159775,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 3.7713897228240967,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57242582897033,
+ "grad_norm": 0.054215673357248306,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 3.8005690574645996,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586387434554975,
+ "grad_norm": 0.05700439214706421,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 3.825161933898926,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600349040139616,
+ "grad_norm": 0.055078551173210144,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 3.786716938018799,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61431064572426,
+ "grad_norm": 0.055569086223840714,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 3.8020172119140625,
+ "step": 2132
+ },
+ {
+ "epoch": 29.6282722513089,
+ "grad_norm": 0.0548536442220211,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 3.8038570880889893,
+ "step": 2133
+ },
+ {
+ "epoch": 29.642233856893544,
+ "grad_norm": 0.05604325234889984,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 3.796006441116333,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656195462478184,
+ "grad_norm": 0.05614005774259567,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 3.7938876152038574,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670157068062828,
+ "grad_norm": 0.05761241167783737,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 3.813568592071533,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68411867364747,
+ "grad_norm": 0.05579913780093193,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 3.826862335205078,
+ "step": 2137
+ },
+ {
+ "epoch": 29.698080279232112,
+ "grad_norm": 0.05784283205866814,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 3.808281421661377,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712041884816752,
+ "grad_norm": 0.05596453696489334,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 3.8119921684265137,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726003490401396,
+ "grad_norm": 0.05699361115694046,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 3.822702407836914,
+ "step": 2140
+ },
+ {
+ "epoch": 29.739965095986037,
+ "grad_norm": 0.05375700071454048,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 3.78023099899292,
+ "step": 2141
+ },
+ {
+ "epoch": 29.75392670157068,
+ "grad_norm": 0.05588206648826599,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 3.791456699371338,
+ "step": 2142
+ },
+ {
+ "epoch": 29.767888307155324,
+ "grad_norm": 0.05579312890768051,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 3.799713134765625,
+ "step": 2143
+ },
+ {
+ "epoch": 29.781849912739965,
+ "grad_norm": 0.05713854730129242,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 3.8211703300476074,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79581151832461,
+ "grad_norm": 0.05645165592432022,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 3.7947046756744385,
+ "step": 2145
+ },
+ {
+ "epoch": 29.80977312390925,
+ "grad_norm": 0.056201402097940445,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 3.811009407043457,
+ "step": 2146
+ },
+ {
+ "epoch": 29.823734729493893,
+ "grad_norm": 0.05659479275345802,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 3.8009982109069824,
+ "step": 2147
+ },
+ {
+ "epoch": 29.837696335078533,
+ "grad_norm": 0.05797034874558449,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 3.7994565963745117,
+ "step": 2148
+ },
+ {
+ "epoch": 29.851657940663177,
+ "grad_norm": 0.05734949931502342,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 3.7696728706359863,
+ "step": 2149
+ },
+ {
+ "epoch": 29.865619546247817,
+ "grad_norm": 0.060396358370780945,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 3.8269689083099365,
+ "step": 2150
+ },
+ {
+ "epoch": 29.87958115183246,
+ "grad_norm": 0.057348065078258514,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 3.7925050258636475,
+ "step": 2151
+ },
+ {
+ "epoch": 29.8935427574171,
+ "grad_norm": 0.058563247323036194,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 3.82887864112854,
+ "step": 2152
+ },
+ {
+ "epoch": 29.907504363001745,
+ "grad_norm": 0.06011585891246796,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 3.8061296939849854,
+ "step": 2153
+ },
+ {
+ "epoch": 29.921465968586386,
+ "grad_norm": 0.058440592139959335,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 3.792586326599121,
+ "step": 2154
+ },
+ {
+ "epoch": 29.93542757417103,
+ "grad_norm": 0.057050157338380814,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 3.7885212898254395,
+ "step": 2155
+ },
+ {
+ "epoch": 29.949389179755673,
+ "grad_norm": 0.05920996889472008,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 3.813566207885742,
+ "step": 2156
+ },
+ {
+ "epoch": 29.963350785340314,
+ "grad_norm": 0.057791322469711304,
+ "learning_rate": 0.00022244633283095,
+ "loss": 3.8399810791015625,
+ "step": 2157
+ },
+ {
+ "epoch": 29.977312390924958,
+ "grad_norm": 0.05644001066684723,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 3.79464054107666,
+ "step": 2158
+ },
+ {
+ "epoch": 29.991273996509598,
+ "grad_norm": 0.06021320819854736,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 3.804086208343506,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.040621865540742874,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 2.386565923690796,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 0.6063579320907593,
+ "eval_runtime": 59.5345,
+ "eval_samples_per_second": 41.018,
+ "eval_steps_per_second": 0.655,
+ "step": 2160
+ },
+ {
+ "epoch": 30.013961605584644,
+ "grad_norm": 0.05961514264345169,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 3.756401777267456,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027923211169284,
+ "grad_norm": 0.057169776409864426,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 3.7327475547790527,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041884816753928,
+ "grad_norm": 0.056084226816892624,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 3.7365498542785645,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05584642233857,
+ "grad_norm": 0.059099674224853516,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 3.7383875846862793,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069808027923212,
+ "grad_norm": 0.05632825568318367,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 3.7642762660980225,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083769633507853,
+ "grad_norm": 0.056937988847494125,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 3.7614428997039795,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097731239092496,
+ "grad_norm": 0.05722609534859657,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 3.7394845485687256,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111692844677137,
+ "grad_norm": 0.056204989552497864,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 3.7473866939544678,
+ "step": 2168
+ },
+ {
+ "epoch": 30.12565445026178,
+ "grad_norm": 0.05640130862593651,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 3.7504067420959473,
+ "step": 2169
+ },
+ {
+ "epoch": 30.13961605584642,
+ "grad_norm": 0.05625637248158455,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 3.754812717437744,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153577661431065,
+ "grad_norm": 0.05596087500452995,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 3.741854429244995,
+ "step": 2171
+ },
+ {
+ "epoch": 30.167539267015705,
+ "grad_norm": 0.054819896817207336,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 3.724705934524536,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18150087260035,
+ "grad_norm": 0.05644344910979271,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 3.7609610557556152,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195462478184993,
+ "grad_norm": 0.05366845801472664,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 3.7557997703552246,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209424083769633,
+ "grad_norm": 0.05517663434147835,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 3.7494120597839355,
+ "step": 2175
+ },
+ {
+ "epoch": 30.223385689354277,
+ "grad_norm": 0.05463553965091705,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 3.7434799671173096,
+ "step": 2176
+ },
+ {
+ "epoch": 30.237347294938917,
+ "grad_norm": 0.05352703109383583,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 3.7435178756713867,
+ "step": 2177
+ },
+ {
+ "epoch": 30.25130890052356,
+ "grad_norm": 0.053156204521656036,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 3.7648849487304688,
+ "step": 2178
+ },
+ {
+ "epoch": 30.2652705061082,
+ "grad_norm": 0.05557673051953316,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 3.746366024017334,
+ "step": 2179
+ },
+ {
+ "epoch": 30.279232111692846,
+ "grad_norm": 0.052989061921834946,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 3.7453460693359375,
+ "step": 2180
+ },
+ {
+ "epoch": 30.293193717277486,
+ "grad_norm": 0.055707018822431564,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 3.800962209701538,
+ "step": 2181
+ },
+ {
+ "epoch": 30.30715532286213,
+ "grad_norm": 0.05560754984617233,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 3.7416906356811523,
+ "step": 2182
+ },
+ {
+ "epoch": 30.32111692844677,
+ "grad_norm": 0.055344413965940475,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 3.7672080993652344,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335078534031414,
+ "grad_norm": 0.05490704998373985,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 3.7664341926574707,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349040139616054,
+ "grad_norm": 0.0578596331179142,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 3.7632641792297363,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363001745200698,
+ "grad_norm": 0.055299609899520874,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 3.7502024173736572,
+ "step": 2186
+ },
+ {
+ "epoch": 30.376963350785342,
+ "grad_norm": 0.055016499012708664,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 3.732128143310547,
+ "step": 2187
+ },
+ {
+ "epoch": 30.390924956369982,
+ "grad_norm": 0.057352419942617416,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 3.723008871078491,
+ "step": 2188
+ },
+ {
+ "epoch": 30.404886561954626,
+ "grad_norm": 0.05596115067601204,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 3.786574363708496,
+ "step": 2189
+ },
+ {
+ "epoch": 30.418848167539267,
+ "grad_norm": 0.05596112087368965,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 3.770382881164551,
+ "step": 2190
+ },
+ {
+ "epoch": 30.43280977312391,
+ "grad_norm": 0.0562058687210083,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 3.7605714797973633,
+ "step": 2191
+ },
+ {
+ "epoch": 30.44677137870855,
+ "grad_norm": 0.056407757103443146,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 3.7485427856445312,
+ "step": 2192
+ },
+ {
+ "epoch": 30.460732984293195,
+ "grad_norm": 0.05596652254462242,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 3.7630670070648193,
+ "step": 2193
+ },
+ {
+ "epoch": 30.474694589877835,
+ "grad_norm": 0.056551866233348846,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 3.7981796264648438,
+ "step": 2194
+ },
+ {
+ "epoch": 30.48865619546248,
+ "grad_norm": 0.057354725897312164,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 3.7544898986816406,
+ "step": 2195
+ },
+ {
+ "epoch": 30.50261780104712,
+ "grad_norm": 0.05791718140244484,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 3.7241907119750977,
+ "step": 2196
+ },
+ {
+ "epoch": 30.516579406631763,
+ "grad_norm": 0.05675847455859184,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 3.794506549835205,
+ "step": 2197
+ },
+ {
+ "epoch": 30.530541012216403,
+ "grad_norm": 0.060349952429533005,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 3.75946044921875,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544502617801047,
+ "grad_norm": 0.05547260120511055,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 3.7278571128845215,
+ "step": 2199
+ },
+ {
+ "epoch": 30.55846422338569,
+ "grad_norm": 0.058400996029376984,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 3.756891965866089,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57242582897033,
+ "grad_norm": 0.057871121913194656,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 3.783863067626953,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586387434554975,
+ "grad_norm": 0.05899912118911743,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 3.7787115573883057,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600349040139616,
+ "grad_norm": 0.056164853274822235,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 3.7599334716796875,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61431064572426,
+ "grad_norm": 0.05775163322687149,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 3.7457070350646973,
+ "step": 2204
+ },
+ {
+ "epoch": 30.6282722513089,
+ "grad_norm": 0.057221125811338425,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 3.7663235664367676,
+ "step": 2205
+ },
+ {
+ "epoch": 30.642233856893544,
+ "grad_norm": 0.05788435414433479,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 3.7676830291748047,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656195462478184,
+ "grad_norm": 0.057100843638181686,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 3.7586870193481445,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670157068062828,
+ "grad_norm": 0.05703200399875641,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 3.7561769485473633,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68411867364747,
+ "grad_norm": 0.05664575845003128,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 3.761331081390381,
+ "step": 2209
+ },
+ {
+ "epoch": 30.698080279232112,
+ "grad_norm": 0.056622982025146484,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 3.737915515899658,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712041884816752,
+ "grad_norm": 0.056481845676898956,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 3.7413077354431152,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726003490401396,
+ "grad_norm": 0.057718146592378616,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 3.751267433166504,
+ "step": 2212
+ },
+ {
+ "epoch": 30.739965095986037,
+ "grad_norm": 0.05772862955927849,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 3.753678321838379,
+ "step": 2213
+ },
+ {
+ "epoch": 30.75392670157068,
+ "grad_norm": 0.056571926921606064,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 3.751033306121826,
+ "step": 2214
+ },
+ {
+ "epoch": 30.767888307155324,
+ "grad_norm": 0.05682666599750519,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 3.763153553009033,
+ "step": 2215
+ },
+ {
+ "epoch": 30.781849912739965,
+ "grad_norm": 0.05993897467851639,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 3.787890911102295,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79581151832461,
+ "grad_norm": 0.05659458413720131,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 3.7655086517333984,
+ "step": 2217
+ },
+ {
+ "epoch": 30.80977312390925,
+ "grad_norm": 0.057996224611997604,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 3.774881601333618,
+ "step": 2218
+ },
+ {
+ "epoch": 30.823734729493893,
+ "grad_norm": 0.05677317455410957,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 3.7600479125976562,
+ "step": 2219
+ },
+ {
+ "epoch": 30.837696335078533,
+ "grad_norm": 0.05664737895131111,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 3.787637233734131,
+ "step": 2220
+ },
+ {
+ "epoch": 30.851657940663177,
+ "grad_norm": 0.05808616429567337,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 3.7529220581054688,
+ "step": 2221
+ },
+ {
+ "epoch": 30.865619546247817,
+ "grad_norm": 0.05666997656226158,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 3.7790794372558594,
+ "step": 2222
+ },
+ {
+ "epoch": 30.87958115183246,
+ "grad_norm": 0.05963682755827904,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 3.762615203857422,
+ "step": 2223
+ },
+ {
+ "epoch": 30.8935427574171,
+ "grad_norm": 0.05558215081691742,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 3.7871851921081543,
+ "step": 2224
+ },
+ {
+ "epoch": 30.907504363001745,
+ "grad_norm": 0.05762261897325516,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 3.7605397701263428,
+ "step": 2225
+ },
+ {
+ "epoch": 30.921465968586386,
+ "grad_norm": 0.0565858818590641,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 3.7693653106689453,
+ "step": 2226
+ },
+ {
+ "epoch": 30.93542757417103,
+ "grad_norm": 0.05821581557393074,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 3.766758918762207,
+ "step": 2227
+ },
+ {
+ "epoch": 30.949389179755673,
+ "grad_norm": 0.055186714977025986,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 3.79498553276062,
+ "step": 2228
+ },
+ {
+ "epoch": 30.963350785340314,
+ "grad_norm": 0.05850847810506821,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 3.799769163131714,
+ "step": 2229
+ },
+ {
+ "epoch": 30.977312390924958,
+ "grad_norm": 0.05700397863984108,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 3.7555625438690186,
+ "step": 2230
+ },
+ {
+ "epoch": 30.991273996509598,
+ "grad_norm": 0.057813212275505066,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 3.777778148651123,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.041730619966983795,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 2.3564491271972656,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 0.6080329418182373,
+ "eval_runtime": 60.4607,
+ "eval_samples_per_second": 40.39,
+ "eval_steps_per_second": 0.645,
+ "step": 2232
+ },
+ {
+ "epoch": 31.013961605584644,
+ "grad_norm": 0.0563117116689682,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 3.719982624053955,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027923211169284,
+ "grad_norm": 0.05564197897911072,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 3.719735622406006,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041884816753928,
+ "grad_norm": 0.05562874674797058,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 3.739713430404663,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05584642233857,
+ "grad_norm": 0.057146716862916946,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 3.736401081085205,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069808027923212,
+ "grad_norm": 0.053163520991802216,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 3.7049455642700195,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083769633507853,
+ "grad_norm": 0.057067256420850754,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 3.723662853240967,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097731239092496,
+ "grad_norm": 0.05276816710829735,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 3.710062026977539,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111692844677137,
+ "grad_norm": 0.055126190185546875,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 3.696173667907715,
+ "step": 2240
+ },
+ {
+ "epoch": 31.12565445026178,
+ "grad_norm": 0.05533793941140175,
+ "learning_rate": 0.000201141724176723,
+ "loss": 3.710416793823242,
+ "step": 2241
+ },
+ {
+ "epoch": 31.13961605584642,
+ "grad_norm": 0.055686868727207184,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 3.7188220024108887,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153577661431065,
+ "grad_norm": 0.055188920348882675,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 3.7026562690734863,
+ "step": 2243
+ },
+ {
+ "epoch": 31.167539267015705,
+ "grad_norm": 0.053627416491508484,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 3.7224509716033936,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18150087260035,
+ "grad_norm": 0.05535290017724037,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 3.710698366165161,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195462478184993,
+ "grad_norm": 0.0528891459107399,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 3.6979012489318848,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209424083769633,
+ "grad_norm": 0.055201176553964615,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 3.7324817180633545,
+ "step": 2247
+ },
+ {
+ "epoch": 31.223385689354277,
+ "grad_norm": 0.05338778346776962,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 3.722564697265625,
+ "step": 2248
+ },
+ {
+ "epoch": 31.237347294938917,
+ "grad_norm": 0.05368843302130699,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 3.7118425369262695,
+ "step": 2249
+ },
+ {
+ "epoch": 31.25130890052356,
+ "grad_norm": 0.052788376808166504,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 3.6875901222229004,
+ "step": 2250
+ },
+ {
+ "epoch": 31.2652705061082,
+ "grad_norm": 0.05427199602127075,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 3.7042768001556396,
+ "step": 2251
+ },
+ {
+ "epoch": 31.279232111692846,
+ "grad_norm": 0.05434224382042885,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 3.7138795852661133,
+ "step": 2252
+ },
+ {
+ "epoch": 31.293193717277486,
+ "grad_norm": 0.054364923387765884,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 3.735142946243286,
+ "step": 2253
+ },
+ {
+ "epoch": 31.30715532286213,
+ "grad_norm": 0.05535430088639259,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 3.7198615074157715,
+ "step": 2254
+ },
+ {
+ "epoch": 31.32111692844677,
+ "grad_norm": 0.05463665351271629,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 3.733370542526245,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335078534031414,
+ "grad_norm": 0.0553361251950264,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 3.702805519104004,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349040139616054,
+ "grad_norm": 0.05423365905880928,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 3.7315120697021484,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363001745200698,
+ "grad_norm": 0.05499394237995148,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 3.731891632080078,
+ "step": 2258
+ },
+ {
+ "epoch": 31.376963350785342,
+ "grad_norm": 0.054968126118183136,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 3.7355899810791016,
+ "step": 2259
+ },
+ {
+ "epoch": 31.390924956369982,
+ "grad_norm": 0.05478367581963539,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 3.7143867015838623,
+ "step": 2260
+ },
+ {
+ "epoch": 31.404886561954626,
+ "grad_norm": 0.05554179847240448,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 3.7227137088775635,
+ "step": 2261
+ },
+ {
+ "epoch": 31.418848167539267,
+ "grad_norm": 0.05399509146809578,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 3.750650405883789,
+ "step": 2262
+ },
+ {
+ "epoch": 31.43280977312391,
+ "grad_norm": 0.056201111525297165,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 3.7478344440460205,
+ "step": 2263
+ },
+ {
+ "epoch": 31.44677137870855,
+ "grad_norm": 0.05437091737985611,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 3.7283718585968018,
+ "step": 2264
+ },
+ {
+ "epoch": 31.460732984293195,
+ "grad_norm": 0.057144906371831894,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 3.7482407093048096,
+ "step": 2265
+ },
+ {
+ "epoch": 31.474694589877835,
+ "grad_norm": 0.05441201478242874,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 3.726378917694092,
+ "step": 2266
+ },
+ {
+ "epoch": 31.48865619546248,
+ "grad_norm": 0.056569576263427734,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 3.7211437225341797,
+ "step": 2267
+ },
+ {
+ "epoch": 31.50261780104712,
+ "grad_norm": 0.05561831593513489,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 3.7278850078582764,
+ "step": 2268
+ },
+ {
+ "epoch": 31.516579406631763,
+ "grad_norm": 0.0569809265434742,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 3.713228702545166,
+ "step": 2269
+ },
+ {
+ "epoch": 31.530541012216403,
+ "grad_norm": 0.05654345825314522,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 3.7354423999786377,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544502617801047,
+ "grad_norm": 0.05482077598571777,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 3.7043380737304688,
+ "step": 2271
+ },
+ {
+ "epoch": 31.55846422338569,
+ "grad_norm": 0.055417660623788834,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 3.7199621200561523,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57242582897033,
+ "grad_norm": 0.054656337946653366,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 3.734508514404297,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586387434554975,
+ "grad_norm": 0.05609512701630592,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 3.7341227531433105,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600349040139616,
+ "grad_norm": 0.05602060630917549,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 3.719949960708618,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61431064572426,
+ "grad_norm": 0.05535270646214485,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 3.7124414443969727,
+ "step": 2276
+ },
+ {
+ "epoch": 31.6282722513089,
+ "grad_norm": 0.05577492713928223,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 3.706301212310791,
+ "step": 2277
+ },
+ {
+ "epoch": 31.642233856893544,
+ "grad_norm": 0.057382550090551376,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 3.7411818504333496,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656195462478184,
+ "grad_norm": 0.056383028626441956,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 3.751235008239746,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670157068062828,
+ "grad_norm": 0.05667940899729729,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 3.7379674911499023,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68411867364747,
+ "grad_norm": 0.05635855346918106,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 3.728001594543457,
+ "step": 2281
+ },
+ {
+ "epoch": 31.698080279232112,
+ "grad_norm": 0.05649873986840248,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 3.740757465362549,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712041884816752,
+ "grad_norm": 0.055871494114398956,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 3.7373616695404053,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726003490401396,
+ "grad_norm": 0.057106468826532364,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 3.7572178840637207,
+ "step": 2284
+ },
+ {
+ "epoch": 31.739965095986037,
+ "grad_norm": 0.05788225308060646,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 3.7474923133850098,
+ "step": 2285
+ },
+ {
+ "epoch": 31.75392670157068,
+ "grad_norm": 0.057152729481458664,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 3.7224936485290527,
+ "step": 2286
+ },
+ {
+ "epoch": 31.767888307155324,
+ "grad_norm": 0.0560687892138958,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 3.7336764335632324,
+ "step": 2287
+ },
+ {
+ "epoch": 31.781849912739965,
+ "grad_norm": 0.05890286713838577,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 3.754326343536377,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79581151832461,
+ "grad_norm": 0.058598410338163376,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 3.7265028953552246,
+ "step": 2289
+ },
+ {
+ "epoch": 31.80977312390925,
+ "grad_norm": 0.05728591978549957,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 3.715481758117676,
+ "step": 2290
+ },
+ {
+ "epoch": 31.823734729493893,
+ "grad_norm": 0.05874952673912048,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 3.7139101028442383,
+ "step": 2291
+ },
+ {
+ "epoch": 31.837696335078533,
+ "grad_norm": 0.056882575154304504,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 3.728553533554077,
+ "step": 2292
+ },
+ {
+ "epoch": 31.851657940663177,
+ "grad_norm": 0.057810861617326736,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 3.757908582687378,
+ "step": 2293
+ },
+ {
+ "epoch": 31.865619546247817,
+ "grad_norm": 0.057366516441106796,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 3.7265639305114746,
+ "step": 2294
+ },
+ {
+ "epoch": 31.87958115183246,
+ "grad_norm": 0.059405453503131866,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 3.740511417388916,
+ "step": 2295
+ },
+ {
+ "epoch": 31.8935427574171,
+ "grad_norm": 0.05881514027714729,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 3.751497268676758,
+ "step": 2296
+ },
+ {
+ "epoch": 31.907504363001745,
+ "grad_norm": 0.06030813232064247,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 3.75439715385437,
+ "step": 2297
+ },
+ {
+ "epoch": 31.921465968586386,
+ "grad_norm": 0.06062643602490425,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 3.7215723991394043,
+ "step": 2298
+ },
+ {
+ "epoch": 31.93542757417103,
+ "grad_norm": 0.05953266844153404,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 3.7635340690612793,
+ "step": 2299
+ },
+ {
+ "epoch": 31.949389179755673,
+ "grad_norm": 0.06019049137830734,
+ "learning_rate": 0.000186516746349841,
+ "loss": 3.7750091552734375,
+ "step": 2300
+ },
+ {
+ "epoch": 31.963350785340314,
+ "grad_norm": 0.058755744248628616,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 3.7646124362945557,
+ "step": 2301
+ },
+ {
+ "epoch": 31.977312390924958,
+ "grad_norm": 0.05986425653100014,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 3.7566630840301514,
+ "step": 2302
+ },
+ {
+ "epoch": 31.991273996509598,
+ "grad_norm": 0.057212650775909424,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 3.714637279510498,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.04401736706495285,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 2.33223032951355,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 0.6091130375862122,
+ "eval_runtime": 60.7334,
+ "eval_samples_per_second": 40.209,
+ "eval_steps_per_second": 0.642,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01396160558464,
+ "grad_norm": 0.05650010704994202,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 3.702322244644165,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02792321116929,
+ "grad_norm": 0.054223258048295975,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 3.699272632598877,
+ "step": 2306
+ },
+ {
+ "epoch": 32.04188481675393,
+ "grad_norm": 0.05616949126124382,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 3.6740047931671143,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05584642233857,
+ "grad_norm": 0.05479196831583977,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 3.6942710876464844,
+ "step": 2308
+ },
+ {
+ "epoch": 32.06980802792321,
+ "grad_norm": 0.05735877528786659,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 3.6724581718444824,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083769633507856,
+ "grad_norm": 0.056006450206041336,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 3.6736817359924316,
+ "step": 2310
+ },
+ {
+ "epoch": 32.0977312390925,
+ "grad_norm": 0.055535659193992615,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 3.710303783416748,
+ "step": 2311
+ },
+ {
+ "epoch": 32.11169284467714,
+ "grad_norm": 0.05510629341006279,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 3.669536590576172,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12565445026178,
+ "grad_norm": 0.05426035821437836,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 3.670957088470459,
+ "step": 2313
+ },
+ {
+ "epoch": 32.139616055846425,
+ "grad_norm": 0.05547798424959183,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 3.7050318717956543,
+ "step": 2314
+ },
+ {
+ "epoch": 32.153577661431065,
+ "grad_norm": 0.05412966012954712,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 3.6849310398101807,
+ "step": 2315
+ },
+ {
+ "epoch": 32.167539267015705,
+ "grad_norm": 0.0554192028939724,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 3.702563762664795,
+ "step": 2316
+ },
+ {
+ "epoch": 32.181500872600346,
+ "grad_norm": 0.054564788937568665,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 3.707547187805176,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19546247818499,
+ "grad_norm": 0.05424094945192337,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 3.650768280029297,
+ "step": 2318
+ },
+ {
+ "epoch": 32.20942408376963,
+ "grad_norm": 0.05562194436788559,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 3.6763458251953125,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223385689354274,
+ "grad_norm": 0.053762275725603104,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 3.690493583679199,
+ "step": 2320
+ },
+ {
+ "epoch": 32.23734729493892,
+ "grad_norm": 0.05513778328895569,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 3.6990034580230713,
+ "step": 2321
+ },
+ {
+ "epoch": 32.25130890052356,
+ "grad_norm": 0.05438566580414772,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 3.702357292175293,
+ "step": 2322
+ },
+ {
+ "epoch": 32.2652705061082,
+ "grad_norm": 0.054521508514881134,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 3.701066255569458,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27923211169284,
+ "grad_norm": 0.05416509136557579,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 3.673935651779175,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29319371727749,
+ "grad_norm": 0.05350432172417641,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 3.6740450859069824,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30715532286213,
+ "grad_norm": 0.054736558347940445,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 3.710815668106079,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32111692844677,
+ "grad_norm": 0.055077534168958664,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 3.6732773780822754,
+ "step": 2327
+ },
+ {
+ "epoch": 32.33507853403141,
+ "grad_norm": 0.054749298840761185,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 3.6978728771209717,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34904013961606,
+ "grad_norm": 0.05378924310207367,
+ "learning_rate": 0.000179445406945268,
+ "loss": 3.6783430576324463,
+ "step": 2329
+ },
+ {
+ "epoch": 32.3630017452007,
+ "grad_norm": 0.0553874596953392,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 3.6876044273376465,
+ "step": 2330
+ },
+ {
+ "epoch": 32.37696335078534,
+ "grad_norm": 0.053926583379507065,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 3.712519884109497,
+ "step": 2331
+ },
+ {
+ "epoch": 32.390924956369986,
+ "grad_norm": 0.055778294801712036,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 3.686643123626709,
+ "step": 2332
+ },
+ {
+ "epoch": 32.404886561954626,
+ "grad_norm": 0.05676745995879173,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 3.685737371444702,
+ "step": 2333
+ },
+ {
+ "epoch": 32.41884816753927,
+ "grad_norm": 0.054717857390642166,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 3.6684136390686035,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43280977312391,
+ "grad_norm": 0.0552208311855793,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 3.696580410003662,
+ "step": 2335
+ },
+ {
+ "epoch": 32.446771378708554,
+ "grad_norm": 0.0549275167286396,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 3.7011008262634277,
+ "step": 2336
+ },
+ {
+ "epoch": 32.460732984293195,
+ "grad_norm": 0.05518808960914612,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 3.7204928398132324,
+ "step": 2337
+ },
+ {
+ "epoch": 32.474694589877835,
+ "grad_norm": 0.05621159076690674,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 3.68092679977417,
+ "step": 2338
+ },
+ {
+ "epoch": 32.488656195462475,
+ "grad_norm": 0.054848797619342804,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 3.677091598510742,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50261780104712,
+ "grad_norm": 0.0562039278447628,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 3.7283501625061035,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51657940663176,
+ "grad_norm": 0.05434577167034149,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 3.710252285003662,
+ "step": 2341
+ },
+ {
+ "epoch": 32.5305410122164,
+ "grad_norm": 0.05737415328621864,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 3.7027816772460938,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544502617801044,
+ "grad_norm": 0.057093650102615356,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 3.7027580738067627,
+ "step": 2343
+ },
+ {
+ "epoch": 32.55846422338569,
+ "grad_norm": 0.0550924576818943,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 3.68161678314209,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57242582897033,
+ "grad_norm": 0.057579416781663895,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 3.67604923248291,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58638743455497,
+ "grad_norm": 0.05515696108341217,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 3.693655490875244,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60034904013962,
+ "grad_norm": 0.05788129195570946,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 3.725567102432251,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61431064572426,
+ "grad_norm": 0.05617021024227142,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 3.7027430534362793,
+ "step": 2348
+ },
+ {
+ "epoch": 32.6282722513089,
+ "grad_norm": 0.05649486556649208,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 3.717970371246338,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64223385689354,
+ "grad_norm": 0.05582305043935776,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 3.6795706748962402,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65619546247819,
+ "grad_norm": 0.058177400380373,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 3.6926321983337402,
+ "step": 2351
+ },
+ {
+ "epoch": 32.67015706806283,
+ "grad_norm": 0.05563648045063019,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 3.7227001190185547,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68411867364747,
+ "grad_norm": 0.058799829334020615,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 3.706979513168335,
+ "step": 2353
+ },
+ {
+ "epoch": 32.69808027923211,
+ "grad_norm": 0.05645359680056572,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 3.6872620582580566,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712041884816756,
+ "grad_norm": 0.058064281940460205,
+ "learning_rate": 0.000173176617304673,
+ "loss": 3.695530414581299,
+ "step": 2355
+ },
+ {
+ "epoch": 32.726003490401396,
+ "grad_norm": 0.05756758525967598,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 3.7229819297790527,
+ "step": 2356
+ },
+ {
+ "epoch": 32.73996509598604,
+ "grad_norm": 0.057592105120420456,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 3.7098472118377686,
+ "step": 2357
+ },
+ {
+ "epoch": 32.753926701570684,
+ "grad_norm": 0.05855923518538475,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 3.704477310180664,
+ "step": 2358
+ },
+ {
+ "epoch": 32.767888307155324,
+ "grad_norm": 0.05809807404875755,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 3.7231266498565674,
+ "step": 2359
+ },
+ {
+ "epoch": 32.781849912739965,
+ "grad_norm": 0.05663153529167175,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 3.708165168762207,
+ "step": 2360
+ },
+ {
+ "epoch": 32.795811518324605,
+ "grad_norm": 0.05867684260010719,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 3.70230770111084,
+ "step": 2361
+ },
+ {
+ "epoch": 32.80977312390925,
+ "grad_norm": 0.05850773677229881,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 3.7339577674865723,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82373472949389,
+ "grad_norm": 0.05651238560676575,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 3.7185606956481934,
+ "step": 2363
+ },
+ {
+ "epoch": 32.83769633507853,
+ "grad_norm": 0.05758710578083992,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 3.677966356277466,
+ "step": 2364
+ },
+ {
+ "epoch": 32.85165794066317,
+ "grad_norm": 0.057175327092409134,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 3.720883846282959,
+ "step": 2365
+ },
+ {
+ "epoch": 32.86561954624782,
+ "grad_norm": 0.05690391734242439,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 3.706535577774048,
+ "step": 2366
+ },
+ {
+ "epoch": 32.87958115183246,
+ "grad_norm": 0.05758208781480789,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 3.743523120880127,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8935427574171,
+ "grad_norm": 0.05831100419163704,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 3.726163387298584,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90750436300174,
+ "grad_norm": 0.05664364993572235,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 3.691837787628174,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92146596858639,
+ "grad_norm": 0.05847351998090744,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 3.7386794090270996,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93542757417103,
+ "grad_norm": 0.05673616752028465,
+ "learning_rate": 0.00016935382741164,
+ "loss": 3.72623348236084,
+ "step": 2371
+ },
+ {
+ "epoch": 32.94938917975567,
+ "grad_norm": 0.058967288583517075,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 3.733576536178589,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96335078534032,
+ "grad_norm": 0.056366659700870514,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 3.7006077766418457,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97731239092496,
+ "grad_norm": 0.05691685527563095,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 3.7104835510253906,
+ "step": 2374
+ },
+ {
+ "epoch": 32.9912739965096,
+ "grad_norm": 0.05809671804308891,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 3.7175397872924805,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.0423070453107357,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 2.3232760429382324,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 0.6105995178222656,
+ "eval_runtime": 61.6632,
+ "eval_samples_per_second": 39.602,
+ "eval_steps_per_second": 0.632,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01396160558464,
+ "grad_norm": 0.05560882017016411,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 3.6668033599853516,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02792321116929,
+ "grad_norm": 0.053790636360645294,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 3.6379659175872803,
+ "step": 2378
+ },
+ {
+ "epoch": 33.04188481675393,
+ "grad_norm": 0.056180961430072784,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 3.6418814659118652,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05584642233857,
+ "grad_norm": 0.054904621094465256,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 3.651149034500122,
+ "step": 2380
+ },
+ {
+ "epoch": 33.06980802792321,
+ "grad_norm": 0.05298716947436333,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 3.686598777770996,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083769633507856,
+ "grad_norm": 0.053739335387945175,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 3.6473655700683594,
+ "step": 2382
+ },
+ {
+ "epoch": 33.0977312390925,
+ "grad_norm": 0.05417626351118088,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 3.6908068656921387,
+ "step": 2383
+ },
+ {
+ "epoch": 33.11169284467714,
+ "grad_norm": 0.0533599928021431,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 3.6613330841064453,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12565445026178,
+ "grad_norm": 0.05309661850333214,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 3.6534783840179443,
+ "step": 2385
+ },
+ {
+ "epoch": 33.139616055846425,
+ "grad_norm": 0.05432313680648804,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 3.664217948913574,
+ "step": 2386
+ },
+ {
+ "epoch": 33.153577661431065,
+ "grad_norm": 0.054764073342084885,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 3.660922050476074,
+ "step": 2387
+ },
+ {
+ "epoch": 33.167539267015705,
+ "grad_norm": 0.05396303907036781,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 3.662687063217163,
+ "step": 2388
+ },
+ {
+ "epoch": 33.181500872600346,
+ "grad_norm": 0.05540551245212555,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 3.6671347618103027,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19546247818499,
+ "grad_norm": 0.05510672926902771,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 3.6410186290740967,
+ "step": 2390
+ },
+ {
+ "epoch": 33.20942408376963,
+ "grad_norm": 0.055180903524160385,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 3.653822422027588,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223385689354274,
+ "grad_norm": 0.054320868104696274,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 3.6835601329803467,
+ "step": 2392
+ },
+ {
+ "epoch": 33.23734729493892,
+ "grad_norm": 0.05568838492035866,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 3.671053171157837,
+ "step": 2393
+ },
+ {
+ "epoch": 33.25130890052356,
+ "grad_norm": 0.056060485541820526,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 3.682260751724243,
+ "step": 2394
+ },
+ {
+ "epoch": 33.2652705061082,
+ "grad_norm": 0.054811324924230576,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 3.6742751598358154,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27923211169284,
+ "grad_norm": 0.05570835620164871,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 3.6490235328674316,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29319371727749,
+ "grad_norm": 0.05547381564974785,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 3.6795434951782227,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30715532286213,
+ "grad_norm": 0.05463161692023277,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 3.655003547668457,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32111692844677,
+ "grad_norm": 0.05583210662007332,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 3.680175304412842,
+ "step": 2399
+ },
+ {
+ "epoch": 33.33507853403141,
+ "grad_norm": 0.056169651448726654,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 3.667175054550171,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34904013961606,
+ "grad_norm": 0.055788591504096985,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 3.6673946380615234,
+ "step": 2401
+ },
+ {
+ "epoch": 33.3630017452007,
+ "grad_norm": 0.05556197836995125,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 3.650700569152832,
+ "step": 2402
+ },
+ {
+ "epoch": 33.37696335078534,
+ "grad_norm": 0.05749715119600296,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 3.6977670192718506,
+ "step": 2403
+ },
+ {
+ "epoch": 33.390924956369986,
+ "grad_norm": 0.05672239512205124,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 3.6604905128479004,
+ "step": 2404
+ },
+ {
+ "epoch": 33.404886561954626,
+ "grad_norm": 0.05661430582404137,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 3.64540958404541,
+ "step": 2405
+ },
+ {
+ "epoch": 33.41884816753927,
+ "grad_norm": 0.05520932003855705,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 3.6648027896881104,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43280977312391,
+ "grad_norm": 0.0557708777487278,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 3.6694397926330566,
+ "step": 2407
+ },
+ {
+ "epoch": 33.446771378708554,
+ "grad_norm": 0.05510152503848076,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 3.6723294258117676,
+ "step": 2408
+ },
+ {
+ "epoch": 33.460732984293195,
+ "grad_norm": 0.05714292824268341,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 3.6610422134399414,
+ "step": 2409
+ },
+ {
+ "epoch": 33.474694589877835,
+ "grad_norm": 0.05525747686624527,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 3.655020236968994,
+ "step": 2410
+ },
+ {
+ "epoch": 33.488656195462475,
+ "grad_norm": 0.05610140040516853,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 3.6707541942596436,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50261780104712,
+ "grad_norm": 0.05564311146736145,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 3.699244976043701,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51657940663176,
+ "grad_norm": 0.05607445910573006,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 3.6682703495025635,
+ "step": 2413
+ },
+ {
+ "epoch": 33.5305410122164,
+ "grad_norm": 0.05434789881110191,
+ "learning_rate": 0.000159218843594243,
+ "loss": 3.6799025535583496,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544502617801044,
+ "grad_norm": 0.055795323103666306,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 3.6433563232421875,
+ "step": 2415
+ },
+ {
+ "epoch": 33.55846422338569,
+ "grad_norm": 0.05676727741956711,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 3.665454864501953,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57242582897033,
+ "grad_norm": 0.05657052621245384,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 3.6618518829345703,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58638743455497,
+ "grad_norm": 0.056275662034749985,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 3.6946463584899902,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60034904013962,
+ "grad_norm": 0.056533295661211014,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 3.6695780754089355,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61431064572426,
+ "grad_norm": 0.05692996829748154,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 3.6850693225860596,
+ "step": 2420
+ },
+ {
+ "epoch": 33.6282722513089,
+ "grad_norm": 0.0577891543507576,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 3.668393611907959,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64223385689354,
+ "grad_norm": 0.05549122393131256,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 3.6772003173828125,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65619546247819,
+ "grad_norm": 0.05636308714747429,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 3.6818785667419434,
+ "step": 2423
+ },
+ {
+ "epoch": 33.67015706806283,
+ "grad_norm": 0.05681503564119339,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 3.6955409049987793,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68411867364747,
+ "grad_norm": 0.05746293440461159,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 3.6917951107025146,
+ "step": 2425
+ },
+ {
+ "epoch": 33.69808027923211,
+ "grad_norm": 0.056621894240379333,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 3.635894775390625,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712041884816756,
+ "grad_norm": 0.05611368268728256,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 3.679736375808716,
+ "step": 2427
+ },
+ {
+ "epoch": 33.726003490401396,
+ "grad_norm": 0.057769183069467545,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 3.6751227378845215,
+ "step": 2428
+ },
+ {
+ "epoch": 33.73996509598604,
+ "grad_norm": 0.05695759132504463,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 3.685178756713867,
+ "step": 2429
+ },
+ {
+ "epoch": 33.753926701570684,
+ "grad_norm": 0.05776882544159889,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 3.6355032920837402,
+ "step": 2430
+ },
+ {
+ "epoch": 33.767888307155324,
+ "grad_norm": 0.056043099611997604,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 3.668668270111084,
+ "step": 2431
+ },
+ {
+ "epoch": 33.781849912739965,
+ "grad_norm": 0.05634639039635658,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 3.6788253784179688,
+ "step": 2432
+ },
+ {
+ "epoch": 33.795811518324605,
+ "grad_norm": 0.05703914538025856,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 3.676163911819458,
+ "step": 2433
+ },
+ {
+ "epoch": 33.80977312390925,
+ "grad_norm": 0.05846793204545975,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 3.657705307006836,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82373472949389,
+ "grad_norm": 0.059191398322582245,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 3.70918607711792,
+ "step": 2435
+ },
+ {
+ "epoch": 33.83769633507853,
+ "grad_norm": 0.056207913905382156,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 3.690638542175293,
+ "step": 2436
+ },
+ {
+ "epoch": 33.85165794066317,
+ "grad_norm": 0.0590900182723999,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 3.6839122772216797,
+ "step": 2437
+ },
+ {
+ "epoch": 33.86561954624782,
+ "grad_norm": 0.05649874731898308,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 3.6689414978027344,
+ "step": 2438
+ },
+ {
+ "epoch": 33.87958115183246,
+ "grad_norm": 0.05947998911142349,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 3.653146743774414,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8935427574171,
+ "grad_norm": 0.05769696831703186,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 3.685544967651367,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90750436300174,
+ "grad_norm": 0.05837882682681084,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 3.6648831367492676,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92146596858639,
+ "grad_norm": 0.05676107108592987,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 3.7186074256896973,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93542757417103,
+ "grad_norm": 0.05771078169345856,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 3.678924083709717,
+ "step": 2443
+ },
+ {
+ "epoch": 33.94938917975567,
+ "grad_norm": 0.05686977133154869,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 3.7026078701019287,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96335078534032,
+ "grad_norm": 0.0571015290915966,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 3.6880578994750977,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97731239092496,
+ "grad_norm": 0.056974101811647415,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 3.6521928310394287,
+ "step": 2446
+ },
+ {
+ "epoch": 33.9912739965096,
+ "grad_norm": 0.05962125584483147,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 3.686126708984375,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.043099287897348404,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 2.308753490447998,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 0.6120397448539734,
+ "eval_runtime": 60.5299,
+ "eval_samples_per_second": 40.344,
+ "eval_steps_per_second": 0.644,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01396160558464,
+ "grad_norm": 0.05469869077205658,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 3.6253180503845215,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02792321116929,
+ "grad_norm": 0.05455945432186127,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 3.651854991912842,
+ "step": 2450
+ },
+ {
+ "epoch": 34.04188481675393,
+ "grad_norm": 0.05369517207145691,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 3.630152702331543,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05584642233857,
+ "grad_norm": 0.05498119816184044,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 3.6293938159942627,
+ "step": 2452
+ },
+ {
+ "epoch": 34.06980802792321,
+ "grad_norm": 0.05530959740281105,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 3.6548690795898438,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083769633507856,
+ "grad_norm": 0.055240705609321594,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 3.6408982276916504,
+ "step": 2454
+ },
+ {
+ "epoch": 34.0977312390925,
+ "grad_norm": 0.05358482152223587,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 3.6274750232696533,
+ "step": 2455
+ },
+ {
+ "epoch": 34.11169284467714,
+ "grad_norm": 0.05580627918243408,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 3.6357412338256836,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12565445026178,
+ "grad_norm": 0.05458527058362961,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 3.6637980937957764,
+ "step": 2457
+ },
+ {
+ "epoch": 34.139616055846425,
+ "grad_norm": 0.05548330023884773,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 3.627671718597412,
+ "step": 2458
+ },
+ {
+ "epoch": 34.153577661431065,
+ "grad_norm": 0.05482204258441925,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 3.642627239227295,
+ "step": 2459
+ },
+ {
+ "epoch": 34.167539267015705,
+ "grad_norm": 0.05435148626565933,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 3.6398377418518066,
+ "step": 2460
+ },
+ {
+ "epoch": 34.181500872600346,
+ "grad_norm": 0.05556599423289299,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 3.62994384765625,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19546247818499,
+ "grad_norm": 0.05394179746508598,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 3.593183994293213,
+ "step": 2462
+ },
+ {
+ "epoch": 34.20942408376963,
+ "grad_norm": 0.05472215637564659,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 3.645197868347168,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223385689354274,
+ "grad_norm": 0.05505116656422615,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 3.655001163482666,
+ "step": 2464
+ },
+ {
+ "epoch": 34.23734729493892,
+ "grad_norm": 0.05473910644650459,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 3.636812210083008,
+ "step": 2465
+ },
+ {
+ "epoch": 34.25130890052356,
+ "grad_norm": 0.05537406727671623,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 3.620631217956543,
+ "step": 2466
+ },
+ {
+ "epoch": 34.2652705061082,
+ "grad_norm": 0.054202884435653687,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 3.6500911712646484,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27923211169284,
+ "grad_norm": 0.0536428838968277,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 3.6268038749694824,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29319371727749,
+ "grad_norm": 0.053046368062496185,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 3.646955966949463,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30715532286213,
+ "grad_norm": 0.05279967561364174,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 3.6142497062683105,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32111692844677,
+ "grad_norm": 0.05348867550492287,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 3.643517017364502,
+ "step": 2471
+ },
+ {
+ "epoch": 34.33507853403141,
+ "grad_norm": 0.05406436324119568,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 3.6528172492980957,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34904013961606,
+ "grad_norm": 0.05249491706490517,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 3.6664018630981445,
+ "step": 2473
+ },
+ {
+ "epoch": 34.3630017452007,
+ "grad_norm": 0.053803686052560806,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 3.609346389770508,
+ "step": 2474
+ },
+ {
+ "epoch": 34.37696335078534,
+ "grad_norm": 0.054445188492536545,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 3.6647889614105225,
+ "step": 2475
+ },
+ {
+ "epoch": 34.390924956369986,
+ "grad_norm": 0.0550406239926815,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 3.656118154525757,
+ "step": 2476
+ },
+ {
+ "epoch": 34.404886561954626,
+ "grad_norm": 0.054776549339294434,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 3.6493523120880127,
+ "step": 2477
+ },
+ {
+ "epoch": 34.41884816753927,
+ "grad_norm": 0.054829373955726624,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 3.634632110595703,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43280977312391,
+ "grad_norm": 0.054575446993112564,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 3.636216878890991,
+ "step": 2479
+ },
+ {
+ "epoch": 34.446771378708554,
+ "grad_norm": 0.056146737188100815,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 3.6584529876708984,
+ "step": 2480
+ },
+ {
+ "epoch": 34.460732984293195,
+ "grad_norm": 0.054963648319244385,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 3.6363425254821777,
+ "step": 2481
+ },
+ {
+ "epoch": 34.474694589877835,
+ "grad_norm": 0.05659656599164009,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 3.6401073932647705,
+ "step": 2482
+ },
+ {
+ "epoch": 34.488656195462475,
+ "grad_norm": 0.05448099970817566,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 3.6161837577819824,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50261780104712,
+ "grad_norm": 0.057917334139347076,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 3.6802477836608887,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51657940663176,
+ "grad_norm": 0.056171346455812454,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 3.652339220046997,
+ "step": 2485
+ },
+ {
+ "epoch": 34.5305410122164,
+ "grad_norm": 0.05800231546163559,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 3.6437389850616455,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544502617801044,
+ "grad_norm": 0.05693773180246353,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 3.6268343925476074,
+ "step": 2487
+ },
+ {
+ "epoch": 34.55846422338569,
+ "grad_norm": 0.05424632877111435,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 3.6246471405029297,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57242582897033,
+ "grad_norm": 0.05645301938056946,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 3.649033308029175,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58638743455497,
+ "grad_norm": 0.056913044303655624,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 3.658982515335083,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60034904013962,
+ "grad_norm": 0.05770258232951164,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 3.621516704559326,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61431064572426,
+ "grad_norm": 0.05483509227633476,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 3.659385919570923,
+ "step": 2492
+ },
+ {
+ "epoch": 34.6282722513089,
+ "grad_norm": 0.057591512799263,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 3.6512298583984375,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64223385689354,
+ "grad_norm": 0.055053744465112686,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 3.630310297012329,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65619546247819,
+ "grad_norm": 0.056070517748594284,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 3.6766881942749023,
+ "step": 2495
+ },
+ {
+ "epoch": 34.67015706806283,
+ "grad_norm": 0.056556008756160736,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 3.6454544067382812,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68411867364747,
+ "grad_norm": 0.05624367296695709,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 3.6482372283935547,
+ "step": 2497
+ },
+ {
+ "epoch": 34.69808027923211,
+ "grad_norm": 0.056266799569129944,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 3.6528916358947754,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712041884816756,
+ "grad_norm": 0.05540672317147255,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 3.636371374130249,
+ "step": 2499
+ },
+ {
+ "epoch": 34.726003490401396,
+ "grad_norm": 0.05527781695127487,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 3.6440672874450684,
+ "step": 2500
+ },
+ {
+ "epoch": 34.73996509598604,
+ "grad_norm": 0.05645914003252983,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 3.6747167110443115,
+ "step": 2501
+ },
+ {
+ "epoch": 34.753926701570684,
+ "grad_norm": 0.055584684014320374,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 3.637603282928467,
+ "step": 2502
+ },
+ {
+ "epoch": 34.767888307155324,
+ "grad_norm": 0.0554841123521328,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 3.6546664237976074,
+ "step": 2503
+ },
+ {
+ "epoch": 34.781849912739965,
+ "grad_norm": 0.0557134784758091,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 3.6366615295410156,
+ "step": 2504
+ },
+ {
+ "epoch": 34.795811518324605,
+ "grad_norm": 0.056569263339042664,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 3.657125949859619,
+ "step": 2505
+ },
+ {
+ "epoch": 34.80977312390925,
+ "grad_norm": 0.05737468600273132,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 3.6409783363342285,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82373472949389,
+ "grad_norm": 0.05602499470114708,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 3.6670620441436768,
+ "step": 2507
+ },
+ {
+ "epoch": 34.83769633507853,
+ "grad_norm": 0.059535399079322815,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 3.6585612297058105,
+ "step": 2508
+ },
+ {
+ "epoch": 34.85165794066317,
+ "grad_norm": 0.05610577017068863,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 3.6544179916381836,
+ "step": 2509
+ },
+ {
+ "epoch": 34.86561954624782,
+ "grad_norm": 0.057512786239385605,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 3.658529281616211,
+ "step": 2510
+ },
+ {
+ "epoch": 34.87958115183246,
+ "grad_norm": 0.05605233088135719,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 3.622344493865967,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8935427574171,
+ "grad_norm": 0.055069468915462494,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 3.6812562942504883,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90750436300174,
+ "grad_norm": 0.05756533145904541,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 3.651491165161133,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92146596858639,
+ "grad_norm": 0.05596763640642166,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 3.633408784866333,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93542757417103,
+ "grad_norm": 0.057164326310157776,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 3.632655620574951,
+ "step": 2515
+ },
+ {
+ "epoch": 34.94938917975567,
+ "grad_norm": 0.059363316744565964,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 3.637542486190796,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96335078534032,
+ "grad_norm": 0.05572650581598282,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 3.666663885116577,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97731239092496,
+ "grad_norm": 0.05825378745794296,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 3.654174327850342,
+ "step": 2518
+ },
+ {
+ "epoch": 34.9912739965096,
+ "grad_norm": 0.058226607739925385,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 3.630509376525879,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.041674062609672546,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 2.2687463760375977,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 0.6132010817527771,
+ "eval_runtime": 58.9507,
+ "eval_samples_per_second": 41.424,
+ "eval_steps_per_second": 0.662,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01396160558464,
+ "grad_norm": 0.053901202976703644,
+ "learning_rate": 0.0001349604788260823,
+ "loss": 3.6262593269348145,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02792321116929,
+ "grad_norm": 0.051058750599622726,
+ "learning_rate": 0.00013474073268599152,
+ "loss": 3.596529483795166,
+ "step": 2522
+ },
+ {
+ "epoch": 35.04188481675393,
+ "grad_norm": 0.05482964590191841,
+ "learning_rate": 0.00013452112210935339,
+ "loss": 3.6197128295898438,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05584642233857,
+ "grad_norm": 0.05166293680667877,
+ "learning_rate": 0.00013430164727310399,
+ "loss": 3.6150989532470703,
+ "step": 2524
+ },
+ {
+ "epoch": 35.06980802792321,
+ "grad_norm": 0.051854558289051056,
+ "learning_rate": 0.00013408230835407027,
+ "loss": 3.594954013824463,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083769633507856,
+ "grad_norm": 0.05247301980853081,
+ "learning_rate": 0.00013386310552896977,
+ "loss": 3.6349754333496094,
+ "step": 2526
+ },
+ {
+ "epoch": 35.0977312390925,
+ "grad_norm": 0.05379469692707062,
+ "learning_rate": 0.00013364403897441016,
+ "loss": 3.611581802368164,
+ "step": 2527
+ },
+ {
+ "epoch": 35.11169284467714,
+ "grad_norm": 0.053294215351343155,
+ "learning_rate": 0.00013342510886688944,
+ "loss": 3.5933470726013184,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12565445026178,
+ "grad_norm": 0.05274510383605957,
+ "learning_rate": 0.00013320631538279548,
+ "loss": 3.6055941581726074,
+ "step": 2529
+ },
+ {
+ "epoch": 35.139616055846425,
+ "grad_norm": 0.05449166148900986,
+ "learning_rate": 0.00013298765869840658,
+ "loss": 3.5891900062561035,
+ "step": 2530
+ },
+ {
+ "epoch": 35.153577661431065,
+ "grad_norm": 0.05245492234826088,
+ "learning_rate": 0.00013276913898989013,
+ "loss": 3.614109992980957,
+ "step": 2531
+ },
+ {
+ "epoch": 35.167539267015705,
+ "grad_norm": 0.05323590710759163,
+ "learning_rate": 0.00013255075643330369,
+ "loss": 3.6126725673675537,
+ "step": 2532
+ },
+ {
+ "epoch": 35.181500872600346,
+ "grad_norm": 0.05381206050515175,
+ "learning_rate": 0.00013233251120459407,
+ "loss": 3.5820865631103516,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19546247818499,
+ "grad_norm": 0.054380469024181366,
+ "learning_rate": 0.0001321144034795976,
+ "loss": 3.6311044692993164,
+ "step": 2534
+ },
+ {
+ "epoch": 35.20942408376963,
+ "grad_norm": 0.053128909319639206,
+ "learning_rate": 0.00013189643343403977,
+ "loss": 3.5665154457092285,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223385689354274,
+ "grad_norm": 0.052764274179935455,
+ "learning_rate": 0.00013167860124353476,
+ "loss": 3.609287738800049,
+ "step": 2536
+ },
+ {
+ "epoch": 35.23734729493892,
+ "grad_norm": 0.05363981053233147,
+ "learning_rate": 0.00013146090708358657,
+ "loss": 3.6260581016540527,
+ "step": 2537
+ },
+ {
+ "epoch": 35.25130890052356,
+ "grad_norm": 0.05316983535885811,
+ "learning_rate": 0.00013124335112958704,
+ "loss": 3.589477300643921,
+ "step": 2538
+ },
+ {
+ "epoch": 35.2652705061082,
+ "grad_norm": 0.05451323837041855,
+ "learning_rate": 0.0001310259335568172,
+ "loss": 3.621201515197754,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27923211169284,
+ "grad_norm": 0.053846996277570724,
+ "learning_rate": 0.00013080865454044647,
+ "loss": 3.610521078109741,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29319371727749,
+ "grad_norm": 0.05316411331295967,
+ "learning_rate": 0.00013059151425553263,
+ "loss": 3.6240267753601074,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30715532286213,
+ "grad_norm": 0.053650014102458954,
+ "learning_rate": 0.00013037451287702183,
+ "loss": 3.6098155975341797,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32111692844677,
+ "grad_norm": 0.054066307842731476,
+ "learning_rate": 0.0001301576505797479,
+ "loss": 3.619396686553955,
+ "step": 2543
+ },
+ {
+ "epoch": 35.33507853403141,
+ "grad_norm": 0.05257592722773552,
+ "learning_rate": 0.00012994092753843293,
+ "loss": 3.655679225921631,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34904013961606,
+ "grad_norm": 0.05437186732888222,
+ "learning_rate": 0.00012972434392768687,
+ "loss": 3.6167373657226562,
+ "step": 2545
+ },
+ {
+ "epoch": 35.3630017452007,
+ "grad_norm": 0.05342245474457741,
+ "learning_rate": 0.00012950789992200714,
+ "loss": 3.5970587730407715,
+ "step": 2546
+ },
+ {
+ "epoch": 35.37696335078534,
+ "grad_norm": 0.054295122623443604,
+ "learning_rate": 0.00012929159569577886,
+ "loss": 3.5975115299224854,
+ "step": 2547
+ },
+ {
+ "epoch": 35.390924956369986,
+ "grad_norm": 0.05349617451429367,
+ "learning_rate": 0.0001290754314232743,
+ "loss": 3.603315591812134,
+ "step": 2548
+ },
+ {
+ "epoch": 35.404886561954626,
+ "grad_norm": 0.05614254251122475,
+ "learning_rate": 0.00012885940727865334,
+ "loss": 3.6229186058044434,
+ "step": 2549
+ },
+ {
+ "epoch": 35.41884816753927,
+ "grad_norm": 0.054123859852552414,
+ "learning_rate": 0.00012864352343596247,
+ "loss": 3.641162395477295,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43280977312391,
+ "grad_norm": 0.05414072424173355,
+ "learning_rate": 0.0001284277800691355,
+ "loss": 3.614049196243286,
+ "step": 2551
+ },
+ {
+ "epoch": 35.446771378708554,
+ "grad_norm": 0.05422188341617584,
+ "learning_rate": 0.00012821217735199298,
+ "loss": 3.5817785263061523,
+ "step": 2552
+ },
+ {
+ "epoch": 35.460732984293195,
+ "grad_norm": 0.05528900399804115,
+ "learning_rate": 0.00012799671545824212,
+ "loss": 3.618821859359741,
+ "step": 2553
+ },
+ {
+ "epoch": 35.474694589877835,
+ "grad_norm": 0.0527317114174366,
+ "learning_rate": 0.0001277813945614768,
+ "loss": 3.6165270805358887,
+ "step": 2554
+ },
+ {
+ "epoch": 35.488656195462475,
+ "grad_norm": 0.05585351586341858,
+ "learning_rate": 0.00012756621483517682,
+ "loss": 3.6158246994018555,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50261780104712,
+ "grad_norm": 0.05375446751713753,
+ "learning_rate": 0.00012735117645270905,
+ "loss": 3.610105037689209,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51657940663176,
+ "grad_norm": 0.055324699729681015,
+ "learning_rate": 0.00012713627958732567,
+ "loss": 3.6106624603271484,
+ "step": 2557
+ },
+ {
+ "epoch": 35.5305410122164,
+ "grad_norm": 0.0544339083135128,
+ "learning_rate": 0.00012692152441216539,
+ "loss": 3.605295181274414,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544502617801044,
+ "grad_norm": 0.05522064492106438,
+ "learning_rate": 0.0001267069111002525,
+ "loss": 3.607207775115967,
+ "step": 2559
+ },
+ {
+ "epoch": 35.55846422338569,
+ "grad_norm": 0.055271610617637634,
+ "learning_rate": 0.00012649243982449718,
+ "loss": 3.610219955444336,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57242582897033,
+ "grad_norm": 0.05532364174723625,
+ "learning_rate": 0.00012627811075769502,
+ "loss": 3.6226587295532227,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58638743455497,
+ "grad_norm": 0.05356820300221443,
+ "learning_rate": 0.00012606392407252687,
+ "loss": 3.628182888031006,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60034904013962,
+ "grad_norm": 0.054853104054927826,
+ "learning_rate": 0.00012584987994155943,
+ "loss": 3.596072196960449,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61431064572426,
+ "grad_norm": 0.05462052673101425,
+ "learning_rate": 0.00012563597853724388,
+ "loss": 3.6201977729797363,
+ "step": 2564
+ },
+ {
+ "epoch": 35.6282722513089,
+ "grad_norm": 0.05532895773649216,
+ "learning_rate": 0.000125422220031917,
+ "loss": 3.642716407775879,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64223385689354,
+ "grad_norm": 0.055425263941287994,
+ "learning_rate": 0.00012520860459779972,
+ "loss": 3.5959243774414062,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65619546247819,
+ "grad_norm": 0.056545358151197433,
+ "learning_rate": 0.0001249951324069986,
+ "loss": 3.5796356201171875,
+ "step": 2567
+ },
+ {
+ "epoch": 35.67015706806283,
+ "grad_norm": 0.055357806384563446,
+ "learning_rate": 0.0001247818036315042,
+ "loss": 3.6106321811676025,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68411867364747,
+ "grad_norm": 0.05461106821894646,
+ "learning_rate": 0.00012456861844319155,
+ "loss": 3.591688632965088,
+ "step": 2569
+ },
+ {
+ "epoch": 35.69808027923211,
+ "grad_norm": 0.055133529007434845,
+ "learning_rate": 0.00012435557701382012,
+ "loss": 3.6243996620178223,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712041884816756,
+ "grad_norm": 0.055249396711587906,
+ "learning_rate": 0.0001241426795150336,
+ "loss": 3.608017683029175,
+ "step": 2571
+ },
+ {
+ "epoch": 35.726003490401396,
+ "grad_norm": 0.056398529559373856,
+ "learning_rate": 0.00012392992611835973,
+ "loss": 3.6282615661621094,
+ "step": 2572
+ },
+ {
+ "epoch": 35.73996509598604,
+ "grad_norm": 0.0552498884499073,
+ "learning_rate": 0.0001237173169952098,
+ "loss": 3.6407179832458496,
+ "step": 2573
+ },
+ {
+ "epoch": 35.753926701570684,
+ "grad_norm": 0.056712064892053604,
+ "learning_rate": 0.00012350485231687954,
+ "loss": 3.616793632507324,
+ "step": 2574
+ },
+ {
+ "epoch": 35.767888307155324,
+ "grad_norm": 0.055920615792274475,
+ "learning_rate": 0.0001232925322545476,
+ "loss": 3.6330060958862305,
+ "step": 2575
+ },
+ {
+ "epoch": 35.781849912739965,
+ "grad_norm": 0.05519915744662285,
+ "learning_rate": 0.0001230803569792765,
+ "loss": 3.61584210395813,
+ "step": 2576
+ },
+ {
+ "epoch": 35.795811518324605,
+ "grad_norm": 0.05472308024764061,
+ "learning_rate": 0.0001228683266620121,
+ "loss": 3.6116104125976562,
+ "step": 2577
+ },
+ {
+ "epoch": 35.80977312390925,
+ "grad_norm": 0.0556313656270504,
+ "learning_rate": 0.00012265644147358326,
+ "loss": 3.634107828140259,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82373472949389,
+ "grad_norm": 0.057370636612176895,
+ "learning_rate": 0.00012244470158470226,
+ "loss": 3.6127190589904785,
+ "step": 2579
+ },
+ {
+ "epoch": 35.83769633507853,
+ "grad_norm": 0.05609549954533577,
+ "learning_rate": 0.00012223310716596387,
+ "loss": 3.6323940753936768,
+ "step": 2580
+ },
+ {
+ "epoch": 35.85165794066317,
+ "grad_norm": 0.054692406207323074,
+ "learning_rate": 0.00012202165838784601,
+ "loss": 3.6339163780212402,
+ "step": 2581
+ },
+ {
+ "epoch": 35.86561954624782,
+ "grad_norm": 0.05703910440206528,
+ "learning_rate": 0.00012181035542070913,
+ "loss": 3.6473350524902344,
+ "step": 2582
+ },
+ {
+ "epoch": 35.87958115183246,
+ "grad_norm": 0.055098868906497955,
+ "learning_rate": 0.00012159919843479617,
+ "loss": 3.6110334396362305,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8935427574171,
+ "grad_norm": 0.0570925697684288,
+ "learning_rate": 0.00012138818760023257,
+ "loss": 3.630516767501831,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90750436300174,
+ "grad_norm": 0.05486847832798958,
+ "learning_rate": 0.00012117732308702592,
+ "loss": 3.622447967529297,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92146596858639,
+ "grad_norm": 0.057239681482315063,
+ "learning_rate": 0.00012096660506506605,
+ "loss": 3.6436092853546143,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93542757417103,
+ "grad_norm": 0.05756063386797905,
+ "learning_rate": 0.00012075603370412443,
+ "loss": 3.6313462257385254,
+ "step": 2587
+ },
+ {
+ "epoch": 35.94938917975567,
+ "grad_norm": 0.05791015550494194,
+ "learning_rate": 0.00012054560917385476,
+ "loss": 3.617950677871704,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96335078534032,
+ "grad_norm": 0.05538728088140488,
+ "learning_rate": 0.00012033533164379225,
+ "loss": 3.665337085723877,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97731239092496,
+ "grad_norm": 0.05843896046280861,
+ "learning_rate": 0.0001201252012833537,
+ "loss": 3.6643996238708496,
+ "step": 2590
+ },
+ {
+ "epoch": 35.9912739965096,
+ "grad_norm": 0.05609741806983948,
+ "learning_rate": 0.00011991521826183747,
+ "loss": 3.6212263107299805,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.04305510222911835,
+ "learning_rate": 0.00011970538274842282,
+ "loss": 2.270846128463745,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 0.6145148873329163,
+ "eval_runtime": 60.714,
+ "eval_samples_per_second": 40.221,
+ "eval_steps_per_second": 0.642,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01396160558464,
+ "grad_norm": 0.054062873125076294,
+ "learning_rate": 0.00011949569491217073,
+ "loss": 3.581265926361084,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02792321116929,
+ "grad_norm": 0.051643744111061096,
+ "learning_rate": 0.00011928615492202269,
+ "loss": 3.586836814880371,
+ "step": 2594
+ },
+ {
+ "epoch": 36.04188481675393,
+ "grad_norm": 0.054067034274339676,
+ "learning_rate": 0.0001190767629468014,
+ "loss": 3.6008591651916504,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05584642233857,
+ "grad_norm": 0.05220582336187363,
+ "learning_rate": 0.00011886751915521007,
+ "loss": 3.5975303649902344,
+ "step": 2596
+ },
+ {
+ "epoch": 36.06980802792321,
+ "grad_norm": 0.053417500108480453,
+ "learning_rate": 0.00011865842371583278,
+ "loss": 3.5652084350585938,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083769633507856,
+ "grad_norm": 0.05344022065401077,
+ "learning_rate": 0.00011844947679713396,
+ "loss": 3.561967372894287,
+ "step": 2598
+ },
+ {
+ "epoch": 36.0977312390925,
+ "grad_norm": 0.05316327139735222,
+ "learning_rate": 0.00011824067856745812,
+ "loss": 3.590935707092285,
+ "step": 2599
+ },
+ {
+ "epoch": 36.11169284467714,
+ "grad_norm": 0.054267000406980515,
+ "learning_rate": 0.00011803202919503064,
+ "loss": 3.5788426399230957,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12565445026178,
+ "grad_norm": 0.05376334860920906,
+ "learning_rate": 0.00011782352884795615,
+ "loss": 3.589162826538086,
+ "step": 2601
+ },
+ {
+ "epoch": 36.139616055846425,
+ "grad_norm": 0.05369187891483307,
+ "learning_rate": 0.00011761517769421983,
+ "loss": 3.558194160461426,
+ "step": 2602
+ },
+ {
+ "epoch": 36.153577661431065,
+ "grad_norm": 0.05407107621431351,
+ "learning_rate": 0.00011740697590168635,
+ "loss": 3.571460247039795,
+ "step": 2603
+ },
+ {
+ "epoch": 36.167539267015705,
+ "grad_norm": 0.05279739201068878,
+ "learning_rate": 0.00011719892363810018,
+ "loss": 3.588043689727783,
+ "step": 2604
+ },
+ {
+ "epoch": 36.181500872600346,
+ "grad_norm": 0.05608730763196945,
+ "learning_rate": 0.00011699102107108539,
+ "loss": 3.5704660415649414,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19546247818499,
+ "grad_norm": 0.05290020629763603,
+ "learning_rate": 0.00011678326836814507,
+ "loss": 3.5612759590148926,
+ "step": 2606
+ },
+ {
+ "epoch": 36.20942408376963,
+ "grad_norm": 0.05398614704608917,
+ "learning_rate": 0.00011657566569666198,
+ "loss": 3.579409122467041,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223385689354274,
+ "grad_norm": 0.051791902631521225,
+ "learning_rate": 0.00011636821322389777,
+ "loss": 3.5976991653442383,
+ "step": 2608
+ },
+ {
+ "epoch": 36.23734729493892,
+ "grad_norm": 0.05329076945781708,
+ "learning_rate": 0.00011616091111699333,
+ "loss": 3.6209723949432373,
+ "step": 2609
+ },
+ {
+ "epoch": 36.25130890052356,
+ "grad_norm": 0.052824556827545166,
+ "learning_rate": 0.00011595375954296786,
+ "loss": 3.5926363468170166,
+ "step": 2610
+ },
+ {
+ "epoch": 36.2652705061082,
+ "grad_norm": 0.05411917343735695,
+ "learning_rate": 0.00011574675866871997,
+ "loss": 3.607084274291992,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27923211169284,
+ "grad_norm": 0.0527656190097332,
+ "learning_rate": 0.00011553990866102651,
+ "loss": 3.579601526260376,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29319371727749,
+ "grad_norm": 0.05329589545726776,
+ "learning_rate": 0.00011533320968654265,
+ "loss": 3.5800671577453613,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30715532286213,
+ "grad_norm": 0.05175339803099632,
+ "learning_rate": 0.00011512666191180204,
+ "loss": 3.594729423522949,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32111692844677,
+ "grad_norm": 0.05224638432264328,
+ "learning_rate": 0.00011492026550321653,
+ "loss": 3.567016839981079,
+ "step": 2615
+ },
+ {
+ "epoch": 36.33507853403141,
+ "grad_norm": 0.053442589938640594,
+ "learning_rate": 0.00011471402062707607,
+ "loss": 3.5964465141296387,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34904013961606,
+ "grad_norm": 0.05282003805041313,
+ "learning_rate": 0.00011450792744954827,
+ "loss": 3.5932140350341797,
+ "step": 2617
+ },
+ {
+ "epoch": 36.3630017452007,
+ "grad_norm": 0.054600607603788376,
+ "learning_rate": 0.00011430198613667861,
+ "loss": 3.5916247367858887,
+ "step": 2618
+ },
+ {
+ "epoch": 36.37696335078534,
+ "grad_norm": 0.053483184427022934,
+ "learning_rate": 0.00011409619685439064,
+ "loss": 3.588195323944092,
+ "step": 2619
+ },
+ {
+ "epoch": 36.390924956369986,
+ "grad_norm": 0.052354127168655396,
+ "learning_rate": 0.00011389055976848477,
+ "loss": 3.6031956672668457,
+ "step": 2620
+ },
+ {
+ "epoch": 36.404886561954626,
+ "grad_norm": 0.05411195009946823,
+ "learning_rate": 0.00011368507504463914,
+ "loss": 3.61525559425354,
+ "step": 2621
+ },
+ {
+ "epoch": 36.41884816753927,
+ "grad_norm": 0.055229391902685165,
+ "learning_rate": 0.00011347974284840914,
+ "loss": 3.58837890625,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43280977312391,
+ "grad_norm": 0.05329221487045288,
+ "learning_rate": 0.00011327456334522725,
+ "loss": 3.565725803375244,
+ "step": 2623
+ },
+ {
+ "epoch": 36.446771378708554,
+ "grad_norm": 0.05464653670787811,
+ "learning_rate": 0.00011306953670040275,
+ "loss": 3.6044445037841797,
+ "step": 2624
+ },
+ {
+ "epoch": 36.460732984293195,
+ "grad_norm": 0.05323687940835953,
+ "learning_rate": 0.00011286466307912184,
+ "loss": 3.5857181549072266,
+ "step": 2625
+ },
+ {
+ "epoch": 36.474694589877835,
+ "grad_norm": 0.054866403341293335,
+ "learning_rate": 0.00011265994264644783,
+ "loss": 3.592473030090332,
+ "step": 2626
+ },
+ {
+ "epoch": 36.488656195462475,
+ "grad_norm": 0.05508954077959061,
+ "learning_rate": 0.00011245537556732001,
+ "loss": 3.5929830074310303,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50261780104712,
+ "grad_norm": 0.057712286710739136,
+ "learning_rate": 0.00011225096200655442,
+ "loss": 3.613492488861084,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51657940663176,
+ "grad_norm": 0.05367693305015564,
+ "learning_rate": 0.00011204670212884342,
+ "loss": 3.5849246978759766,
+ "step": 2629
+ },
+ {
+ "epoch": 36.5305410122164,
+ "grad_norm": 0.055529702454805374,
+ "learning_rate": 0.00011184259609875557,
+ "loss": 3.60062313079834,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544502617801044,
+ "grad_norm": 0.05650486797094345,
+ "learning_rate": 0.00011163864408073522,
+ "loss": 3.570746898651123,
+ "step": 2631
+ },
+ {
+ "epoch": 36.55846422338569,
+ "grad_norm": 0.055417656898498535,
+ "learning_rate": 0.00011143484623910293,
+ "loss": 3.5839993953704834,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57242582897033,
+ "grad_norm": 0.055162668228149414,
+ "learning_rate": 0.00011123120273805496,
+ "loss": 3.5858545303344727,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58638743455497,
+ "grad_norm": 0.054991088807582855,
+ "learning_rate": 0.0001110277137416632,
+ "loss": 3.5967633724212646,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60034904013962,
+ "grad_norm": 0.055452100932598114,
+ "learning_rate": 0.00011082437941387512,
+ "loss": 3.6110477447509766,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61431064572426,
+ "grad_norm": 0.05554793030023575,
+ "learning_rate": 0.00011062119991851322,
+ "loss": 3.5871753692626953,
+ "step": 2636
+ },
+ {
+ "epoch": 36.6282722513089,
+ "grad_norm": 0.054689038544893265,
+ "learning_rate": 0.000110418175419276,
+ "loss": 3.590219020843506,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64223385689354,
+ "grad_norm": 0.05510031804442406,
+ "learning_rate": 0.00011021530607973626,
+ "loss": 3.5875062942504883,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65619546247819,
+ "grad_norm": 0.05421333760023117,
+ "learning_rate": 0.00011001259206334235,
+ "loss": 3.602912187576294,
+ "step": 2639
+ },
+ {
+ "epoch": 36.67015706806283,
+ "grad_norm": 0.05532202124595642,
+ "learning_rate": 0.00010981003353341721,
+ "loss": 3.5950570106506348,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68411867364747,
+ "grad_norm": 0.05355961248278618,
+ "learning_rate": 0.00010960763065315864,
+ "loss": 3.5744097232818604,
+ "step": 2641
+ },
+ {
+ "epoch": 36.69808027923211,
+ "grad_norm": 0.05564635619521141,
+ "learning_rate": 0.00010940538358563906,
+ "loss": 3.5986971855163574,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712041884816756,
+ "grad_norm": 0.054115377366542816,
+ "learning_rate": 0.00010920329249380509,
+ "loss": 3.596283197402954,
+ "step": 2643
+ },
+ {
+ "epoch": 36.726003490401396,
+ "grad_norm": 0.05481656640768051,
+ "learning_rate": 0.00010900135754047799,
+ "loss": 3.5836915969848633,
+ "step": 2644
+ },
+ {
+ "epoch": 36.73996509598604,
+ "grad_norm": 0.05519521236419678,
+ "learning_rate": 0.00010879957888835307,
+ "loss": 3.6295974254608154,
+ "step": 2645
+ },
+ {
+ "epoch": 36.753926701570684,
+ "grad_norm": 0.05460044741630554,
+ "learning_rate": 0.00010859795669999981,
+ "loss": 3.5932111740112305,
+ "step": 2646
+ },
+ {
+ "epoch": 36.767888307155324,
+ "grad_norm": 0.054085709154605865,
+ "learning_rate": 0.00010839649113786137,
+ "loss": 3.584069013595581,
+ "step": 2647
+ },
+ {
+ "epoch": 36.781849912739965,
+ "grad_norm": 0.05407894030213356,
+ "learning_rate": 0.00010819518236425514,
+ "loss": 3.5967845916748047,
+ "step": 2648
+ },
+ {
+ "epoch": 36.795811518324605,
+ "grad_norm": 0.05513349175453186,
+ "learning_rate": 0.00010799403054137197,
+ "loss": 3.598283290863037,
+ "step": 2649
+ },
+ {
+ "epoch": 36.80977312390925,
+ "grad_norm": 0.0558769516646862,
+ "learning_rate": 0.00010779303583127609,
+ "loss": 3.60675048828125,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82373472949389,
+ "grad_norm": 0.05440812557935715,
+ "learning_rate": 0.00010759219839590535,
+ "loss": 3.5989885330200195,
+ "step": 2651
+ },
+ {
+ "epoch": 36.83769633507853,
+ "grad_norm": 0.05533481389284134,
+ "learning_rate": 0.00010739151839707089,
+ "loss": 3.6296310424804688,
+ "step": 2652
+ },
+ {
+ "epoch": 36.85165794066317,
+ "grad_norm": 0.055396221578121185,
+ "learning_rate": 0.00010719099599645706,
+ "loss": 3.6243319511413574,
+ "step": 2653
+ },
+ {
+ "epoch": 36.86561954624782,
+ "grad_norm": 0.05608600378036499,
+ "learning_rate": 0.0001069906313556208,
+ "loss": 3.6124401092529297,
+ "step": 2654
+ },
+ {
+ "epoch": 36.87958115183246,
+ "grad_norm": 0.05493833124637604,
+ "learning_rate": 0.00010679042463599263,
+ "loss": 3.6008474826812744,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8935427574171,
+ "grad_norm": 0.055183470249176025,
+ "learning_rate": 0.00010659037599887545,
+ "loss": 3.5892064571380615,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90750436300174,
+ "grad_norm": 0.05589289590716362,
+ "learning_rate": 0.0001063904856054446,
+ "loss": 3.594189405441284,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92146596858639,
+ "grad_norm": 0.05483611673116684,
+ "learning_rate": 0.00010619075361674836,
+ "loss": 3.586012601852417,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93542757417103,
+ "grad_norm": 0.055471133440732956,
+ "learning_rate": 0.0001059911801937071,
+ "loss": 3.602313995361328,
+ "step": 2659
+ },
+ {
+ "epoch": 36.94938917975567,
+ "grad_norm": 0.0559491291642189,
+ "learning_rate": 0.00010579176549711365,
+ "loss": 3.6010072231292725,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96335078534032,
+ "grad_norm": 0.055661238729953766,
+ "learning_rate": 0.00010559250968763265,
+ "loss": 3.6093027591705322,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97731239092496,
+ "grad_norm": 0.05550997704267502,
+ "learning_rate": 0.00010539341292580086,
+ "loss": 3.614647150039673,
+ "step": 2662
+ },
+ {
+ "epoch": 36.9912739965096,
+ "grad_norm": 0.056113287806510925,
+ "learning_rate": 0.00010519447537202729,
+ "loss": 3.5756256580352783,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.04090592637658119,
+ "learning_rate": 0.00010499569718659201,
+ "loss": 2.237126588821411,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 0.6159548759460449,
+ "eval_runtime": 60.4892,
+ "eval_samples_per_second": 40.371,
+ "eval_steps_per_second": 0.645,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01396160558464,
+ "grad_norm": 0.05119108036160469,
+ "learning_rate": 0.00010479707852964713,
+ "loss": 3.5844168663024902,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02792321116929,
+ "grad_norm": 0.050482187420129776,
+ "learning_rate": 0.00010459861956121611,
+ "loss": 3.5596923828125,
+ "step": 2666
+ },
+ {
+ "epoch": 37.04188481675393,
+ "grad_norm": 0.05054338648915291,
+ "learning_rate": 0.00010440032044119383,
+ "loss": 3.533599376678467,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05584642233857,
+ "grad_norm": 0.050228655338287354,
+ "learning_rate": 0.00010420218132934614,
+ "loss": 3.557422637939453,
+ "step": 2668
+ },
+ {
+ "epoch": 37.06980802792321,
+ "grad_norm": 0.04899684339761734,
+ "learning_rate": 0.00010400420238531023,
+ "loss": 3.5452675819396973,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083769633507856,
+ "grad_norm": 0.049607474356889725,
+ "learning_rate": 0.00010380638376859416,
+ "loss": 3.571603298187256,
+ "step": 2670
+ },
+ {
+ "epoch": 37.0977312390925,
+ "grad_norm": 0.04938557744026184,
+ "learning_rate": 0.00010360872563857682,
+ "loss": 3.5608606338500977,
+ "step": 2671
+ },
+ {
+ "epoch": 37.11169284467714,
+ "grad_norm": 0.05051542446017265,
+ "learning_rate": 0.0001034112281545079,
+ "loss": 3.5389974117279053,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12565445026178,
+ "grad_norm": 0.04959443211555481,
+ "learning_rate": 0.00010321389147550725,
+ "loss": 3.5671193599700928,
+ "step": 2673
+ },
+ {
+ "epoch": 37.139616055846425,
+ "grad_norm": 0.0502617284655571,
+ "learning_rate": 0.00010301671576056588,
+ "loss": 3.5699734687805176,
+ "step": 2674
+ },
+ {
+ "epoch": 37.153577661431065,
+ "grad_norm": 0.051576778292655945,
+ "learning_rate": 0.00010281970116854436,
+ "loss": 3.562732219696045,
+ "step": 2675
+ },
+ {
+ "epoch": 37.167539267015705,
+ "grad_norm": 0.05000219866633415,
+ "learning_rate": 0.00010262284785817392,
+ "loss": 3.545559883117676,
+ "step": 2676
+ },
+ {
+ "epoch": 37.181500872600346,
+ "grad_norm": 0.049805279821157455,
+ "learning_rate": 0.00010242615598805574,
+ "loss": 3.5819478034973145,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19546247818499,
+ "grad_norm": 0.05044523999094963,
+ "learning_rate": 0.00010222962571666088,
+ "loss": 3.57983136177063,
+ "step": 2678
+ },
+ {
+ "epoch": 37.20942408376963,
+ "grad_norm": 0.05141332373023033,
+ "learning_rate": 0.00010203325720233032,
+ "loss": 3.5591063499450684,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223385689354274,
+ "grad_norm": 0.05136275291442871,
+ "learning_rate": 0.00010183705060327433,
+ "loss": 3.5776171684265137,
+ "step": 2680
+ },
+ {
+ "epoch": 37.23734729493892,
+ "grad_norm": 0.05119960755109787,
+ "learning_rate": 0.00010164100607757346,
+ "loss": 3.5721652507781982,
+ "step": 2681
+ },
+ {
+ "epoch": 37.25130890052356,
+ "grad_norm": 0.05113738030195236,
+ "learning_rate": 0.00010144512378317687,
+ "loss": 3.5741496086120605,
+ "step": 2682
+ },
+ {
+ "epoch": 37.2652705061082,
+ "grad_norm": 0.05131804198026657,
+ "learning_rate": 0.00010124940387790354,
+ "loss": 3.577519416809082,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27923211169284,
+ "grad_norm": 0.05085485801100731,
+ "learning_rate": 0.00010105384651944148,
+ "loss": 3.5454201698303223,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29319371727749,
+ "grad_norm": 0.05160171166062355,
+ "learning_rate": 0.00010085845186534769,
+ "loss": 3.574143886566162,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30715532286213,
+ "grad_norm": 0.05132637917995453,
+ "learning_rate": 0.00010066322007304819,
+ "loss": 3.5619382858276367,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32111692844677,
+ "grad_norm": 0.05075863003730774,
+ "learning_rate": 0.00010046815129983757,
+ "loss": 3.5536012649536133,
+ "step": 2687
+ },
+ {
+ "epoch": 37.33507853403141,
+ "grad_norm": 0.052456241101026535,
+ "learning_rate": 0.00010027324570287925,
+ "loss": 3.586211681365967,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34904013961606,
+ "grad_norm": 0.051105350255966187,
+ "learning_rate": 0.00010007850343920519,
+ "loss": 3.5617852210998535,
+ "step": 2689
+ },
+ {
+ "epoch": 37.3630017452007,
+ "grad_norm": 0.052098311483860016,
+ "learning_rate": 9.988392466571572e-05,
+ "loss": 3.5693583488464355,
+ "step": 2690
+ },
+ {
+ "epoch": 37.37696335078534,
+ "grad_norm": 0.05069877579808235,
+ "learning_rate": 9.968950953917945e-05,
+ "loss": 3.547999620437622,
+ "step": 2691
+ },
+ {
+ "epoch": 37.390924956369986,
+ "grad_norm": 0.05264188349246979,
+ "learning_rate": 9.949525821623309e-05,
+ "loss": 3.554316520690918,
+ "step": 2692
+ },
+ {
+ "epoch": 37.404886561954626,
+ "grad_norm": 0.05224690958857536,
+ "learning_rate": 9.930117085338156e-05,
+ "loss": 3.531437397003174,
+ "step": 2693
+ },
+ {
+ "epoch": 37.41884816753927,
+ "grad_norm": 0.05134997144341469,
+ "learning_rate": 9.910724760699731e-05,
+ "loss": 3.5924055576324463,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43280977312391,
+ "grad_norm": 0.05177623778581619,
+ "learning_rate": 9.891348863332093e-05,
+ "loss": 3.580904006958008,
+ "step": 2695
+ },
+ {
+ "epoch": 37.446771378708554,
+ "grad_norm": 0.052150946110486984,
+ "learning_rate": 9.871989408846051e-05,
+ "loss": 3.568195343017578,
+ "step": 2696
+ },
+ {
+ "epoch": 37.460732984293195,
+ "grad_norm": 0.05220150575041771,
+ "learning_rate": 9.852646412839178e-05,
+ "loss": 3.5756545066833496,
+ "step": 2697
+ },
+ {
+ "epoch": 37.474694589877835,
+ "grad_norm": 0.053251419216394424,
+ "learning_rate": 9.833319890895756e-05,
+ "loss": 3.5686798095703125,
+ "step": 2698
+ },
+ {
+ "epoch": 37.488656195462475,
+ "grad_norm": 0.05219034478068352,
+ "learning_rate": 9.814009858586817e-05,
+ "loss": 3.5607855319976807,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50261780104712,
+ "grad_norm": 0.05180756375193596,
+ "learning_rate": 9.794716331470131e-05,
+ "loss": 3.567056179046631,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51657940663176,
+ "grad_norm": 0.0524761863052845,
+ "learning_rate": 9.77543932509012e-05,
+ "loss": 3.5638341903686523,
+ "step": 2701
+ },
+ {
+ "epoch": 37.5305410122164,
+ "grad_norm": 0.05237184837460518,
+ "learning_rate": 9.756178854977925e-05,
+ "loss": 3.575751781463623,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544502617801044,
+ "grad_norm": 0.05274934321641922,
+ "learning_rate": 9.736934936651362e-05,
+ "loss": 3.5704338550567627,
+ "step": 2703
+ },
+ {
+ "epoch": 37.55846422338569,
+ "grad_norm": 0.053565170615911484,
+ "learning_rate": 9.717707585614916e-05,
+ "loss": 3.566765785217285,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57242582897033,
+ "grad_norm": 0.05329638719558716,
+ "learning_rate": 9.698496817359698e-05,
+ "loss": 3.588357448577881,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58638743455497,
+ "grad_norm": 0.05382629856467247,
+ "learning_rate": 9.679302647363476e-05,
+ "loss": 3.578054428100586,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60034904013962,
+ "grad_norm": 0.05376645550131798,
+ "learning_rate": 9.660125091090675e-05,
+ "loss": 3.5657711029052734,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61431064572426,
+ "grad_norm": 0.05371774733066559,
+ "learning_rate": 9.64096416399228e-05,
+ "loss": 3.591853141784668,
+ "step": 2708
+ },
+ {
+ "epoch": 37.6282722513089,
+ "grad_norm": 0.055276691913604736,
+ "learning_rate": 9.621819881505918e-05,
+ "loss": 3.5756590366363525,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64223385689354,
+ "grad_norm": 0.05425848439335823,
+ "learning_rate": 9.602692259055767e-05,
+ "loss": 3.5689821243286133,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65619546247819,
+ "grad_norm": 0.0549544058740139,
+ "learning_rate": 9.583581312052646e-05,
+ "loss": 3.5869078636169434,
+ "step": 2711
+ },
+ {
+ "epoch": 37.67015706806283,
+ "grad_norm": 0.05412036553025246,
+ "learning_rate": 9.564487055893867e-05,
+ "loss": 3.57331919670105,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68411867364747,
+ "grad_norm": 0.05490295961499214,
+ "learning_rate": 9.545409505963338e-05,
+ "loss": 3.581331729888916,
+ "step": 2713
+ },
+ {
+ "epoch": 37.69808027923211,
+ "grad_norm": 0.05448489263653755,
+ "learning_rate": 9.526348677631499e-05,
+ "loss": 3.605595588684082,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712041884816756,
+ "grad_norm": 0.05397019162774086,
+ "learning_rate": 9.507304586255311e-05,
+ "loss": 3.5643043518066406,
+ "step": 2715
+ },
+ {
+ "epoch": 37.726003490401396,
+ "grad_norm": 0.05399048700928688,
+ "learning_rate": 9.488277247178267e-05,
+ "loss": 3.5841126441955566,
+ "step": 2716
+ },
+ {
+ "epoch": 37.73996509598604,
+ "grad_norm": 0.055010151118040085,
+ "learning_rate": 9.469266675730319e-05,
+ "loss": 3.5840625762939453,
+ "step": 2717
+ },
+ {
+ "epoch": 37.753926701570684,
+ "grad_norm": 0.05478813126683235,
+ "learning_rate": 9.450272887227983e-05,
+ "loss": 3.56626033782959,
+ "step": 2718
+ },
+ {
+ "epoch": 37.767888307155324,
+ "grad_norm": 0.05396709218621254,
+ "learning_rate": 9.431295896974182e-05,
+ "loss": 3.5726799964904785,
+ "step": 2719
+ },
+ {
+ "epoch": 37.781849912739965,
+ "grad_norm": 0.05348159372806549,
+ "learning_rate": 9.412335720258341e-05,
+ "loss": 3.572139263153076,
+ "step": 2720
+ },
+ {
+ "epoch": 37.795811518324605,
+ "grad_norm": 0.055067192763090134,
+ "learning_rate": 9.393392372356335e-05,
+ "loss": 3.5794320106506348,
+ "step": 2721
+ },
+ {
+ "epoch": 37.80977312390925,
+ "grad_norm": 0.054990511387586594,
+ "learning_rate": 9.374465868530477e-05,
+ "loss": 3.573702335357666,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82373472949389,
+ "grad_norm": 0.05417406186461449,
+ "learning_rate": 9.355556224029515e-05,
+ "loss": 3.574542999267578,
+ "step": 2723
+ },
+ {
+ "epoch": 37.83769633507853,
+ "grad_norm": 0.05412792041897774,
+ "learning_rate": 9.336663454088593e-05,
+ "loss": 3.5837464332580566,
+ "step": 2724
+ },
+ {
+ "epoch": 37.85165794066317,
+ "grad_norm": 0.055187419056892395,
+ "learning_rate": 9.317787573929282e-05,
+ "loss": 3.5841941833496094,
+ "step": 2725
+ },
+ {
+ "epoch": 37.86561954624782,
+ "grad_norm": 0.05523138865828514,
+ "learning_rate": 9.298928598759541e-05,
+ "loss": 3.582552909851074,
+ "step": 2726
+ },
+ {
+ "epoch": 37.87958115183246,
+ "grad_norm": 0.05365348979830742,
+ "learning_rate": 9.280086543773698e-05,
+ "loss": 3.5693717002868652,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8935427574171,
+ "grad_norm": 0.0547465942800045,
+ "learning_rate": 9.26126142415247e-05,
+ "loss": 3.551072597503662,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90750436300174,
+ "grad_norm": 0.05491182208061218,
+ "learning_rate": 9.2424532550629e-05,
+ "loss": 3.573578357696533,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92146596858639,
+ "grad_norm": 0.055363260209560394,
+ "learning_rate": 9.223662051658408e-05,
+ "loss": 3.5630593299865723,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93542757417103,
+ "grad_norm": 0.05482657626271248,
+ "learning_rate": 9.204887829078709e-05,
+ "loss": 3.5579466819763184,
+ "step": 2731
+ },
+ {
+ "epoch": 37.94938917975567,
+ "grad_norm": 0.05555591359734535,
+ "learning_rate": 9.186130602449861e-05,
+ "loss": 3.572491407394409,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96335078534032,
+ "grad_norm": 0.0549536868929863,
+ "learning_rate": 9.167390386884224e-05,
+ "loss": 3.576451301574707,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97731239092496,
+ "grad_norm": 0.0547570176422596,
+ "learning_rate": 9.148667197480455e-05,
+ "loss": 3.5588786602020264,
+ "step": 2734
+ },
+ {
+ "epoch": 37.9912739965096,
+ "grad_norm": 0.055114034563302994,
+ "learning_rate": 9.129961049323494e-05,
+ "loss": 3.578408718109131,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.04000422731041908,
+ "learning_rate": 9.111271957484519e-05,
+ "loss": 2.212750196456909,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 0.6171724796295166,
+ "eval_runtime": 60.0184,
+ "eval_samples_per_second": 40.687,
+ "eval_steps_per_second": 0.65,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01396160558464,
+ "grad_norm": 0.05178803205490112,
+ "learning_rate": 9.092599937021031e-05,
+ "loss": 3.5207724571228027,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02792321116929,
+ "grad_norm": 0.048786960542201996,
+ "learning_rate": 9.073945002976715e-05,
+ "loss": 3.524625778198242,
+ "step": 2738
+ },
+ {
+ "epoch": 38.04188481675393,
+ "grad_norm": 0.05165952071547508,
+ "learning_rate": 9.055307170381522e-05,
+ "loss": 3.5370655059814453,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05584642233857,
+ "grad_norm": 0.05022848770022392,
+ "learning_rate": 9.036686454251613e-05,
+ "loss": 3.52911376953125,
+ "step": 2740
+ },
+ {
+ "epoch": 38.06980802792321,
+ "grad_norm": 0.04977148398756981,
+ "learning_rate": 9.018082869589369e-05,
+ "loss": 3.548173666000366,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083769633507856,
+ "grad_norm": 0.04955532029271126,
+ "learning_rate": 8.999496431383368e-05,
+ "loss": 3.5462260246276855,
+ "step": 2742
+ },
+ {
+ "epoch": 38.0977312390925,
+ "grad_norm": 0.051178816705942154,
+ "learning_rate": 8.980927154608341e-05,
+ "loss": 3.539468288421631,
+ "step": 2743
+ },
+ {
+ "epoch": 38.11169284467714,
+ "grad_norm": 0.05035502463579178,
+ "learning_rate": 8.962375054225255e-05,
+ "loss": 3.5545785427093506,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12565445026178,
+ "grad_norm": 0.04895094409584999,
+ "learning_rate": 8.943840145181177e-05,
+ "loss": 3.535142421722412,
+ "step": 2745
+ },
+ {
+ "epoch": 38.139616055846425,
+ "grad_norm": 0.050836317241191864,
+ "learning_rate": 8.925322442409361e-05,
+ "loss": 3.5562942028045654,
+ "step": 2746
+ },
+ {
+ "epoch": 38.153577661431065,
+ "grad_norm": 0.04898427426815033,
+ "learning_rate": 8.906821960829163e-05,
+ "loss": 3.533984661102295,
+ "step": 2747
+ },
+ {
+ "epoch": 38.167539267015705,
+ "grad_norm": 0.05146025866270065,
+ "learning_rate": 8.888338715346113e-05,
+ "loss": 3.5528769493103027,
+ "step": 2748
+ },
+ {
+ "epoch": 38.181500872600346,
+ "grad_norm": 0.049640901386737823,
+ "learning_rate": 8.869872720851831e-05,
+ "loss": 3.5450353622436523,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19546247818499,
+ "grad_norm": 0.04954954609274864,
+ "learning_rate": 8.851423992224012e-05,
+ "loss": 3.566117763519287,
+ "step": 2750
+ },
+ {
+ "epoch": 38.20942408376963,
+ "grad_norm": 0.04950375854969025,
+ "learning_rate": 8.832992544326473e-05,
+ "loss": 3.549238443374634,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223385689354274,
+ "grad_norm": 0.05026113614439964,
+ "learning_rate": 8.814578392009104e-05,
+ "loss": 3.5252106189727783,
+ "step": 2752
+ },
+ {
+ "epoch": 38.23734729493892,
+ "grad_norm": 0.050842445343732834,
+ "learning_rate": 8.796181550107857e-05,
+ "loss": 3.5351614952087402,
+ "step": 2753
+ },
+ {
+ "epoch": 38.25130890052356,
+ "grad_norm": 0.050328366458415985,
+ "learning_rate": 8.777802033444712e-05,
+ "loss": 3.5556483268737793,
+ "step": 2754
+ },
+ {
+ "epoch": 38.2652705061082,
+ "grad_norm": 0.05114470049738884,
+ "learning_rate": 8.75943985682774e-05,
+ "loss": 3.5427448749542236,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27923211169284,
+ "grad_norm": 0.05121666193008423,
+ "learning_rate": 8.741095035051017e-05,
+ "loss": 3.5361504554748535,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29319371727749,
+ "grad_norm": 0.0501880943775177,
+ "learning_rate": 8.722767582894613e-05,
+ "loss": 3.5314698219299316,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30715532286213,
+ "grad_norm": 0.05276653543114662,
+ "learning_rate": 8.704457515124636e-05,
+ "loss": 3.5526633262634277,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32111692844677,
+ "grad_norm": 0.05037686228752136,
+ "learning_rate": 8.686164846493176e-05,
+ "loss": 3.537220001220703,
+ "step": 2759
+ },
+ {
+ "epoch": 38.33507853403141,
+ "grad_norm": 0.051579032093286514,
+ "learning_rate": 8.667889591738317e-05,
+ "loss": 3.5302658081054688,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34904013961606,
+ "grad_norm": 0.05040337145328522,
+ "learning_rate": 8.649631765584083e-05,
+ "loss": 3.528028964996338,
+ "step": 2761
+ },
+ {
+ "epoch": 38.3630017452007,
+ "grad_norm": 0.05162221938371658,
+ "learning_rate": 8.631391382740482e-05,
+ "loss": 3.5771780014038086,
+ "step": 2762
+ },
+ {
+ "epoch": 38.37696335078534,
+ "grad_norm": 0.05134383216500282,
+ "learning_rate": 8.613168457903459e-05,
+ "loss": 3.532721996307373,
+ "step": 2763
+ },
+ {
+ "epoch": 38.390924956369986,
+ "grad_norm": 0.05137534812092781,
+ "learning_rate": 8.594963005754901e-05,
+ "loss": 3.550847291946411,
+ "step": 2764
+ },
+ {
+ "epoch": 38.404886561954626,
+ "grad_norm": 0.05198913812637329,
+ "learning_rate": 8.57677504096261e-05,
+ "loss": 3.556427478790283,
+ "step": 2765
+ },
+ {
+ "epoch": 38.41884816753927,
+ "grad_norm": 0.05211595818400383,
+ "learning_rate": 8.558604578180301e-05,
+ "loss": 3.579143524169922,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43280977312391,
+ "grad_norm": 0.05178852006793022,
+ "learning_rate": 8.540451632047593e-05,
+ "loss": 3.548065423965454,
+ "step": 2767
+ },
+ {
+ "epoch": 38.446771378708554,
+ "grad_norm": 0.05172461271286011,
+ "learning_rate": 8.522316217189972e-05,
+ "loss": 3.523592948913574,
+ "step": 2768
+ },
+ {
+ "epoch": 38.460732984293195,
+ "grad_norm": 0.05164046585559845,
+ "learning_rate": 8.504198348218821e-05,
+ "loss": 3.5595390796661377,
+ "step": 2769
+ },
+ {
+ "epoch": 38.474694589877835,
+ "grad_norm": 0.05212831869721413,
+ "learning_rate": 8.486098039731384e-05,
+ "loss": 3.517983913421631,
+ "step": 2770
+ },
+ {
+ "epoch": 38.488656195462475,
+ "grad_norm": 0.05294293537735939,
+ "learning_rate": 8.46801530631075e-05,
+ "loss": 3.5454447269439697,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50261780104712,
+ "grad_norm": 0.05115564912557602,
+ "learning_rate": 8.449950162525859e-05,
+ "loss": 3.562196731567383,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51657940663176,
+ "grad_norm": 0.05171671882271767,
+ "learning_rate": 8.431902622931443e-05,
+ "loss": 3.526003360748291,
+ "step": 2773
+ },
+ {
+ "epoch": 38.5305410122164,
+ "grad_norm": 0.052197013050317764,
+ "learning_rate": 8.413872702068119e-05,
+ "loss": 3.5459299087524414,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544502617801044,
+ "grad_norm": 0.05189663916826248,
+ "learning_rate": 8.395860414462238e-05,
+ "loss": 3.572622299194336,
+ "step": 2775
+ },
+ {
+ "epoch": 38.55846422338569,
+ "grad_norm": 0.05171465501189232,
+ "learning_rate": 8.377865774625985e-05,
+ "loss": 3.535243034362793,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57242582897033,
+ "grad_norm": 0.052626628428697586,
+ "learning_rate": 8.359888797057316e-05,
+ "loss": 3.555469036102295,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58638743455497,
+ "grad_norm": 0.05242423340678215,
+ "learning_rate": 8.341929496239958e-05,
+ "loss": 3.5629148483276367,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60034904013962,
+ "grad_norm": 0.051777057349681854,
+ "learning_rate": 8.323987886643402e-05,
+ "loss": 3.5607125759124756,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61431064572426,
+ "grad_norm": 0.05259648710489273,
+ "learning_rate": 8.306063982722855e-05,
+ "loss": 3.5630576610565186,
+ "step": 2780
+ },
+ {
+ "epoch": 38.6282722513089,
+ "grad_norm": 0.05220061168074608,
+ "learning_rate": 8.288157798919318e-05,
+ "loss": 3.545198917388916,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64223385689354,
+ "grad_norm": 0.05149109661579132,
+ "learning_rate": 8.27026934965945e-05,
+ "loss": 3.5277156829833984,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65619546247819,
+ "grad_norm": 0.05252330005168915,
+ "learning_rate": 8.25239864935566e-05,
+ "loss": 3.541299343109131,
+ "step": 2783
+ },
+ {
+ "epoch": 38.67015706806283,
+ "grad_norm": 0.05302104353904724,
+ "learning_rate": 8.23454571240605e-05,
+ "loss": 3.5433006286621094,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68411867364747,
+ "grad_norm": 0.053501565009355545,
+ "learning_rate": 8.216710553194416e-05,
+ "loss": 3.5382585525512695,
+ "step": 2785
+ },
+ {
+ "epoch": 38.69808027923211,
+ "grad_norm": 0.053820304572582245,
+ "learning_rate": 8.198893186090222e-05,
+ "loss": 3.5610904693603516,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712041884816756,
+ "grad_norm": 0.05330271273851395,
+ "learning_rate": 8.181093625448585e-05,
+ "loss": 3.5504231452941895,
+ "step": 2787
+ },
+ {
+ "epoch": 38.726003490401396,
+ "grad_norm": 0.05421965569257736,
+ "learning_rate": 8.163311885610299e-05,
+ "loss": 3.550283432006836,
+ "step": 2788
+ },
+ {
+ "epoch": 38.73996509598604,
+ "grad_norm": 0.053781405091285706,
+ "learning_rate": 8.145547980901791e-05,
+ "loss": 3.5624003410339355,
+ "step": 2789
+ },
+ {
+ "epoch": 38.753926701570684,
+ "grad_norm": 0.05342188850045204,
+ "learning_rate": 8.127801925635126e-05,
+ "loss": 3.524596691131592,
+ "step": 2790
+ },
+ {
+ "epoch": 38.767888307155324,
+ "grad_norm": 0.05213013291358948,
+ "learning_rate": 8.110073734107954e-05,
+ "loss": 3.5316734313964844,
+ "step": 2791
+ },
+ {
+ "epoch": 38.781849912739965,
+ "grad_norm": 0.052924077957868576,
+ "learning_rate": 8.092363420603584e-05,
+ "loss": 3.5611703395843506,
+ "step": 2792
+ },
+ {
+ "epoch": 38.795811518324605,
+ "grad_norm": 0.05358070507645607,
+ "learning_rate": 8.074670999390896e-05,
+ "loss": 3.545983076095581,
+ "step": 2793
+ },
+ {
+ "epoch": 38.80977312390925,
+ "grad_norm": 0.053544074296951294,
+ "learning_rate": 8.056996484724342e-05,
+ "loss": 3.5701904296875,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82373472949389,
+ "grad_norm": 0.054532721638679504,
+ "learning_rate": 8.039339890843958e-05,
+ "loss": 3.543095111846924,
+ "step": 2795
+ },
+ {
+ "epoch": 38.83769633507853,
+ "grad_norm": 0.0532798208296299,
+ "learning_rate": 8.02170123197535e-05,
+ "loss": 3.5167088508605957,
+ "step": 2796
+ },
+ {
+ "epoch": 38.85165794066317,
+ "grad_norm": 0.05314210429787636,
+ "learning_rate": 8.004080522329674e-05,
+ "loss": 3.5611109733581543,
+ "step": 2797
+ },
+ {
+ "epoch": 38.86561954624782,
+ "grad_norm": 0.053358301520347595,
+ "learning_rate": 7.986477776103601e-05,
+ "loss": 3.550701141357422,
+ "step": 2798
+ },
+ {
+ "epoch": 38.87958115183246,
+ "grad_norm": 0.0538727343082428,
+ "learning_rate": 7.968893007479343e-05,
+ "loss": 3.562330722808838,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8935427574171,
+ "grad_norm": 0.05346875637769699,
+ "learning_rate": 7.951326230624658e-05,
+ "loss": 3.6011950969696045,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90750436300174,
+ "grad_norm": 0.05399498715996742,
+ "learning_rate": 7.933777459692756e-05,
+ "loss": 3.552870750427246,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92146596858639,
+ "grad_norm": 0.053817737847566605,
+ "learning_rate": 7.916246708822373e-05,
+ "loss": 3.5696794986724854,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93542757417103,
+ "grad_norm": 0.05272316932678223,
+ "learning_rate": 7.898733992137715e-05,
+ "loss": 3.55513334274292,
+ "step": 2803
+ },
+ {
+ "epoch": 38.94938917975567,
+ "grad_norm": 0.053616899996995926,
+ "learning_rate": 7.881239323748475e-05,
+ "loss": 3.5549991130828857,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96335078534032,
+ "grad_norm": 0.053872063755989075,
+ "learning_rate": 7.863762717749771e-05,
+ "loss": 3.537346839904785,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97731239092496,
+ "grad_norm": 0.054122194647789,
+ "learning_rate": 7.846304188222189e-05,
+ "loss": 3.552178144454956,
+ "step": 2806
+ },
+ {
+ "epoch": 38.9912739965096,
+ "grad_norm": 0.053457897156476974,
+ "learning_rate": 7.828863749231777e-05,
+ "loss": 3.574368715286255,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.04022035375237465,
+ "learning_rate": 7.811441414829958e-05,
+ "loss": 2.2162294387817383,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 0.6184564828872681,
+ "eval_runtime": 60.5596,
+ "eval_samples_per_second": 40.324,
+ "eval_steps_per_second": 0.644,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01396160558464,
+ "grad_norm": 0.04818584397435188,
+ "learning_rate": 7.794037199053598e-05,
+ "loss": 3.5161890983581543,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02792321116929,
+ "grad_norm": 0.04768325760960579,
+ "learning_rate": 7.776651115924959e-05,
+ "loss": 3.5117123126983643,
+ "step": 2810
+ },
+ {
+ "epoch": 39.04188481675393,
+ "grad_norm": 0.049668096005916595,
+ "learning_rate": 7.759283179451704e-05,
+ "loss": 3.5076825618743896,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05584642233857,
+ "grad_norm": 0.04928773269057274,
+ "learning_rate": 7.741933403626848e-05,
+ "loss": 3.522817611694336,
+ "step": 2812
+ },
+ {
+ "epoch": 39.06980802792321,
+ "grad_norm": 0.049023669213056564,
+ "learning_rate": 7.724601802428805e-05,
+ "loss": 3.4985058307647705,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083769633507856,
+ "grad_norm": 0.04910000413656235,
+ "learning_rate": 7.707288389821329e-05,
+ "loss": 3.5706303119659424,
+ "step": 2814
+ },
+ {
+ "epoch": 39.0977312390925,
+ "grad_norm": 0.04971414431929588,
+ "learning_rate": 7.689993179753519e-05,
+ "loss": 3.5187277793884277,
+ "step": 2815
+ },
+ {
+ "epoch": 39.11169284467714,
+ "grad_norm": 0.049427829682826996,
+ "learning_rate": 7.672716186159829e-05,
+ "loss": 3.5230023860931396,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12565445026178,
+ "grad_norm": 0.04888913407921791,
+ "learning_rate": 7.655457422959993e-05,
+ "loss": 3.525777816772461,
+ "step": 2817
+ },
+ {
+ "epoch": 39.139616055846425,
+ "grad_norm": 0.04879172518849373,
+ "learning_rate": 7.638216904059122e-05,
+ "loss": 3.5179781913757324,
+ "step": 2818
+ },
+ {
+ "epoch": 39.153577661431065,
+ "grad_norm": 0.04999098926782608,
+ "learning_rate": 7.620994643347559e-05,
+ "loss": 3.53208327293396,
+ "step": 2819
+ },
+ {
+ "epoch": 39.167539267015705,
+ "grad_norm": 0.051719557493925095,
+ "learning_rate": 7.60379065470098e-05,
+ "loss": 3.500316619873047,
+ "step": 2820
+ },
+ {
+ "epoch": 39.181500872600346,
+ "grad_norm": 0.05016033351421356,
+ "learning_rate": 7.586604951980326e-05,
+ "loss": 3.4964962005615234,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19546247818499,
+ "grad_norm": 0.048761315643787384,
+ "learning_rate": 7.569437549031813e-05,
+ "loss": 3.5171115398406982,
+ "step": 2822
+ },
+ {
+ "epoch": 39.20942408376963,
+ "grad_norm": 0.05029286816716194,
+ "learning_rate": 7.55228845968691e-05,
+ "loss": 3.5414791107177734,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223385689354274,
+ "grad_norm": 0.0502147302031517,
+ "learning_rate": 7.535157697762313e-05,
+ "loss": 3.519467830657959,
+ "step": 2824
+ },
+ {
+ "epoch": 39.23734729493892,
+ "grad_norm": 0.05098641291260719,
+ "learning_rate": 7.518045277059979e-05,
+ "loss": 3.539252519607544,
+ "step": 2825
+ },
+ {
+ "epoch": 39.25130890052356,
+ "grad_norm": 0.04990671947598457,
+ "learning_rate": 7.500951211367068e-05,
+ "loss": 3.5308918952941895,
+ "step": 2826
+ },
+ {
+ "epoch": 39.2652705061082,
+ "grad_norm": 0.04992193728685379,
+ "learning_rate": 7.483875514455979e-05,
+ "loss": 3.521590232849121,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27923211169284,
+ "grad_norm": 0.05016932636499405,
+ "learning_rate": 7.466818200084264e-05,
+ "loss": 3.5208449363708496,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29319371727749,
+ "grad_norm": 0.050798505544662476,
+ "learning_rate": 7.449779281994712e-05,
+ "loss": 3.5616073608398438,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30715532286213,
+ "grad_norm": 0.05047072842717171,
+ "learning_rate": 7.43275877391528e-05,
+ "loss": 3.540660858154297,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32111692844677,
+ "grad_norm": 0.051101770251989365,
+ "learning_rate": 7.415756689559061e-05,
+ "loss": 3.532931327819824,
+ "step": 2831
+ },
+ {
+ "epoch": 39.33507853403141,
+ "grad_norm": 0.049927644431591034,
+ "learning_rate": 7.398773042624341e-05,
+ "loss": 3.5045418739318848,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34904013961606,
+ "grad_norm": 0.05085239186882973,
+ "learning_rate": 7.381807846794532e-05,
+ "loss": 3.4864392280578613,
+ "step": 2833
+ },
+ {
+ "epoch": 39.3630017452007,
+ "grad_norm": 0.05125649273395538,
+ "learning_rate": 7.3648611157382e-05,
+ "loss": 3.51094913482666,
+ "step": 2834
+ },
+ {
+ "epoch": 39.37696335078534,
+ "grad_norm": 0.05072709918022156,
+ "learning_rate": 7.34793286310899e-05,
+ "loss": 3.531437873840332,
+ "step": 2835
+ },
+ {
+ "epoch": 39.390924956369986,
+ "grad_norm": 0.050076793879270554,
+ "learning_rate": 7.331023102545716e-05,
+ "loss": 3.5222954750061035,
+ "step": 2836
+ },
+ {
+ "epoch": 39.404886561954626,
+ "grad_norm": 0.050684161484241486,
+ "learning_rate": 7.314131847672268e-05,
+ "loss": 3.525660991668701,
+ "step": 2837
+ },
+ {
+ "epoch": 39.41884816753927,
+ "grad_norm": 0.05221536383032799,
+ "learning_rate": 7.297259112097608e-05,
+ "loss": 3.547537326812744,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43280977312391,
+ "grad_norm": 0.050070326775312424,
+ "learning_rate": 7.280404909415801e-05,
+ "loss": 3.524541139602661,
+ "step": 2839
+ },
+ {
+ "epoch": 39.446771378708554,
+ "grad_norm": 0.05107896402478218,
+ "learning_rate": 7.263569253205973e-05,
+ "loss": 3.4862544536590576,
+ "step": 2840
+ },
+ {
+ "epoch": 39.460732984293195,
+ "grad_norm": 0.05145040526986122,
+ "learning_rate": 7.24675215703232e-05,
+ "loss": 3.5242197513580322,
+ "step": 2841
+ },
+ {
+ "epoch": 39.474694589877835,
+ "grad_norm": 0.05019835755228996,
+ "learning_rate": 7.229953634444055e-05,
+ "loss": 3.5409276485443115,
+ "step": 2842
+ },
+ {
+ "epoch": 39.488656195462475,
+ "grad_norm": 0.051458705216646194,
+ "learning_rate": 7.213173698975444e-05,
+ "loss": 3.5637054443359375,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50261780104712,
+ "grad_norm": 0.05186862125992775,
+ "learning_rate": 7.1964123641458e-05,
+ "loss": 3.5018935203552246,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51657940663176,
+ "grad_norm": 0.05114978924393654,
+ "learning_rate": 7.179669643459405e-05,
+ "loss": 3.5413007736206055,
+ "step": 2845
+ },
+ {
+ "epoch": 39.5305410122164,
+ "grad_norm": 0.051313482224941254,
+ "learning_rate": 7.162945550405575e-05,
+ "loss": 3.5065975189208984,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544502617801044,
+ "grad_norm": 0.0509142242372036,
+ "learning_rate": 7.146240098458613e-05,
+ "loss": 3.521883964538574,
+ "step": 2847
+ },
+ {
+ "epoch": 39.55846422338569,
+ "grad_norm": 0.05225839465856552,
+ "learning_rate": 7.129553301077801e-05,
+ "loss": 3.534417152404785,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57242582897033,
+ "grad_norm": 0.05145575851202011,
+ "learning_rate": 7.112885171707378e-05,
+ "loss": 3.5341527462005615,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58638743455497,
+ "grad_norm": 0.0504966638982296,
+ "learning_rate": 7.096235723776562e-05,
+ "loss": 3.529374599456787,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60034904013962,
+ "grad_norm": 0.05249404534697533,
+ "learning_rate": 7.079604970699514e-05,
+ "loss": 3.537140130996704,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61431064572426,
+ "grad_norm": 0.052380647510290146,
+ "learning_rate": 7.062992925875332e-05,
+ "loss": 3.529085159301758,
+ "step": 2852
+ },
+ {
+ "epoch": 39.6282722513089,
+ "grad_norm": 0.05162700638175011,
+ "learning_rate": 7.046399602688045e-05,
+ "loss": 3.549781084060669,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64223385689354,
+ "grad_norm": 0.05245045945048332,
+ "learning_rate": 7.029825014506572e-05,
+ "loss": 3.5424249172210693,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65619546247819,
+ "grad_norm": 0.05116629973053932,
+ "learning_rate": 7.013269174684795e-05,
+ "loss": 3.5407791137695312,
+ "step": 2855
+ },
+ {
+ "epoch": 39.67015706806283,
+ "grad_norm": 0.0507299080491066,
+ "learning_rate": 6.996732096561422e-05,
+ "loss": 3.550326108932495,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68411867364747,
+ "grad_norm": 0.05050551891326904,
+ "learning_rate": 6.980213793460092e-05,
+ "loss": 3.50667667388916,
+ "step": 2857
+ },
+ {
+ "epoch": 39.69808027923211,
+ "grad_norm": 0.052720312029123306,
+ "learning_rate": 6.963714278689304e-05,
+ "loss": 3.5426530838012695,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712041884816756,
+ "grad_norm": 0.05040572211146355,
+ "learning_rate": 6.947233565542417e-05,
+ "loss": 3.50482177734375,
+ "step": 2859
+ },
+ {
+ "epoch": 39.726003490401396,
+ "grad_norm": 0.05204465985298157,
+ "learning_rate": 6.930771667297653e-05,
+ "loss": 3.5070583820343018,
+ "step": 2860
+ },
+ {
+ "epoch": 39.73996509598604,
+ "grad_norm": 0.05265089496970177,
+ "learning_rate": 6.914328597218042e-05,
+ "loss": 3.53663969039917,
+ "step": 2861
+ },
+ {
+ "epoch": 39.753926701570684,
+ "grad_norm": 0.05254863202571869,
+ "learning_rate": 6.8979043685515e-05,
+ "loss": 3.5625903606414795,
+ "step": 2862
+ },
+ {
+ "epoch": 39.767888307155324,
+ "grad_norm": 0.05163634940981865,
+ "learning_rate": 6.881498994530708e-05,
+ "loss": 3.541079044342041,
+ "step": 2863
+ },
+ {
+ "epoch": 39.781849912739965,
+ "grad_norm": 0.05227374657988548,
+ "learning_rate": 6.865112488373186e-05,
+ "loss": 3.52247953414917,
+ "step": 2864
+ },
+ {
+ "epoch": 39.795811518324605,
+ "grad_norm": 0.051065072417259216,
+ "learning_rate": 6.84874486328125e-05,
+ "loss": 3.5435099601745605,
+ "step": 2865
+ },
+ {
+ "epoch": 39.80977312390925,
+ "grad_norm": 0.05301819369196892,
+ "learning_rate": 6.832396132441993e-05,
+ "loss": 3.551717758178711,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82373472949389,
+ "grad_norm": 0.052438024431467056,
+ "learning_rate": 6.816066309027309e-05,
+ "loss": 3.545943260192871,
+ "step": 2867
+ },
+ {
+ "epoch": 39.83769633507853,
+ "grad_norm": 0.05212542042136192,
+ "learning_rate": 6.799755406193815e-05,
+ "loss": 3.5160136222839355,
+ "step": 2868
+ },
+ {
+ "epoch": 39.85165794066317,
+ "grad_norm": 0.05442605912685394,
+ "learning_rate": 6.783463437082929e-05,
+ "loss": 3.5255932807922363,
+ "step": 2869
+ },
+ {
+ "epoch": 39.86561954624782,
+ "grad_norm": 0.05317089334130287,
+ "learning_rate": 6.767190414820792e-05,
+ "loss": 3.5328879356384277,
+ "step": 2870
+ },
+ {
+ "epoch": 39.87958115183246,
+ "grad_norm": 0.05377183482050896,
+ "learning_rate": 6.750936352518284e-05,
+ "loss": 3.512451171875,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8935427574171,
+ "grad_norm": 0.05238938704133034,
+ "learning_rate": 6.734701263271011e-05,
+ "loss": 3.5253477096557617,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90750436300174,
+ "grad_norm": 0.0536922849714756,
+ "learning_rate": 6.718485160159289e-05,
+ "loss": 3.532573938369751,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92146596858639,
+ "grad_norm": 0.05387766659259796,
+ "learning_rate": 6.702288056248145e-05,
+ "loss": 3.535872459411621,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93542757417103,
+ "grad_norm": 0.05426036939024925,
+ "learning_rate": 6.686109964587285e-05,
+ "loss": 3.534602165222168,
+ "step": 2875
+ },
+ {
+ "epoch": 39.94938917975567,
+ "grad_norm": 0.05351133272051811,
+ "learning_rate": 6.669950898211106e-05,
+ "loss": 3.507835865020752,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96335078534032,
+ "grad_norm": 0.05391782522201538,
+ "learning_rate": 6.653810870138676e-05,
+ "loss": 3.543053150177002,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97731239092496,
+ "grad_norm": 0.05348298326134682,
+ "learning_rate": 6.637689893373729e-05,
+ "loss": 3.528005361557007,
+ "step": 2878
+ },
+ {
+ "epoch": 39.9912739965096,
+ "grad_norm": 0.05260506272315979,
+ "learning_rate": 6.621587980904651e-05,
+ "loss": 3.542234420776367,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.040236134082078934,
+ "learning_rate": 6.605505145704437e-05,
+ "loss": 2.2050814628601074,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 0.6198474764823914,
+ "eval_runtime": 59.2234,
+ "eval_samples_per_second": 41.234,
+ "eval_steps_per_second": 0.659,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01396160558464,
+ "grad_norm": 0.048478540033102036,
+ "learning_rate": 6.58944140073077e-05,
+ "loss": 3.524498224258423,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02792321116929,
+ "grad_norm": 0.04602562263607979,
+ "learning_rate": 6.573396758925891e-05,
+ "loss": 3.5231332778930664,
+ "step": 2882
+ },
+ {
+ "epoch": 40.04188481675393,
+ "grad_norm": 0.04725645110011101,
+ "learning_rate": 6.557371233216692e-05,
+ "loss": 3.4903368949890137,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05584642233857,
+ "grad_norm": 0.04776613414287567,
+ "learning_rate": 6.541364836514646e-05,
+ "loss": 3.4806885719299316,
+ "step": 2884
+ },
+ {
+ "epoch": 40.06980802792321,
+ "grad_norm": 0.04646219685673714,
+ "learning_rate": 6.525377581715829e-05,
+ "loss": 3.5238184928894043,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083769633507856,
+ "grad_norm": 0.04782753065228462,
+ "learning_rate": 6.509409481700864e-05,
+ "loss": 3.4997220039367676,
+ "step": 2886
+ },
+ {
+ "epoch": 40.0977312390925,
+ "grad_norm": 0.04805218428373337,
+ "learning_rate": 6.49346054933496e-05,
+ "loss": 3.493037700653076,
+ "step": 2887
+ },
+ {
+ "epoch": 40.11169284467714,
+ "grad_norm": 0.047421909868717194,
+ "learning_rate": 6.477530797467911e-05,
+ "loss": 3.510756492614746,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12565445026178,
+ "grad_norm": 0.04791928082704544,
+ "learning_rate": 6.461620238934006e-05,
+ "loss": 3.503185510635376,
+ "step": 2889
+ },
+ {
+ "epoch": 40.139616055846425,
+ "grad_norm": 0.047295596450567245,
+ "learning_rate": 6.445728886552109e-05,
+ "loss": 3.5190024375915527,
+ "step": 2890
+ },
+ {
+ "epoch": 40.153577661431065,
+ "grad_norm": 0.047632548958063126,
+ "learning_rate": 6.429856753125573e-05,
+ "loss": 3.5584514141082764,
+ "step": 2891
+ },
+ {
+ "epoch": 40.167539267015705,
+ "grad_norm": 0.04803183302283287,
+ "learning_rate": 6.414003851442318e-05,
+ "loss": 3.5070176124572754,
+ "step": 2892
+ },
+ {
+ "epoch": 40.181500872600346,
+ "grad_norm": 0.047644875943660736,
+ "learning_rate": 6.398170194274722e-05,
+ "loss": 3.489863395690918,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19546247818499,
+ "grad_norm": 0.04730125144124031,
+ "learning_rate": 6.38235579437968e-05,
+ "loss": 3.5007224082946777,
+ "step": 2894
+ },
+ {
+ "epoch": 40.20942408376963,
+ "grad_norm": 0.04779860004782677,
+ "learning_rate": 6.366560664498572e-05,
+ "loss": 3.5284690856933594,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223385689354274,
+ "grad_norm": 0.04776514694094658,
+ "learning_rate": 6.350784817357242e-05,
+ "loss": 3.462017774581909,
+ "step": 2896
+ },
+ {
+ "epoch": 40.23734729493892,
+ "grad_norm": 0.04804418608546257,
+ "learning_rate": 6.335028265666023e-05,
+ "loss": 3.5062084197998047,
+ "step": 2897
+ },
+ {
+ "epoch": 40.25130890052356,
+ "grad_norm": 0.047954075038433075,
+ "learning_rate": 6.319291022119653e-05,
+ "loss": 3.4866459369659424,
+ "step": 2898
+ },
+ {
+ "epoch": 40.2652705061082,
+ "grad_norm": 0.0490020215511322,
+ "learning_rate": 6.303573099397378e-05,
+ "loss": 3.4901952743530273,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27923211169284,
+ "grad_norm": 0.04876933991909027,
+ "learning_rate": 6.287874510162821e-05,
+ "loss": 3.5119152069091797,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29319371727749,
+ "grad_norm": 0.0485818088054657,
+ "learning_rate": 6.272195267064063e-05,
+ "loss": 3.4974474906921387,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30715532286213,
+ "grad_norm": 0.04771488532423973,
+ "learning_rate": 6.25653538273358e-05,
+ "loss": 3.523752212524414,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32111692844677,
+ "grad_norm": 0.048272743821144104,
+ "learning_rate": 6.240894869788267e-05,
+ "loss": 3.4994726181030273,
+ "step": 2903
+ },
+ {
+ "epoch": 40.33507853403141,
+ "grad_norm": 0.04918459802865982,
+ "learning_rate": 6.225273740829404e-05,
+ "loss": 3.5262579917907715,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34904013961606,
+ "grad_norm": 0.04880300536751747,
+ "learning_rate": 6.209672008442635e-05,
+ "loss": 3.4975790977478027,
+ "step": 2905
+ },
+ {
+ "epoch": 40.3630017452007,
+ "grad_norm": 0.049021199345588684,
+ "learning_rate": 6.19408968519801e-05,
+ "loss": 3.5101144313812256,
+ "step": 2906
+ },
+ {
+ "epoch": 40.37696335078534,
+ "grad_norm": 0.04934848099946976,
+ "learning_rate": 6.178526783649916e-05,
+ "loss": 3.4800543785095215,
+ "step": 2907
+ },
+ {
+ "epoch": 40.390924956369986,
+ "grad_norm": 0.048395898193120956,
+ "learning_rate": 6.162983316337109e-05,
+ "loss": 3.5071773529052734,
+ "step": 2908
+ },
+ {
+ "epoch": 40.404886561954626,
+ "grad_norm": 0.049257755279541016,
+ "learning_rate": 6.147459295782676e-05,
+ "loss": 3.509805202484131,
+ "step": 2909
+ },
+ {
+ "epoch": 40.41884816753927,
+ "grad_norm": 0.051024191081523895,
+ "learning_rate": 6.131954734494045e-05,
+ "loss": 3.5087671279907227,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43280977312391,
+ "grad_norm": 0.04894109442830086,
+ "learning_rate": 6.11646964496296e-05,
+ "loss": 3.5116591453552246,
+ "step": 2911
+ },
+ {
+ "epoch": 40.446771378708554,
+ "grad_norm": 0.04954129084944725,
+ "learning_rate": 6.101004039665471e-05,
+ "loss": 3.4806766510009766,
+ "step": 2912
+ },
+ {
+ "epoch": 40.460732984293195,
+ "grad_norm": 0.04909701645374298,
+ "learning_rate": 6.085557931061937e-05,
+ "loss": 3.5208868980407715,
+ "step": 2913
+ },
+ {
+ "epoch": 40.474694589877835,
+ "grad_norm": 0.04900852590799332,
+ "learning_rate": 6.070131331597015e-05,
+ "loss": 3.5316522121429443,
+ "step": 2914
+ },
+ {
+ "epoch": 40.488656195462475,
+ "grad_norm": 0.04930015280842781,
+ "learning_rate": 6.054724253699636e-05,
+ "loss": 3.4901177883148193,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50261780104712,
+ "grad_norm": 0.04921771585941315,
+ "learning_rate": 6.03933670978301e-05,
+ "loss": 3.500121593475342,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51657940663176,
+ "grad_norm": 0.05016683414578438,
+ "learning_rate": 6.0239687122445826e-05,
+ "loss": 3.5284152030944824,
+ "step": 2917
+ },
+ {
+ "epoch": 40.5305410122164,
+ "grad_norm": 0.049245234578847885,
+ "learning_rate": 6.0086202734661e-05,
+ "loss": 3.5167784690856934,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544502617801044,
+ "grad_norm": 0.049730222672224045,
+ "learning_rate": 5.9932914058135004e-05,
+ "loss": 3.5100388526916504,
+ "step": 2919
+ },
+ {
+ "epoch": 40.55846422338569,
+ "grad_norm": 0.05063595995306969,
+ "learning_rate": 5.977982121636984e-05,
+ "loss": 3.5240533351898193,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57242582897033,
+ "grad_norm": 0.049625277519226074,
+ "learning_rate": 5.962692433270962e-05,
+ "loss": 3.4961838722229004,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58638743455497,
+ "grad_norm": 0.048995789140462875,
+ "learning_rate": 5.9474223530340634e-05,
+ "loss": 3.519697427749634,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60034904013962,
+ "grad_norm": 0.05007295310497284,
+ "learning_rate": 5.932171893229124e-05,
+ "loss": 3.4762754440307617,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61431064572426,
+ "grad_norm": 0.05034065619111061,
+ "learning_rate": 5.916941066143137e-05,
+ "loss": 3.510772705078125,
+ "step": 2924
+ },
+ {
+ "epoch": 40.6282722513089,
+ "grad_norm": 0.05077948421239853,
+ "learning_rate": 5.9017298840473364e-05,
+ "loss": 3.50581693649292,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64223385689354,
+ "grad_norm": 0.05035000666975975,
+ "learning_rate": 5.8865383591970775e-05,
+ "loss": 3.509791374206543,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65619546247819,
+ "grad_norm": 0.04955745488405228,
+ "learning_rate": 5.8713665038319e-05,
+ "loss": 3.5151772499084473,
+ "step": 2927
+ },
+ {
+ "epoch": 40.67015706806283,
+ "grad_norm": 0.05059986561536789,
+ "learning_rate": 5.856214330175498e-05,
+ "loss": 3.5117061138153076,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68411867364747,
+ "grad_norm": 0.050653424113988876,
+ "learning_rate": 5.841081850435704e-05,
+ "loss": 3.49762225151062,
+ "step": 2929
+ },
+ {
+ "epoch": 40.69808027923211,
+ "grad_norm": 0.049989622086286545,
+ "learning_rate": 5.825969076804488e-05,
+ "loss": 3.5447728633880615,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712041884816756,
+ "grad_norm": 0.05025019496679306,
+ "learning_rate": 5.810876021457925e-05,
+ "loss": 3.500591993331909,
+ "step": 2931
+ },
+ {
+ "epoch": 40.726003490401396,
+ "grad_norm": 0.04964078962802887,
+ "learning_rate": 5.795802696556222e-05,
+ "loss": 3.536098003387451,
+ "step": 2932
+ },
+ {
+ "epoch": 40.73996509598604,
+ "grad_norm": 0.05067792907357216,
+ "learning_rate": 5.78074911424369e-05,
+ "loss": 3.4908149242401123,
+ "step": 2933
+ },
+ {
+ "epoch": 40.753926701570684,
+ "grad_norm": 0.05067761242389679,
+ "learning_rate": 5.765715286648734e-05,
+ "loss": 3.5012640953063965,
+ "step": 2934
+ },
+ {
+ "epoch": 40.767888307155324,
+ "grad_norm": 0.05116443336009979,
+ "learning_rate": 5.7507012258838154e-05,
+ "loss": 3.5251986980438232,
+ "step": 2935
+ },
+ {
+ "epoch": 40.781849912739965,
+ "grad_norm": 0.05042244866490364,
+ "learning_rate": 5.7357069440455095e-05,
+ "loss": 3.493983030319214,
+ "step": 2936
+ },
+ {
+ "epoch": 40.795811518324605,
+ "grad_norm": 0.050487346947193146,
+ "learning_rate": 5.7207324532144454e-05,
+ "loss": 3.51411771774292,
+ "step": 2937
+ },
+ {
+ "epoch": 40.80977312390925,
+ "grad_norm": 0.050840575248003006,
+ "learning_rate": 5.7057777654552785e-05,
+ "loss": 3.528165340423584,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82373472949389,
+ "grad_norm": 0.052268560975790024,
+ "learning_rate": 5.690842892816741e-05,
+ "loss": 3.513277292251587,
+ "step": 2939
+ },
+ {
+ "epoch": 40.83769633507853,
+ "grad_norm": 0.05126330256462097,
+ "learning_rate": 5.675927847331593e-05,
+ "loss": 3.4808921813964844,
+ "step": 2940
+ },
+ {
+ "epoch": 40.85165794066317,
+ "grad_norm": 0.05069379135966301,
+ "learning_rate": 5.661032641016619e-05,
+ "loss": 3.517787456512451,
+ "step": 2941
+ },
+ {
+ "epoch": 40.86561954624782,
+ "grad_norm": 0.051330823451280594,
+ "learning_rate": 5.6461572858726086e-05,
+ "loss": 3.5521302223205566,
+ "step": 2942
+ },
+ {
+ "epoch": 40.87958115183246,
+ "grad_norm": 0.05037380009889603,
+ "learning_rate": 5.6313017938843595e-05,
+ "loss": 3.505967617034912,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8935427574171,
+ "grad_norm": 0.05150613188743591,
+ "learning_rate": 5.6164661770207026e-05,
+ "loss": 3.518454074859619,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90750436300174,
+ "grad_norm": 0.05246536061167717,
+ "learning_rate": 5.6016504472344e-05,
+ "loss": 3.518848180770874,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92146596858639,
+ "grad_norm": 0.052759598940610886,
+ "learning_rate": 5.5868546164622284e-05,
+ "loss": 3.511413097381592,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93542757417103,
+ "grad_norm": 0.05075768381357193,
+ "learning_rate": 5.572078696624917e-05,
+ "loss": 3.4923768043518066,
+ "step": 2947
+ },
+ {
+ "epoch": 40.94938917975567,
+ "grad_norm": 0.051101621240377426,
+ "learning_rate": 5.557322699627168e-05,
+ "loss": 3.545301675796509,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96335078534032,
+ "grad_norm": 0.05053454637527466,
+ "learning_rate": 5.542586637357607e-05,
+ "loss": 3.4888358116149902,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97731239092496,
+ "grad_norm": 0.05186663195490837,
+ "learning_rate": 5.527870521688817e-05,
+ "loss": 3.5280933380126953,
+ "step": 2950
+ },
+ {
+ "epoch": 40.9912739965096,
+ "grad_norm": 0.05131891742348671,
+ "learning_rate": 5.513174364477311e-05,
+ "loss": 3.5003533363342285,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.03856140375137329,
+ "learning_rate": 5.498498177563518e-05,
+ "loss": 2.1855430603027344,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 0.621021568775177,
+ "eval_runtime": 60.3507,
+ "eval_samples_per_second": 40.463,
+ "eval_steps_per_second": 0.646,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01396160558464,
+ "grad_norm": 0.04582727700471878,
+ "learning_rate": 5.4838419727717716e-05,
+ "loss": 3.492494821548462,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02792321116929,
+ "grad_norm": 0.04572693631052971,
+ "learning_rate": 5.4692057619103144e-05,
+ "loss": 3.4678001403808594,
+ "step": 2954
+ },
+ {
+ "epoch": 41.04188481675393,
+ "grad_norm": 0.04531959071755409,
+ "learning_rate": 5.454589556771286e-05,
+ "loss": 3.5139541625976562,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05584642233857,
+ "grad_norm": 0.046629469841718674,
+ "learning_rate": 5.439993369130681e-05,
+ "loss": 3.483097553253174,
+ "step": 2956
+ },
+ {
+ "epoch": 41.06980802792321,
+ "grad_norm": 0.04704796150326729,
+ "learning_rate": 5.4254172107483924e-05,
+ "loss": 3.484652042388916,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083769633507856,
+ "grad_norm": 0.04632614925503731,
+ "learning_rate": 5.410861093368169e-05,
+ "loss": 3.4961838722229004,
+ "step": 2958
+ },
+ {
+ "epoch": 41.0977312390925,
+ "grad_norm": 0.046637099236249924,
+ "learning_rate": 5.3963250287176094e-05,
+ "loss": 3.5225467681884766,
+ "step": 2959
+ },
+ {
+ "epoch": 41.11169284467714,
+ "grad_norm": 0.04615108296275139,
+ "learning_rate": 5.381809028508168e-05,
+ "loss": 3.478269577026367,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12565445026178,
+ "grad_norm": 0.04565376043319702,
+ "learning_rate": 5.367313104435104e-05,
+ "loss": 3.4609298706054688,
+ "step": 2961
+ },
+ {
+ "epoch": 41.139616055846425,
+ "grad_norm": 0.04716965928673744,
+ "learning_rate": 5.352837268177544e-05,
+ "loss": 3.498220443725586,
+ "step": 2962
+ },
+ {
+ "epoch": 41.153577661431065,
+ "grad_norm": 0.04637673497200012,
+ "learning_rate": 5.3383815313983916e-05,
+ "loss": 3.4881229400634766,
+ "step": 2963
+ },
+ {
+ "epoch": 41.167539267015705,
+ "grad_norm": 0.047194816172122955,
+ "learning_rate": 5.323945905744379e-05,
+ "loss": 3.4718289375305176,
+ "step": 2964
+ },
+ {
+ "epoch": 41.181500872600346,
+ "grad_norm": 0.04609799385070801,
+ "learning_rate": 5.3095304028460284e-05,
+ "loss": 3.460082530975342,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19546247818499,
+ "grad_norm": 0.04652848839759827,
+ "learning_rate": 5.2951350343176495e-05,
+ "loss": 3.4704749584198,
+ "step": 2966
+ },
+ {
+ "epoch": 41.20942408376963,
+ "grad_norm": 0.046700265258550644,
+ "learning_rate": 5.2807598117573384e-05,
+ "loss": 3.4888997077941895,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223385689354274,
+ "grad_norm": 0.046414151787757874,
+ "learning_rate": 5.266404746746939e-05,
+ "loss": 3.506500244140625,
+ "step": 2968
+ },
+ {
+ "epoch": 41.23734729493892,
+ "grad_norm": 0.04814400523900986,
+ "learning_rate": 5.252069850852072e-05,
+ "loss": 3.515608787536621,
+ "step": 2969
+ },
+ {
+ "epoch": 41.25130890052356,
+ "grad_norm": 0.046656325459480286,
+ "learning_rate": 5.237755135622107e-05,
+ "loss": 3.4709742069244385,
+ "step": 2970
+ },
+ {
+ "epoch": 41.2652705061082,
+ "grad_norm": 0.04764151945710182,
+ "learning_rate": 5.223460612590159e-05,
+ "loss": 3.5001049041748047,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27923211169284,
+ "grad_norm": 0.04743436351418495,
+ "learning_rate": 5.209186293273044e-05,
+ "loss": 3.4994988441467285,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29319371727749,
+ "grad_norm": 0.04705984145402908,
+ "learning_rate": 5.1949321891713446e-05,
+ "loss": 3.469075918197632,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30715532286213,
+ "grad_norm": 0.04596104845404625,
+ "learning_rate": 5.180698311769338e-05,
+ "loss": 3.5107405185699463,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32111692844677,
+ "grad_norm": 0.04772106930613518,
+ "learning_rate": 5.166484672534983e-05,
+ "loss": 3.4777815341949463,
+ "step": 2975
+ },
+ {
+ "epoch": 41.33507853403141,
+ "grad_norm": 0.04846428334712982,
+ "learning_rate": 5.152291282919961e-05,
+ "loss": 3.5055108070373535,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34904013961606,
+ "grad_norm": 0.04617132991552353,
+ "learning_rate": 5.138118154359632e-05,
+ "loss": 3.5047807693481445,
+ "step": 2977
+ },
+ {
+ "epoch": 41.3630017452007,
+ "grad_norm": 0.04738635942339897,
+ "learning_rate": 5.1239652982730335e-05,
+ "loss": 3.458731174468994,
+ "step": 2978
+ },
+ {
+ "epoch": 41.37696335078534,
+ "grad_norm": 0.046908650547266006,
+ "learning_rate": 5.1098327260628466e-05,
+ "loss": 3.4803903102874756,
+ "step": 2979
+ },
+ {
+ "epoch": 41.390924956369986,
+ "grad_norm": 0.04845135658979416,
+ "learning_rate": 5.09572044911545e-05,
+ "loss": 3.4671194553375244,
+ "step": 2980
+ },
+ {
+ "epoch": 41.404886561954626,
+ "grad_norm": 0.0482504665851593,
+ "learning_rate": 5.0816284788008527e-05,
+ "loss": 3.4812138080596924,
+ "step": 2981
+ },
+ {
+ "epoch": 41.41884816753927,
+ "grad_norm": 0.04697469249367714,
+ "learning_rate": 5.067556826472684e-05,
+ "loss": 3.4910900592803955,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43280977312391,
+ "grad_norm": 0.04753987118601799,
+ "learning_rate": 5.053505503468228e-05,
+ "loss": 3.504899024963379,
+ "step": 2983
+ },
+ {
+ "epoch": 41.446771378708554,
+ "grad_norm": 0.04830222949385643,
+ "learning_rate": 5.0394745211083785e-05,
+ "loss": 3.487231492996216,
+ "step": 2984
+ },
+ {
+ "epoch": 41.460732984293195,
+ "grad_norm": 0.04917479306459427,
+ "learning_rate": 5.025463890697655e-05,
+ "loss": 3.491119146347046,
+ "step": 2985
+ },
+ {
+ "epoch": 41.474694589877835,
+ "grad_norm": 0.047515422105789185,
+ "learning_rate": 5.011473623524159e-05,
+ "loss": 3.4952025413513184,
+ "step": 2986
+ },
+ {
+ "epoch": 41.488656195462475,
+ "grad_norm": 0.04785816743969917,
+ "learning_rate": 4.9975037308595864e-05,
+ "loss": 3.5042450428009033,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50261780104712,
+ "grad_norm": 0.04761168360710144,
+ "learning_rate": 4.983554223959257e-05,
+ "loss": 3.4989638328552246,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51657940663176,
+ "grad_norm": 0.04817018285393715,
+ "learning_rate": 4.9696251140620134e-05,
+ "loss": 3.473546028137207,
+ "step": 2989
+ },
+ {
+ "epoch": 41.5305410122164,
+ "grad_norm": 0.047623272985219955,
+ "learning_rate": 4.9557164123902924e-05,
+ "loss": 3.5037989616394043,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544502617801044,
+ "grad_norm": 0.046707481145858765,
+ "learning_rate": 4.9418281301500844e-05,
+ "loss": 3.521557569503784,
+ "step": 2991
+ },
+ {
+ "epoch": 41.55846422338569,
+ "grad_norm": 0.048793643712997437,
+ "learning_rate": 4.9279602785309365e-05,
+ "loss": 3.4975666999816895,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57242582897033,
+ "grad_norm": 0.048892680555582047,
+ "learning_rate": 4.914112868705911e-05,
+ "loss": 3.499896764755249,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58638743455497,
+ "grad_norm": 0.050392571836709976,
+ "learning_rate": 4.900285911831624e-05,
+ "loss": 3.494752883911133,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60034904013962,
+ "grad_norm": 0.049144912511110306,
+ "learning_rate": 4.886479419048201e-05,
+ "loss": 3.4834141731262207,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61431064572426,
+ "grad_norm": 0.04847243055701256,
+ "learning_rate": 4.872693401479292e-05,
+ "loss": 3.452786922454834,
+ "step": 2996
+ },
+ {
+ "epoch": 41.6282722513089,
+ "grad_norm": 0.04898214712738991,
+ "learning_rate": 4.858927870232039e-05,
+ "loss": 3.495100975036621,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64223385689354,
+ "grad_norm": 0.048891909420490265,
+ "learning_rate": 4.845182836397068e-05,
+ "loss": 3.499159574508667,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65619546247819,
+ "grad_norm": 0.049408622086048126,
+ "learning_rate": 4.831458311048532e-05,
+ "loss": 3.4985241889953613,
+ "step": 2999
+ },
+ {
+ "epoch": 41.67015706806283,
+ "grad_norm": 0.04940970614552498,
+ "learning_rate": 4.817754305244008e-05,
+ "loss": 3.5193381309509277,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68411867364747,
+ "grad_norm": 0.04961863160133362,
+ "learning_rate": 4.804070830024577e-05,
+ "loss": 3.5064754486083984,
+ "step": 3001
+ },
+ {
+ "epoch": 41.69808027923211,
+ "grad_norm": 0.049569353461265564,
+ "learning_rate": 4.7904078964147654e-05,
+ "loss": 3.4965577125549316,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712041884816756,
+ "grad_norm": 0.04980974644422531,
+ "learning_rate": 4.776765515422557e-05,
+ "loss": 3.5100278854370117,
+ "step": 3003
+ },
+ {
+ "epoch": 41.726003490401396,
+ "grad_norm": 0.0492531843483448,
+ "learning_rate": 4.763143698039371e-05,
+ "loss": 3.4880599975585938,
+ "step": 3004
+ },
+ {
+ "epoch": 41.73996509598604,
+ "grad_norm": 0.05068065971136093,
+ "learning_rate": 4.749542455240043e-05,
+ "loss": 3.5037388801574707,
+ "step": 3005
+ },
+ {
+ "epoch": 41.753926701570684,
+ "grad_norm": 0.04909749701619148,
+ "learning_rate": 4.735961797982876e-05,
+ "loss": 3.4796335697174072,
+ "step": 3006
+ },
+ {
+ "epoch": 41.767888307155324,
+ "grad_norm": 0.05050486698746681,
+ "learning_rate": 4.7224017372095414e-05,
+ "loss": 3.4703454971313477,
+ "step": 3007
+ },
+ {
+ "epoch": 41.781849912739965,
+ "grad_norm": 0.0496763177216053,
+ "learning_rate": 4.708862283845143e-05,
+ "loss": 3.494141101837158,
+ "step": 3008
+ },
+ {
+ "epoch": 41.795811518324605,
+ "grad_norm": 0.0503719225525856,
+ "learning_rate": 4.695343448798163e-05,
+ "loss": 3.5245471000671387,
+ "step": 3009
+ },
+ {
+ "epoch": 41.80977312390925,
+ "grad_norm": 0.04876178130507469,
+ "learning_rate": 4.681845242960492e-05,
+ "loss": 3.485657215118408,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82373472949389,
+ "grad_norm": 0.05160895735025406,
+ "learning_rate": 4.668367677207398e-05,
+ "loss": 3.5181169509887695,
+ "step": 3011
+ },
+ {
+ "epoch": 41.83769633507853,
+ "grad_norm": 0.04962378367781639,
+ "learning_rate": 4.654910762397499e-05,
+ "loss": 3.4959146976470947,
+ "step": 3012
+ },
+ {
+ "epoch": 41.85165794066317,
+ "grad_norm": 0.04914521425962448,
+ "learning_rate": 4.6414745093727913e-05,
+ "loss": 3.4929981231689453,
+ "step": 3013
+ },
+ {
+ "epoch": 41.86561954624782,
+ "grad_norm": 0.049995288252830505,
+ "learning_rate": 4.6280589289586255e-05,
+ "loss": 3.4831490516662598,
+ "step": 3014
+ },
+ {
+ "epoch": 41.87958115183246,
+ "grad_norm": 0.04932008311152458,
+ "learning_rate": 4.614664031963692e-05,
+ "loss": 3.4770588874816895,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8935427574171,
+ "grad_norm": 0.050988808274269104,
+ "learning_rate": 4.601289829180004e-05,
+ "loss": 3.5015549659729004,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90750436300174,
+ "grad_norm": 0.049767497926950455,
+ "learning_rate": 4.587936331382934e-05,
+ "loss": 3.480461597442627,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92146596858639,
+ "grad_norm": 0.050207316875457764,
+ "learning_rate": 4.574603549331151e-05,
+ "loss": 3.522735118865967,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93542757417103,
+ "grad_norm": 0.04978947341442108,
+ "learning_rate": 4.561291493766625e-05,
+ "loss": 3.5139577388763428,
+ "step": 3019
+ },
+ {
+ "epoch": 41.94938917975567,
+ "grad_norm": 0.04990898072719574,
+ "learning_rate": 4.5480001754146455e-05,
+ "loss": 3.512300491333008,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96335078534032,
+ "grad_norm": 0.050290387123823166,
+ "learning_rate": 4.5347296049837875e-05,
+ "loss": 3.4925060272216797,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97731239092496,
+ "grad_norm": 0.049074798822402954,
+ "learning_rate": 4.52147979316592e-05,
+ "loss": 3.4772558212280273,
+ "step": 3022
+ },
+ {
+ "epoch": 41.9912739965096,
+ "grad_norm": 0.04929618537425995,
+ "learning_rate": 4.5082507506361574e-05,
+ "loss": 3.4907312393188477,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.03784853219985962,
+ "learning_rate": 4.495042488052904e-05,
+ "loss": 2.182621955871582,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 0.6221460700035095,
+ "eval_runtime": 60.4686,
+ "eval_samples_per_second": 40.385,
+ "eval_steps_per_second": 0.645,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01396160558464,
+ "grad_norm": 0.0446116179227829,
+ "learning_rate": 4.481855016057839e-05,
+ "loss": 3.5182933807373047,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02792321116929,
+ "grad_norm": 0.04415718466043472,
+ "learning_rate": 4.468688345275848e-05,
+ "loss": 3.4862308502197266,
+ "step": 3026
+ },
+ {
+ "epoch": 42.04188481675393,
+ "grad_norm": 0.043365973979234695,
+ "learning_rate": 4.455542486315086e-05,
+ "loss": 3.4973888397216797,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05584642233857,
+ "grad_norm": 0.045916847884655,
+ "learning_rate": 4.442417449766938e-05,
+ "loss": 3.4444098472595215,
+ "step": 3028
+ },
+ {
+ "epoch": 42.06980802792321,
+ "grad_norm": 0.04477037489414215,
+ "learning_rate": 4.429313246206014e-05,
+ "loss": 3.4581756591796875,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083769633507856,
+ "grad_norm": 0.04387771338224411,
+ "learning_rate": 4.416229886190119e-05,
+ "loss": 3.4740853309631348,
+ "step": 3030
+ },
+ {
+ "epoch": 42.0977312390925,
+ "grad_norm": 0.043969713151454926,
+ "learning_rate": 4.403167380260281e-05,
+ "loss": 3.4641780853271484,
+ "step": 3031
+ },
+ {
+ "epoch": 42.11169284467714,
+ "grad_norm": 0.045231763273477554,
+ "learning_rate": 4.3901257389407475e-05,
+ "loss": 3.481898069381714,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12565445026178,
+ "grad_norm": 0.043941181153059006,
+ "learning_rate": 4.3771049727389075e-05,
+ "loss": 3.479301929473877,
+ "step": 3033
+ },
+ {
+ "epoch": 42.139616055846425,
+ "grad_norm": 0.04559679329395294,
+ "learning_rate": 4.364105092145377e-05,
+ "loss": 3.474754810333252,
+ "step": 3034
+ },
+ {
+ "epoch": 42.153577661431065,
+ "grad_norm": 0.044603366404771805,
+ "learning_rate": 4.3511261076339e-05,
+ "loss": 3.469996452331543,
+ "step": 3035
+ },
+ {
+ "epoch": 42.167539267015705,
+ "grad_norm": 0.04484036564826965,
+ "learning_rate": 4.3381680296614394e-05,
+ "loss": 3.474886417388916,
+ "step": 3036
+ },
+ {
+ "epoch": 42.181500872600346,
+ "grad_norm": 0.04458961635828018,
+ "learning_rate": 4.3252308686680626e-05,
+ "loss": 3.4669456481933594,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19546247818499,
+ "grad_norm": 0.04605228826403618,
+ "learning_rate": 4.3123146350770146e-05,
+ "loss": 3.4829368591308594,
+ "step": 3038
+ },
+ {
+ "epoch": 42.20942408376963,
+ "grad_norm": 0.04483795538544655,
+ "learning_rate": 4.2994193392946726e-05,
+ "loss": 3.4611988067626953,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223385689354274,
+ "grad_norm": 0.04524310305714607,
+ "learning_rate": 4.286544991710541e-05,
+ "loss": 3.4750466346740723,
+ "step": 3040
+ },
+ {
+ "epoch": 42.23734729493892,
+ "grad_norm": 0.04514177143573761,
+ "learning_rate": 4.2736916026972576e-05,
+ "loss": 3.456035614013672,
+ "step": 3041
+ },
+ {
+ "epoch": 42.25130890052356,
+ "grad_norm": 0.04549764469265938,
+ "learning_rate": 4.260859182610542e-05,
+ "loss": 3.450291633605957,
+ "step": 3042
+ },
+ {
+ "epoch": 42.2652705061082,
+ "grad_norm": 0.04712701961398125,
+ "learning_rate": 4.2480477417892776e-05,
+ "loss": 3.4654693603515625,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27923211169284,
+ "grad_norm": 0.045718107372522354,
+ "learning_rate": 4.235257290555381e-05,
+ "loss": 3.491769790649414,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29319371727749,
+ "grad_norm": 0.045064907521009445,
+ "learning_rate": 4.222487839213903e-05,
+ "loss": 3.49348783493042,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30715532286213,
+ "grad_norm": 0.04526374116539955,
+ "learning_rate": 4.209739398052956e-05,
+ "loss": 3.4869022369384766,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32111692844677,
+ "grad_norm": 0.045719344168901443,
+ "learning_rate": 4.1970119773437316e-05,
+ "loss": 3.484297752380371,
+ "step": 3047
+ },
+ {
+ "epoch": 42.33507853403141,
+ "grad_norm": 0.047497447580099106,
+ "learning_rate": 4.184305587340483e-05,
+ "loss": 3.4783167839050293,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34904013961606,
+ "grad_norm": 0.044837966561317444,
+ "learning_rate": 4.171620238280511e-05,
+ "loss": 3.460331678390503,
+ "step": 3049
+ },
+ {
+ "epoch": 42.3630017452007,
+ "grad_norm": 0.0457470640540123,
+ "learning_rate": 4.158955940384179e-05,
+ "loss": 3.4783971309661865,
+ "step": 3050
+ },
+ {
+ "epoch": 42.37696335078534,
+ "grad_norm": 0.04576839506626129,
+ "learning_rate": 4.146312703854881e-05,
+ "loss": 3.4972984790802,
+ "step": 3051
+ },
+ {
+ "epoch": 42.390924956369986,
+ "grad_norm": 0.04589150846004486,
+ "learning_rate": 4.133690538879046e-05,
+ "loss": 3.4743404388427734,
+ "step": 3052
+ },
+ {
+ "epoch": 42.404886561954626,
+ "grad_norm": 0.04680817574262619,
+ "learning_rate": 4.1210894556261226e-05,
+ "loss": 3.4869651794433594,
+ "step": 3053
+ },
+ {
+ "epoch": 42.41884816753927,
+ "grad_norm": 0.04661082476377487,
+ "learning_rate": 4.108509464248578e-05,
+ "loss": 3.47898530960083,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43280977312391,
+ "grad_norm": 0.04658796265721321,
+ "learning_rate": 4.095950574881891e-05,
+ "loss": 3.446974277496338,
+ "step": 3055
+ },
+ {
+ "epoch": 42.446771378708554,
+ "grad_norm": 0.0446607768535614,
+ "learning_rate": 4.083412797644514e-05,
+ "loss": 3.481013298034668,
+ "step": 3056
+ },
+ {
+ "epoch": 42.460732984293195,
+ "grad_norm": 0.04672912880778313,
+ "learning_rate": 4.0708961426379214e-05,
+ "loss": 3.493116855621338,
+ "step": 3057
+ },
+ {
+ "epoch": 42.474694589877835,
+ "grad_norm": 0.04644620046019554,
+ "learning_rate": 4.0584006199465484e-05,
+ "loss": 3.476749897003174,
+ "step": 3058
+ },
+ {
+ "epoch": 42.488656195462475,
+ "grad_norm": 0.0467224083840847,
+ "learning_rate": 4.0459262396378165e-05,
+ "loss": 3.440511703491211,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50261780104712,
+ "grad_norm": 0.045566949993371964,
+ "learning_rate": 4.033473011762116e-05,
+ "loss": 3.4723963737487793,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51657940663176,
+ "grad_norm": 0.04642454907298088,
+ "learning_rate": 4.0210409463527656e-05,
+ "loss": 3.4766931533813477,
+ "step": 3061
+ },
+ {
+ "epoch": 42.5305410122164,
+ "grad_norm": 0.046170495450496674,
+ "learning_rate": 4.008630053426082e-05,
+ "loss": 3.4428517818450928,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544502617801044,
+ "grad_norm": 0.046656228601932526,
+ "learning_rate": 3.996240342981279e-05,
+ "loss": 3.490520477294922,
+ "step": 3063
+ },
+ {
+ "epoch": 42.55846422338569,
+ "grad_norm": 0.04764910787343979,
+ "learning_rate": 3.983871825000529e-05,
+ "loss": 3.478440761566162,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57242582897033,
+ "grad_norm": 0.04596332833170891,
+ "learning_rate": 3.971524509448925e-05,
+ "loss": 3.4757537841796875,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58638743455497,
+ "grad_norm": 0.047473009675741196,
+ "learning_rate": 3.959198406274472e-05,
+ "loss": 3.4758293628692627,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60034904013962,
+ "grad_norm": 0.046200625598430634,
+ "learning_rate": 3.946893525408097e-05,
+ "loss": 3.476270914077759,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61431064572426,
+ "grad_norm": 0.0474882535636425,
+ "learning_rate": 3.934609876763604e-05,
+ "loss": 3.4846224784851074,
+ "step": 3068
+ },
+ {
+ "epoch": 42.6282722513089,
+ "grad_norm": 0.046172887086868286,
+ "learning_rate": 3.922347470237728e-05,
+ "loss": 3.4917209148406982,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64223385689354,
+ "grad_norm": 0.047964416444301605,
+ "learning_rate": 3.910106315710053e-05,
+ "loss": 3.493190050125122,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65619546247819,
+ "grad_norm": 0.048526931554079056,
+ "learning_rate": 3.897886423043061e-05,
+ "loss": 3.471339225769043,
+ "step": 3071
+ },
+ {
+ "epoch": 42.67015706806283,
+ "grad_norm": 0.04729713499546051,
+ "learning_rate": 3.8856878020820975e-05,
+ "loss": 3.4923644065856934,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68411867364747,
+ "grad_norm": 0.0482463575899601,
+ "learning_rate": 3.873510462655375e-05,
+ "loss": 3.496248722076416,
+ "step": 3073
+ },
+ {
+ "epoch": 42.69808027923211,
+ "grad_norm": 0.04724344238638878,
+ "learning_rate": 3.861354414573954e-05,
+ "loss": 3.4864397048950195,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712041884816756,
+ "grad_norm": 0.04727556183934212,
+ "learning_rate": 3.849219667631735e-05,
+ "loss": 3.483755111694336,
+ "step": 3075
+ },
+ {
+ "epoch": 42.726003490401396,
+ "grad_norm": 0.04790109395980835,
+ "learning_rate": 3.8371062316054764e-05,
+ "loss": 3.4795761108398438,
+ "step": 3076
+ },
+ {
+ "epoch": 42.73996509598604,
+ "grad_norm": 0.04786139354109764,
+ "learning_rate": 3.8250141162547466e-05,
+ "loss": 3.502222776412964,
+ "step": 3077
+ },
+ {
+ "epoch": 42.753926701570684,
+ "grad_norm": 0.048031169921159744,
+ "learning_rate": 3.812943331321956e-05,
+ "loss": 3.456325054168701,
+ "step": 3078
+ },
+ {
+ "epoch": 42.767888307155324,
+ "grad_norm": 0.04753265529870987,
+ "learning_rate": 3.8008938865322955e-05,
+ "loss": 3.482483386993408,
+ "step": 3079
+ },
+ {
+ "epoch": 42.781849912739965,
+ "grad_norm": 0.04813816025853157,
+ "learning_rate": 3.788865791593813e-05,
+ "loss": 3.4626238346099854,
+ "step": 3080
+ },
+ {
+ "epoch": 42.795811518324605,
+ "grad_norm": 0.04639475792646408,
+ "learning_rate": 3.7768590561973064e-05,
+ "loss": 3.484825611114502,
+ "step": 3081
+ },
+ {
+ "epoch": 42.80977312390925,
+ "grad_norm": 0.048058658838272095,
+ "learning_rate": 3.764873690016394e-05,
+ "loss": 3.479367733001709,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82373472949389,
+ "grad_norm": 0.04669356718659401,
+ "learning_rate": 3.7529097027074654e-05,
+ "loss": 3.4717469215393066,
+ "step": 3083
+ },
+ {
+ "epoch": 42.83769633507853,
+ "grad_norm": 0.04729066789150238,
+ "learning_rate": 3.7409671039096936e-05,
+ "loss": 3.448122978210449,
+ "step": 3084
+ },
+ {
+ "epoch": 42.85165794066317,
+ "grad_norm": 0.04941864311695099,
+ "learning_rate": 3.7290459032450184e-05,
+ "loss": 3.4794251918792725,
+ "step": 3085
+ },
+ {
+ "epoch": 42.86561954624782,
+ "grad_norm": 0.047708433121442795,
+ "learning_rate": 3.717146110318123e-05,
+ "loss": 3.4708971977233887,
+ "step": 3086
+ },
+ {
+ "epoch": 42.87958115183246,
+ "grad_norm": 0.04743608459830284,
+ "learning_rate": 3.7052677347164614e-05,
+ "loss": 3.487725019454956,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8935427574171,
+ "grad_norm": 0.047558411955833435,
+ "learning_rate": 3.693410786010231e-05,
+ "loss": 3.488107204437256,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90750436300174,
+ "grad_norm": 0.0468779057264328,
+ "learning_rate": 3.6815752737523537e-05,
+ "loss": 3.4700937271118164,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92146596858639,
+ "grad_norm": 0.04860694333910942,
+ "learning_rate": 3.6697612074784944e-05,
+ "loss": 3.4898247718811035,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93542757417103,
+ "grad_norm": 0.0477910116314888,
+ "learning_rate": 3.657968596707029e-05,
+ "loss": 3.497626781463623,
+ "step": 3091
+ },
+ {
+ "epoch": 42.94938917975567,
+ "grad_norm": 0.04593159258365631,
+ "learning_rate": 3.64619745093906e-05,
+ "loss": 3.4628067016601562,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96335078534032,
+ "grad_norm": 0.048044558614492416,
+ "learning_rate": 3.6344477796583714e-05,
+ "loss": 3.4711837768554688,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97731239092496,
+ "grad_norm": 0.04747995361685753,
+ "learning_rate": 3.622719592331472e-05,
+ "loss": 3.483624219894409,
+ "step": 3094
+ },
+ {
+ "epoch": 42.9912739965096,
+ "grad_norm": 0.04785338044166565,
+ "learning_rate": 3.611012898407544e-05,
+ "loss": 3.4870688915252686,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.036172445863485336,
+ "learning_rate": 3.599327707318464e-05,
+ "loss": 2.202502965927124,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 0.6233968734741211,
+ "eval_runtime": 59.5357,
+ "eval_samples_per_second": 41.017,
+ "eval_steps_per_second": 0.655,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01396160558464,
+ "grad_norm": 0.041919440031051636,
+ "learning_rate": 3.587664028478779e-05,
+ "loss": 3.4489996433258057,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02792321116929,
+ "grad_norm": 0.041333068162202835,
+ "learning_rate": 3.5760218712857046e-05,
+ "loss": 3.459347724914551,
+ "step": 3098
+ },
+ {
+ "epoch": 43.04188481675393,
+ "grad_norm": 0.04244478419423103,
+ "learning_rate": 3.564401245119119e-05,
+ "loss": 3.469055414199829,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05584642233857,
+ "grad_norm": 0.04218090698122978,
+ "learning_rate": 3.552802159341543e-05,
+ "loss": 3.4696848392486572,
+ "step": 3100
+ },
+ {
+ "epoch": 43.06980802792321,
+ "grad_norm": 0.04310792684555054,
+ "learning_rate": 3.541224623298155e-05,
+ "loss": 3.445007801055908,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083769633507856,
+ "grad_norm": 0.043118201196193695,
+ "learning_rate": 3.529668646316765e-05,
+ "loss": 3.4675815105438232,
+ "step": 3102
+ },
+ {
+ "epoch": 43.0977312390925,
+ "grad_norm": 0.04221854358911514,
+ "learning_rate": 3.5181342377078166e-05,
+ "loss": 3.4767966270446777,
+ "step": 3103
+ },
+ {
+ "epoch": 43.11169284467714,
+ "grad_norm": 0.043059397488832474,
+ "learning_rate": 3.5066214067643785e-05,
+ "loss": 3.4651613235473633,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12565445026178,
+ "grad_norm": 0.04187485948204994,
+ "learning_rate": 3.495130162762119e-05,
+ "loss": 3.446305751800537,
+ "step": 3105
+ },
+ {
+ "epoch": 43.139616055846425,
+ "grad_norm": 0.04304322227835655,
+ "learning_rate": 3.483660514959343e-05,
+ "loss": 3.4663028717041016,
+ "step": 3106
+ },
+ {
+ "epoch": 43.153577661431065,
+ "grad_norm": 0.04343954101204872,
+ "learning_rate": 3.4722124725969235e-05,
+ "loss": 3.4554028511047363,
+ "step": 3107
+ },
+ {
+ "epoch": 43.167539267015705,
+ "grad_norm": 0.04271049425005913,
+ "learning_rate": 3.460786044898346e-05,
+ "loss": 3.4608445167541504,
+ "step": 3108
+ },
+ {
+ "epoch": 43.181500872600346,
+ "grad_norm": 0.042723026126623154,
+ "learning_rate": 3.4493812410696784e-05,
+ "loss": 3.4895715713500977,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19546247818499,
+ "grad_norm": 0.044030871242284775,
+ "learning_rate": 3.437998070299564e-05,
+ "loss": 3.4766366481781006,
+ "step": 3110
+ },
+ {
+ "epoch": 43.20942408376963,
+ "grad_norm": 0.04285202547907829,
+ "learning_rate": 3.426636541759225e-05,
+ "loss": 3.458911418914795,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223385689354274,
+ "grad_norm": 0.04367215558886528,
+ "learning_rate": 3.415296664602425e-05,
+ "loss": 3.4974024295806885,
+ "step": 3112
+ },
+ {
+ "epoch": 43.23734729493892,
+ "grad_norm": 0.04419645667076111,
+ "learning_rate": 3.403978447965507e-05,
+ "loss": 3.485166549682617,
+ "step": 3113
+ },
+ {
+ "epoch": 43.25130890052356,
+ "grad_norm": 0.04392880201339722,
+ "learning_rate": 3.392681900967355e-05,
+ "loss": 3.459162473678589,
+ "step": 3114
+ },
+ {
+ "epoch": 43.2652705061082,
+ "grad_norm": 0.04220719635486603,
+ "learning_rate": 3.381407032709395e-05,
+ "loss": 3.4838223457336426,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27923211169284,
+ "grad_norm": 0.04468657821416855,
+ "learning_rate": 3.3701538522755714e-05,
+ "loss": 3.4502077102661133,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29319371727749,
+ "grad_norm": 0.04435672610998154,
+ "learning_rate": 3.35892236873238e-05,
+ "loss": 3.4528560638427734,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30715532286213,
+ "grad_norm": 0.044321656227111816,
+ "learning_rate": 3.3477125911288284e-05,
+ "loss": 3.463057041168213,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32111692844677,
+ "grad_norm": 0.044570039957761765,
+ "learning_rate": 3.3365245284964225e-05,
+ "loss": 3.42390775680542,
+ "step": 3119
+ },
+ {
+ "epoch": 43.33507853403141,
+ "grad_norm": 0.04383990913629532,
+ "learning_rate": 3.3253581898491836e-05,
+ "loss": 3.468905210494995,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34904013961606,
+ "grad_norm": 0.044629745185375214,
+ "learning_rate": 3.3142135841836276e-05,
+ "loss": 3.470247983932495,
+ "step": 3121
+ },
+ {
+ "epoch": 43.3630017452007,
+ "grad_norm": 0.04445461556315422,
+ "learning_rate": 3.303090720478773e-05,
+ "loss": 3.4802684783935547,
+ "step": 3122
+ },
+ {
+ "epoch": 43.37696335078534,
+ "grad_norm": 0.04420622065663338,
+ "learning_rate": 3.291989607696092e-05,
+ "loss": 3.470055103302002,
+ "step": 3123
+ },
+ {
+ "epoch": 43.390924956369986,
+ "grad_norm": 0.04278041422367096,
+ "learning_rate": 3.280910254779565e-05,
+ "loss": 3.4415011405944824,
+ "step": 3124
+ },
+ {
+ "epoch": 43.404886561954626,
+ "grad_norm": 0.04414992034435272,
+ "learning_rate": 3.269852670655629e-05,
+ "loss": 3.4492762088775635,
+ "step": 3125
+ },
+ {
+ "epoch": 43.41884816753927,
+ "grad_norm": 0.044683173298835754,
+ "learning_rate": 3.258816864233169e-05,
+ "loss": 3.473803997039795,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43280977312391,
+ "grad_norm": 0.043872565031051636,
+ "learning_rate": 3.247802844403536e-05,
+ "loss": 3.4689323902130127,
+ "step": 3127
+ },
+ {
+ "epoch": 43.446771378708554,
+ "grad_norm": 0.044481247663497925,
+ "learning_rate": 3.236810620040532e-05,
+ "loss": 3.474726438522339,
+ "step": 3128
+ },
+ {
+ "epoch": 43.460732984293195,
+ "grad_norm": 0.044009409844875336,
+ "learning_rate": 3.225840200000396e-05,
+ "loss": 3.4411134719848633,
+ "step": 3129
+ },
+ {
+ "epoch": 43.474694589877835,
+ "grad_norm": 0.04399845376610756,
+ "learning_rate": 3.214891593121788e-05,
+ "loss": 3.4436731338500977,
+ "step": 3130
+ },
+ {
+ "epoch": 43.488656195462475,
+ "grad_norm": 0.04343521595001221,
+ "learning_rate": 3.203964808225801e-05,
+ "loss": 3.452680826187134,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50261780104712,
+ "grad_norm": 0.04366961866617203,
+ "learning_rate": 3.193059854115968e-05,
+ "loss": 3.468557357788086,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51657940663176,
+ "grad_norm": 0.04462398961186409,
+ "learning_rate": 3.182176739578195e-05,
+ "loss": 3.4636874198913574,
+ "step": 3133
+ },
+ {
+ "epoch": 43.5305410122164,
+ "grad_norm": 0.04596228525042534,
+ "learning_rate": 3.171315473380815e-05,
+ "loss": 3.483391284942627,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544502617801044,
+ "grad_norm": 0.04510573297739029,
+ "learning_rate": 3.160476064274555e-05,
+ "loss": 3.4479193687438965,
+ "step": 3135
+ },
+ {
+ "epoch": 43.55846422338569,
+ "grad_norm": 0.044678136706352234,
+ "learning_rate": 3.149658520992537e-05,
+ "loss": 3.471526622772217,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57242582897033,
+ "grad_norm": 0.043749045580625534,
+ "learning_rate": 3.138862852250249e-05,
+ "loss": 3.4809961318969727,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58638743455497,
+ "grad_norm": 0.044585809111595154,
+ "learning_rate": 3.1280890667455694e-05,
+ "loss": 3.4486989974975586,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60034904013962,
+ "grad_norm": 0.04449521377682686,
+ "learning_rate": 3.1173371731587496e-05,
+ "loss": 3.4768877029418945,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61431064572426,
+ "grad_norm": 0.04507734626531601,
+ "learning_rate": 3.1066071801523905e-05,
+ "loss": 3.4656381607055664,
+ "step": 3140
+ },
+ {
+ "epoch": 43.6282722513089,
+ "grad_norm": 0.04444066062569618,
+ "learning_rate": 3.095899096371456e-05,
+ "loss": 3.449824333190918,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64223385689354,
+ "grad_norm": 0.04459245502948761,
+ "learning_rate": 3.0852129304432506e-05,
+ "loss": 3.4757838249206543,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65619546247819,
+ "grad_norm": 0.04539329186081886,
+ "learning_rate": 3.074548690977434e-05,
+ "loss": 3.467472791671753,
+ "step": 3143
+ },
+ {
+ "epoch": 43.67015706806283,
+ "grad_norm": 0.04432376101613045,
+ "learning_rate": 3.0639063865659834e-05,
+ "loss": 3.4804115295410156,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68411867364747,
+ "grad_norm": 0.044664494693279266,
+ "learning_rate": 3.0532860257832144e-05,
+ "loss": 3.458524703979492,
+ "step": 3145
+ },
+ {
+ "epoch": 43.69808027923211,
+ "grad_norm": 0.04534987732768059,
+ "learning_rate": 3.04268761718576e-05,
+ "loss": 3.472733497619629,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712041884816756,
+ "grad_norm": 0.04504983872175217,
+ "learning_rate": 3.0321111693125648e-05,
+ "loss": 3.447598934173584,
+ "step": 3147
+ },
+ {
+ "epoch": 43.726003490401396,
+ "grad_norm": 0.04506697878241539,
+ "learning_rate": 3.0215566906848906e-05,
+ "loss": 3.479231595993042,
+ "step": 3148
+ },
+ {
+ "epoch": 43.73996509598604,
+ "grad_norm": 0.04495571181178093,
+ "learning_rate": 3.0110241898062726e-05,
+ "loss": 3.4603395462036133,
+ "step": 3149
+ },
+ {
+ "epoch": 43.753926701570684,
+ "grad_norm": 0.04532467573881149,
+ "learning_rate": 3.0005136751625748e-05,
+ "loss": 3.4643521308898926,
+ "step": 3150
+ },
+ {
+ "epoch": 43.767888307155324,
+ "grad_norm": 0.046149928122758865,
+ "learning_rate": 2.9900251552219178e-05,
+ "loss": 3.466956615447998,
+ "step": 3151
+ },
+ {
+ "epoch": 43.781849912739965,
+ "grad_norm": 0.045516159385442734,
+ "learning_rate": 2.9795586384347202e-05,
+ "loss": 3.4780664443969727,
+ "step": 3152
+ },
+ {
+ "epoch": 43.795811518324605,
+ "grad_norm": 0.04423840343952179,
+ "learning_rate": 2.9691141332336494e-05,
+ "loss": 3.4812045097351074,
+ "step": 3153
+ },
+ {
+ "epoch": 43.80977312390925,
+ "grad_norm": 0.047003038227558136,
+ "learning_rate": 2.9586916480336735e-05,
+ "loss": 3.4673991203308105,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82373472949389,
+ "grad_norm": 0.045799724757671356,
+ "learning_rate": 2.9482911912320006e-05,
+ "loss": 3.4602036476135254,
+ "step": 3155
+ },
+ {
+ "epoch": 43.83769633507853,
+ "grad_norm": 0.04607130587100983,
+ "learning_rate": 2.9379127712080756e-05,
+ "loss": 3.4666333198547363,
+ "step": 3156
+ },
+ {
+ "epoch": 43.85165794066317,
+ "grad_norm": 0.04715318605303764,
+ "learning_rate": 2.9275563963236107e-05,
+ "loss": 3.4574172496795654,
+ "step": 3157
+ },
+ {
+ "epoch": 43.86561954624782,
+ "grad_norm": 0.044526107609272,
+ "learning_rate": 2.9172220749225548e-05,
+ "loss": 3.4708940982818604,
+ "step": 3158
+ },
+ {
+ "epoch": 43.87958115183246,
+ "grad_norm": 0.04484846815466881,
+ "learning_rate": 2.906909815331085e-05,
+ "loss": 3.466261386871338,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8935427574171,
+ "grad_norm": 0.04613117873668671,
+ "learning_rate": 2.896619625857588e-05,
+ "loss": 3.4540066719055176,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90750436300174,
+ "grad_norm": 0.0452883206307888,
+ "learning_rate": 2.8863515147927003e-05,
+ "loss": 3.4572741985321045,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92146596858639,
+ "grad_norm": 0.045408450067043304,
+ "learning_rate": 2.8761054904092514e-05,
+ "loss": 3.43621563911438,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93542757417103,
+ "grad_norm": 0.045606113970279694,
+ "learning_rate": 2.8658815609622665e-05,
+ "loss": 3.4194536209106445,
+ "step": 3163
+ },
+ {
+ "epoch": 43.94938917975567,
+ "grad_norm": 0.04547363147139549,
+ "learning_rate": 2.855679734688993e-05,
+ "loss": 3.4742746353149414,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96335078534032,
+ "grad_norm": 0.046189647167921066,
+ "learning_rate": 2.8455000198088502e-05,
+ "loss": 3.459886074066162,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97731239092496,
+ "grad_norm": 0.045051224529743195,
+ "learning_rate": 2.8353424245234582e-05,
+ "loss": 3.460083484649658,
+ "step": 3166
+ },
+ {
+ "epoch": 43.9912739965096,
+ "grad_norm": 0.04479827359318733,
+ "learning_rate": 2.825206957016599e-05,
+ "loss": 3.4630260467529297,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.03426015004515648,
+ "learning_rate": 2.8150936254542376e-05,
+ "loss": 2.1829705238342285,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 0.6243340373039246,
+ "eval_runtime": 59.6021,
+ "eval_samples_per_second": 40.972,
+ "eval_steps_per_second": 0.654,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01396160558464,
+ "grad_norm": 0.039502255618572235,
+ "learning_rate": 2.8050024379845118e-05,
+ "loss": 3.453655481338501,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02792321116929,
+ "grad_norm": 0.039988402277231216,
+ "learning_rate": 2.7949334027376986e-05,
+ "loss": 3.478839874267578,
+ "step": 3170
+ },
+ {
+ "epoch": 44.04188481675393,
+ "grad_norm": 0.04035341739654541,
+ "learning_rate": 2.7848865278262427e-05,
+ "loss": 3.432953357696533,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05584642233857,
+ "grad_norm": 0.03990163281559944,
+ "learning_rate": 2.7748618213447268e-05,
+ "loss": 3.463597297668457,
+ "step": 3172
+ },
+ {
+ "epoch": 44.06980802792321,
+ "grad_norm": 0.04220886528491974,
+ "learning_rate": 2.764859291369885e-05,
+ "loss": 3.427955150604248,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083769633507856,
+ "grad_norm": 0.041814520955085754,
+ "learning_rate": 2.7548789459605603e-05,
+ "loss": 3.471323013305664,
+ "step": 3174
+ },
+ {
+ "epoch": 44.0977312390925,
+ "grad_norm": 0.040427837520837784,
+ "learning_rate": 2.744920793157743e-05,
+ "loss": 3.4614925384521484,
+ "step": 3175
+ },
+ {
+ "epoch": 44.11169284467714,
+ "grad_norm": 0.04101467877626419,
+ "learning_rate": 2.7349848409845462e-05,
+ "loss": 3.4553816318511963,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12565445026178,
+ "grad_norm": 0.04197322204709053,
+ "learning_rate": 2.725071097446174e-05,
+ "loss": 3.4764347076416016,
+ "step": 3177
+ },
+ {
+ "epoch": 44.139616055846425,
+ "grad_norm": 0.042121630162000656,
+ "learning_rate": 2.715179570529959e-05,
+ "loss": 3.4732165336608887,
+ "step": 3178
+ },
+ {
+ "epoch": 44.153577661431065,
+ "grad_norm": 0.041242565959692,
+ "learning_rate": 2.7053102682053134e-05,
+ "loss": 3.4677939414978027,
+ "step": 3179
+ },
+ {
+ "epoch": 44.167539267015705,
+ "grad_norm": 0.040816668421030045,
+ "learning_rate": 2.6954631984237766e-05,
+ "loss": 3.4444072246551514,
+ "step": 3180
+ },
+ {
+ "epoch": 44.181500872600346,
+ "grad_norm": 0.04216597601771355,
+ "learning_rate": 2.6856383691189347e-05,
+ "loss": 3.450758934020996,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19546247818499,
+ "grad_norm": 0.04147190600633621,
+ "learning_rate": 2.675835788206485e-05,
+ "loss": 3.44083571434021,
+ "step": 3182
+ },
+ {
+ "epoch": 44.20942408376963,
+ "grad_norm": 0.04059838876128197,
+ "learning_rate": 2.6660554635841912e-05,
+ "loss": 3.446469783782959,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223385689354274,
+ "grad_norm": 0.042010124772787094,
+ "learning_rate": 2.6562974031318788e-05,
+ "loss": 3.419384002685547,
+ "step": 3184
+ },
+ {
+ "epoch": 44.23734729493892,
+ "grad_norm": 0.042179301381111145,
+ "learning_rate": 2.6465616147114485e-05,
+ "loss": 3.4438977241516113,
+ "step": 3185
+ },
+ {
+ "epoch": 44.25130890052356,
+ "grad_norm": 0.04143347218632698,
+ "learning_rate": 2.6368481061668344e-05,
+ "loss": 3.478097438812256,
+ "step": 3186
+ },
+ {
+ "epoch": 44.2652705061082,
+ "grad_norm": 0.041317183524370193,
+ "learning_rate": 2.6271568853240557e-05,
+ "loss": 3.4608001708984375,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27923211169284,
+ "grad_norm": 0.041217777878046036,
+ "learning_rate": 2.6174879599911347e-05,
+ "loss": 3.4443745613098145,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29319371727749,
+ "grad_norm": 0.04093816131353378,
+ "learning_rate": 2.6078413379581627e-05,
+ "loss": 3.443121910095215,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30715532286213,
+ "grad_norm": 0.040228866040706635,
+ "learning_rate": 2.5982170269972397e-05,
+ "loss": 3.441892623901367,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32111692844677,
+ "grad_norm": 0.042651765048503876,
+ "learning_rate": 2.588615034862507e-05,
+ "loss": 3.4565064907073975,
+ "step": 3191
+ },
+ {
+ "epoch": 44.33507853403141,
+ "grad_norm": 0.042239658534526825,
+ "learning_rate": 2.5790353692901158e-05,
+ "loss": 3.451909065246582,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34904013961606,
+ "grad_norm": 0.04031132906675339,
+ "learning_rate": 2.5694780379982238e-05,
+ "loss": 3.4511477947235107,
+ "step": 3193
+ },
+ {
+ "epoch": 44.3630017452007,
+ "grad_norm": 0.04258335009217262,
+ "learning_rate": 2.5599430486870055e-05,
+ "loss": 3.484065294265747,
+ "step": 3194
+ },
+ {
+ "epoch": 44.37696335078534,
+ "grad_norm": 0.041619960218667984,
+ "learning_rate": 2.5504304090386225e-05,
+ "loss": 3.4708070755004883,
+ "step": 3195
+ },
+ {
+ "epoch": 44.390924956369986,
+ "grad_norm": 0.041788406670093536,
+ "learning_rate": 2.5409401267172474e-05,
+ "loss": 3.4092917442321777,
+ "step": 3196
+ },
+ {
+ "epoch": 44.404886561954626,
+ "grad_norm": 0.0416782908141613,
+ "learning_rate": 2.5314722093690176e-05,
+ "loss": 3.4138784408569336,
+ "step": 3197
+ },
+ {
+ "epoch": 44.41884816753927,
+ "grad_norm": 0.042210739105939865,
+ "learning_rate": 2.522026664622071e-05,
+ "loss": 3.453545093536377,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43280977312391,
+ "grad_norm": 0.04266810044646263,
+ "learning_rate": 2.5126035000865124e-05,
+ "loss": 3.424774169921875,
+ "step": 3199
+ },
+ {
+ "epoch": 44.446771378708554,
+ "grad_norm": 0.04081276059150696,
+ "learning_rate": 2.503202723354413e-05,
+ "loss": 3.4538846015930176,
+ "step": 3200
+ },
+ {
+ "epoch": 44.460732984293195,
+ "grad_norm": 0.0413636788725853,
+ "learning_rate": 2.4938243419998047e-05,
+ "loss": 3.4658732414245605,
+ "step": 3201
+ },
+ {
+ "epoch": 44.474694589877835,
+ "grad_norm": 0.04199403524398804,
+ "learning_rate": 2.484468363578689e-05,
+ "loss": 3.4627909660339355,
+ "step": 3202
+ },
+ {
+ "epoch": 44.488656195462475,
+ "grad_norm": 0.04227396100759506,
+ "learning_rate": 2.475134795629004e-05,
+ "loss": 3.4517412185668945,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50261780104712,
+ "grad_norm": 0.04252436012029648,
+ "learning_rate": 2.4658236456706357e-05,
+ "loss": 3.4830689430236816,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51657940663176,
+ "grad_norm": 0.04169167950749397,
+ "learning_rate": 2.456534921205403e-05,
+ "loss": 3.450204372406006,
+ "step": 3205
+ },
+ {
+ "epoch": 44.5305410122164,
+ "grad_norm": 0.041747767478227615,
+ "learning_rate": 2.4472686297170822e-05,
+ "loss": 3.472952365875244,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544502617801044,
+ "grad_norm": 0.0426458977162838,
+ "learning_rate": 2.4380247786713403e-05,
+ "loss": 3.4559519290924072,
+ "step": 3207
+ },
+ {
+ "epoch": 44.55846422338569,
+ "grad_norm": 0.04284683242440224,
+ "learning_rate": 2.428803375515788e-05,
+ "loss": 3.4351186752319336,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57242582897033,
+ "grad_norm": 0.041979581117630005,
+ "learning_rate": 2.4196044276799397e-05,
+ "loss": 3.4520907402038574,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58638743455497,
+ "grad_norm": 0.042444370687007904,
+ "learning_rate": 2.410427942575228e-05,
+ "loss": 3.4589779376983643,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60034904013962,
+ "grad_norm": 0.04146238788962364,
+ "learning_rate": 2.4012739275949746e-05,
+ "loss": 3.468356132507324,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61431064572426,
+ "grad_norm": 0.043098706752061844,
+ "learning_rate": 2.3921423901144008e-05,
+ "loss": 3.422999620437622,
+ "step": 3212
+ },
+ {
+ "epoch": 44.6282722513089,
+ "grad_norm": 0.042554691433906555,
+ "learning_rate": 2.3830333374906342e-05,
+ "loss": 3.441878080368042,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64223385689354,
+ "grad_norm": 0.04226307198405266,
+ "learning_rate": 2.3739467770626637e-05,
+ "loss": 3.4962363243103027,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65619546247819,
+ "grad_norm": 0.041877422481775284,
+ "learning_rate": 2.3648827161513686e-05,
+ "loss": 3.471935510635376,
+ "step": 3215
+ },
+ {
+ "epoch": 44.67015706806283,
+ "grad_norm": 0.04196307063102722,
+ "learning_rate": 2.3558411620595038e-05,
+ "loss": 3.468240737915039,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68411867364747,
+ "grad_norm": 0.04305512458086014,
+ "learning_rate": 2.3468221220716844e-05,
+ "loss": 3.457876205444336,
+ "step": 3217
+ },
+ {
+ "epoch": 44.69808027923211,
+ "grad_norm": 0.04215867817401886,
+ "learning_rate": 2.33782560345438e-05,
+ "loss": 3.4575209617614746,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712041884816756,
+ "grad_norm": 0.041869666427373886,
+ "learning_rate": 2.328851613455932e-05,
+ "loss": 3.442275285720825,
+ "step": 3219
+ },
+ {
+ "epoch": 44.726003490401396,
+ "grad_norm": 0.04279697686433792,
+ "learning_rate": 2.3199001593065204e-05,
+ "loss": 3.4583919048309326,
+ "step": 3220
+ },
+ {
+ "epoch": 44.73996509598604,
+ "grad_norm": 0.04206348955631256,
+ "learning_rate": 2.3109712482181684e-05,
+ "loss": 3.4497156143188477,
+ "step": 3221
+ },
+ {
+ "epoch": 44.753926701570684,
+ "grad_norm": 0.0420316718518734,
+ "learning_rate": 2.3020648873847414e-05,
+ "loss": 3.4462804794311523,
+ "step": 3222
+ },
+ {
+ "epoch": 44.767888307155324,
+ "grad_norm": 0.0421106293797493,
+ "learning_rate": 2.2931810839819235e-05,
+ "loss": 3.461982250213623,
+ "step": 3223
+ },
+ {
+ "epoch": 44.781849912739965,
+ "grad_norm": 0.04225604981184006,
+ "learning_rate": 2.284319845167253e-05,
+ "loss": 3.440438747406006,
+ "step": 3224
+ },
+ {
+ "epoch": 44.795811518324605,
+ "grad_norm": 0.04363993555307388,
+ "learning_rate": 2.2754811780800532e-05,
+ "loss": 3.4595792293548584,
+ "step": 3225
+ },
+ {
+ "epoch": 44.80977312390925,
+ "grad_norm": 0.04277531057596207,
+ "learning_rate": 2.2666650898414827e-05,
+ "loss": 3.444552421569824,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82373472949389,
+ "grad_norm": 0.042936112731695175,
+ "learning_rate": 2.2578715875545095e-05,
+ "loss": 3.4602608680725098,
+ "step": 3227
+ },
+ {
+ "epoch": 44.83769633507853,
+ "grad_norm": 0.041564665734767914,
+ "learning_rate": 2.2491006783038976e-05,
+ "loss": 3.4477994441986084,
+ "step": 3228
+ },
+ {
+ "epoch": 44.85165794066317,
+ "grad_norm": 0.043317802250385284,
+ "learning_rate": 2.2403523691562092e-05,
+ "loss": 3.481790542602539,
+ "step": 3229
+ },
+ {
+ "epoch": 44.86561954624782,
+ "grad_norm": 0.042523398995399475,
+ "learning_rate": 2.2316266671598003e-05,
+ "loss": 3.432340621948242,
+ "step": 3230
+ },
+ {
+ "epoch": 44.87958115183246,
+ "grad_norm": 0.04249822348356247,
+ "learning_rate": 2.222923579344805e-05,
+ "loss": 3.4209306240081787,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8935427574171,
+ "grad_norm": 0.04233477637171745,
+ "learning_rate": 2.2142431127231502e-05,
+ "loss": 3.454221248626709,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90750436300174,
+ "grad_norm": 0.04137618467211723,
+ "learning_rate": 2.2055852742885322e-05,
+ "loss": 3.463860511779785,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92146596858639,
+ "grad_norm": 0.04398854076862335,
+ "learning_rate": 2.196950071016409e-05,
+ "loss": 3.4566519260406494,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93542757417103,
+ "grad_norm": 0.04244855418801308,
+ "learning_rate": 2.1883375098640134e-05,
+ "loss": 3.4292609691619873,
+ "step": 3235
+ },
+ {
+ "epoch": 44.94938917975567,
+ "grad_norm": 0.04317071661353111,
+ "learning_rate": 2.1797475977703352e-05,
+ "loss": 3.4438283443450928,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96335078534032,
+ "grad_norm": 0.04333546757698059,
+ "learning_rate": 2.1711803416560964e-05,
+ "loss": 3.420079231262207,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97731239092496,
+ "grad_norm": 0.04234454035758972,
+ "learning_rate": 2.16263574842379e-05,
+ "loss": 3.4233388900756836,
+ "step": 3238
+ },
+ {
+ "epoch": 44.9912739965096,
+ "grad_norm": 0.04304066300392151,
+ "learning_rate": 2.15411382495764e-05,
+ "loss": 3.4663498401641846,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.03293231502175331,
+ "learning_rate": 2.1456145781236057e-05,
+ "loss": 2.176724433898926,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01396160558464,
+ "grad_norm": 0.037529222667217255,
+ "learning_rate": 2.1371380147693752e-05,
+ "loss": 3.431626796722412,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02792321116929,
+ "grad_norm": 0.0367698110640049,
+ "learning_rate": 2.1286841417243662e-05,
+ "loss": 3.419940948486328,
+ "step": 3242
+ },
+ {
+ "epoch": 45.04188481675393,
+ "grad_norm": 0.03786632791161537,
+ "learning_rate": 2.1202529657997137e-05,
+ "loss": 3.452460289001465,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05584642233857,
+ "grad_norm": 0.03760818764567375,
+ "learning_rate": 2.1118444937882538e-05,
+ "loss": 3.4500484466552734,
+ "step": 3244
+ },
+ {
+ "epoch": 45.06980802792321,
+ "grad_norm": 0.03770846873521805,
+ "learning_rate": 2.1034587324645456e-05,
+ "loss": 3.4246275424957275,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083769633507856,
+ "grad_norm": 0.038266588002443314,
+ "learning_rate": 2.095095688584846e-05,
+ "loss": 3.418670654296875,
+ "step": 3246
+ },
+ {
+ "epoch": 45.0977312390925,
+ "grad_norm": 0.03750491142272949,
+ "learning_rate": 2.086755368887107e-05,
+ "loss": 3.427488327026367,
+ "step": 3247
+ },
+ {
+ "epoch": 45.11169284467714,
+ "grad_norm": 0.03955574706196785,
+ "learning_rate": 2.0784377800909802e-05,
+ "loss": 3.4298300743103027,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12565445026178,
+ "grad_norm": 0.0371951162815094,
+ "learning_rate": 2.0701429288977847e-05,
+ "loss": 3.427586078643799,
+ "step": 3249
+ },
+ {
+ "epoch": 45.139616055846425,
+ "grad_norm": 0.038450200110673904,
+ "learning_rate": 2.0618708219905455e-05,
+ "loss": 3.4494733810424805,
+ "step": 3250
+ },
+ {
+ "epoch": 45.153577661431065,
+ "grad_norm": 0.039204902946949005,
+ "learning_rate": 2.0536214660339398e-05,
+ "loss": 3.4210152626037598,
+ "step": 3251
+ },
+ {
+ "epoch": 45.167539267015705,
+ "grad_norm": 0.039056919515132904,
+ "learning_rate": 2.0453948676743267e-05,
+ "loss": 3.436870813369751,
+ "step": 3252
+ },
+ {
+ "epoch": 45.181500872600346,
+ "grad_norm": 0.039004355669021606,
+ "learning_rate": 2.037191033539731e-05,
+ "loss": 3.445186138153076,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19546247818499,
+ "grad_norm": 0.039315756410360336,
+ "learning_rate": 2.029009970239831e-05,
+ "loss": 3.456547737121582,
+ "step": 3254
+ },
+ {
+ "epoch": 45.20942408376963,
+ "grad_norm": 0.03935930132865906,
+ "learning_rate": 2.0208516843659678e-05,
+ "loss": 3.459127187728882,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223385689354274,
+ "grad_norm": 0.03802679851651192,
+ "learning_rate": 2.012716182491114e-05,
+ "loss": 3.4773435592651367,
+ "step": 3256
+ },
+ {
+ "epoch": 45.23734729493892,
+ "grad_norm": 0.0384269542992115,
+ "learning_rate": 2.0046034711699048e-05,
+ "loss": 3.433504581451416,
+ "step": 3257
+ },
+ {
+ "epoch": 45.25130890052356,
+ "grad_norm": 0.03953392058610916,
+ "learning_rate": 1.9965135569386032e-05,
+ "loss": 3.4244377613067627,
+ "step": 3258
+ },
+ {
+ "epoch": 45.2652705061082,
+ "grad_norm": 0.03861505910754204,
+ "learning_rate": 1.9884464463151077e-05,
+ "loss": 3.4536099433898926,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27923211169284,
+ "grad_norm": 0.039188891649246216,
+ "learning_rate": 1.980402145798933e-05,
+ "loss": 3.4352147579193115,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29319371727749,
+ "grad_norm": 0.03887999802827835,
+ "learning_rate": 1.9723806618712484e-05,
+ "loss": 3.4650840759277344,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30715532286213,
+ "grad_norm": 0.03883611038327217,
+ "learning_rate": 1.9643820009947983e-05,
+ "loss": 3.4430532455444336,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32111692844677,
+ "grad_norm": 0.03958422318100929,
+ "learning_rate": 1.9564061696139687e-05,
+ "loss": 3.4469406604766846,
+ "step": 3263
+ },
+ {
+ "epoch": 45.33507853403141,
+ "grad_norm": 0.03871805965900421,
+ "learning_rate": 1.948453174154744e-05,
+ "loss": 3.459047794342041,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34904013961606,
+ "grad_norm": 0.03840525820851326,
+ "learning_rate": 1.9405230210247042e-05,
+ "loss": 3.4470202922821045,
+ "step": 3265
+ },
+ {
+ "epoch": 45.3630017452007,
+ "grad_norm": 0.03913816064596176,
+ "learning_rate": 1.9326157166130365e-05,
+ "loss": 3.4735798835754395,
+ "step": 3266
+ },
+ {
+ "epoch": 45.37696335078534,
+ "grad_norm": 0.038954757153987885,
+ "learning_rate": 1.9247312672905017e-05,
+ "loss": 3.4590110778808594,
+ "step": 3267
+ },
+ {
+ "epoch": 45.390924956369986,
+ "grad_norm": 0.038997333496809006,
+ "learning_rate": 1.9168696794094758e-05,
+ "loss": 3.4293274879455566,
+ "step": 3268
+ },
+ {
+ "epoch": 45.404886561954626,
+ "grad_norm": 0.039304427802562714,
+ "learning_rate": 1.9090309593038812e-05,
+ "loss": 3.4162559509277344,
+ "step": 3269
+ },
+ {
+ "epoch": 45.41884816753927,
+ "grad_norm": 0.03842204436659813,
+ "learning_rate": 1.9012151132892425e-05,
+ "loss": 3.4435534477233887,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43280977312391,
+ "grad_norm": 0.03963358700275421,
+ "learning_rate": 1.893422147662644e-05,
+ "loss": 3.434021472930908,
+ "step": 3271
+ },
+ {
+ "epoch": 45.446771378708554,
+ "grad_norm": 0.03829001635313034,
+ "learning_rate": 1.8856520687027325e-05,
+ "loss": 3.4448766708374023,
+ "step": 3272
+ },
+ {
+ "epoch": 45.460732984293195,
+ "grad_norm": 0.038870956748723984,
+ "learning_rate": 1.8779048826697318e-05,
+ "loss": 3.465782880783081,
+ "step": 3273
+ },
+ {
+ "epoch": 45.474694589877835,
+ "grad_norm": 0.039185624569654465,
+ "learning_rate": 1.8701805958054e-05,
+ "loss": 3.439093589782715,
+ "step": 3274
+ },
+ {
+ "epoch": 45.488656195462475,
+ "grad_norm": 0.04019121453166008,
+ "learning_rate": 1.8624792143330544e-05,
+ "loss": 3.433955192565918,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50261780104712,
+ "grad_norm": 0.040416400879621506,
+ "learning_rate": 1.8548007444575656e-05,
+ "loss": 3.436868190765381,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51657940663176,
+ "grad_norm": 0.03936008736491203,
+ "learning_rate": 1.8471451923653322e-05,
+ "loss": 3.460071325302124,
+ "step": 3277
+ },
+ {
+ "epoch": 45.5305410122164,
+ "grad_norm": 0.039932943880558014,
+ "learning_rate": 1.8395125642242987e-05,
+ "loss": 3.441723346710205,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544502617801044,
+ "grad_norm": 0.03912197798490524,
+ "learning_rate": 1.8319028661839343e-05,
+ "loss": 3.4285101890563965,
+ "step": 3279
+ },
+ {
+ "epoch": 45.55846422338569,
+ "grad_norm": 0.03936731070280075,
+ "learning_rate": 1.8243161043752384e-05,
+ "loss": 3.45601487159729,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57242582897033,
+ "grad_norm": 0.039453133940696716,
+ "learning_rate": 1.8167522849107228e-05,
+ "loss": 3.437004566192627,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58638743455497,
+ "grad_norm": 0.03929216042160988,
+ "learning_rate": 1.809211413884418e-05,
+ "loss": 3.4422807693481445,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60034904013962,
+ "grad_norm": 0.039395686239004135,
+ "learning_rate": 1.8016934973718762e-05,
+ "loss": 3.4438841342926025,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61431064572426,
+ "grad_norm": 0.04027940705418587,
+ "learning_rate": 1.7941985414301403e-05,
+ "loss": 3.4653074741363525,
+ "step": 3284
+ },
+ {
+ "epoch": 45.6282722513089,
+ "grad_norm": 0.03892998769879341,
+ "learning_rate": 1.786726552097766e-05,
+ "loss": 3.435762882232666,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64223385689354,
+ "grad_norm": 0.0392686203122139,
+ "learning_rate": 1.7792775353947908e-05,
+ "loss": 3.4253759384155273,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65619546247819,
+ "grad_norm": 0.040689513087272644,
+ "learning_rate": 1.7718514973227634e-05,
+ "loss": 3.438173294067383,
+ "step": 3287
+ },
+ {
+ "epoch": 45.67015706806283,
+ "grad_norm": 0.039168741554021835,
+ "learning_rate": 1.7644484438647057e-05,
+ "loss": 3.4670262336730957,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68411867364747,
+ "grad_norm": 0.04095490649342537,
+ "learning_rate": 1.75706838098512e-05,
+ "loss": 3.4413394927978516,
+ "step": 3289
+ },
+ {
+ "epoch": 45.69808027923211,
+ "grad_norm": 0.0398528017103672,
+ "learning_rate": 1.749711314629992e-05,
+ "loss": 3.4605154991149902,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712041884816756,
+ "grad_norm": 0.04048970341682434,
+ "learning_rate": 1.742377250726779e-05,
+ "loss": 3.464498996734619,
+ "step": 3291
+ },
+ {
+ "epoch": 45.726003490401396,
+ "grad_norm": 0.04015368968248367,
+ "learning_rate": 1.7350661951844065e-05,
+ "loss": 3.4415974617004395,
+ "step": 3292
+ },
+ {
+ "epoch": 45.73996509598604,
+ "grad_norm": 0.038970813155174255,
+ "learning_rate": 1.727778153893248e-05,
+ "loss": 3.4493942260742188,
+ "step": 3293
+ },
+ {
+ "epoch": 45.753926701570684,
+ "grad_norm": 0.039729416370391846,
+ "learning_rate": 1.720513132725162e-05,
+ "loss": 3.449537754058838,
+ "step": 3294
+ },
+ {
+ "epoch": 45.767888307155324,
+ "grad_norm": 0.039825472980737686,
+ "learning_rate": 1.713271137533433e-05,
+ "loss": 3.4090218544006348,
+ "step": 3295
+ },
+ {
+ "epoch": 45.781849912739965,
+ "grad_norm": 0.039860207587480545,
+ "learning_rate": 1.7060521741528127e-05,
+ "loss": 3.4335269927978516,
+ "step": 3296
+ },
+ {
+ "epoch": 45.795811518324605,
+ "grad_norm": 0.0395670011639595,
+ "learning_rate": 1.698856248399482e-05,
+ "loss": 3.4387285709381104,
+ "step": 3297
+ },
+ {
+ "epoch": 45.80977312390925,
+ "grad_norm": 0.03963136300444603,
+ "learning_rate": 1.6916833660710785e-05,
+ "loss": 3.4522385597229004,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82373472949389,
+ "grad_norm": 0.03982813283801079,
+ "learning_rate": 1.6845335329466606e-05,
+ "loss": 3.4439194202423096,
+ "step": 3299
+ },
+ {
+ "epoch": 45.83769633507853,
+ "grad_norm": 0.03938654810190201,
+ "learning_rate": 1.6774067547867113e-05,
+ "loss": 3.425381898880005,
+ "step": 3300
+ },
+ {
+ "epoch": 45.85165794066317,
+ "grad_norm": 0.03889504447579384,
+ "learning_rate": 1.6703030373331585e-05,
+ "loss": 3.4470620155334473,
+ "step": 3301
+ },
+ {
+ "epoch": 45.86561954624782,
+ "grad_norm": 0.03958441689610481,
+ "learning_rate": 1.6632223863093328e-05,
+ "loss": 3.4461910724639893,
+ "step": 3302
+ },
+ {
+ "epoch": 45.87958115183246,
+ "grad_norm": 0.03922349214553833,
+ "learning_rate": 1.656164807419988e-05,
+ "loss": 3.4262583255767822,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8935427574171,
+ "grad_norm": 0.04015881195664406,
+ "learning_rate": 1.6491303063512866e-05,
+ "loss": 3.4820480346679688,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90750436300174,
+ "grad_norm": 0.03984055295586586,
+ "learning_rate": 1.6421188887708013e-05,
+ "loss": 3.444671154022217,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92146596858639,
+ "grad_norm": 0.04056533798575401,
+ "learning_rate": 1.6351305603275067e-05,
+ "loss": 3.4515886306762695,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93542757417103,
+ "grad_norm": 0.03927317634224892,
+ "learning_rate": 1.6281653266517658e-05,
+ "loss": 3.4190826416015625,
+ "step": 3307
+ },
+ {
+ "epoch": 45.94938917975567,
+ "grad_norm": 0.039910025894641876,
+ "learning_rate": 1.6212231933553485e-05,
+ "loss": 3.4318108558654785,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96335078534032,
+ "grad_norm": 0.04135388880968094,
+ "learning_rate": 1.6143041660313997e-05,
+ "loss": 3.479313373565674,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97731239092496,
+ "grad_norm": 0.039951108396053314,
+ "learning_rate": 1.6074082502544668e-05,
+ "loss": 3.4217677116394043,
+ "step": 3310
+ },
+ {
+ "epoch": 45.9912739965096,
+ "grad_norm": 0.0405874103307724,
+ "learning_rate": 1.6005354515804503e-05,
+ "loss": 3.4322991371154785,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.031679119914770126,
+ "learning_rate": 1.59368577554664e-05,
+ "loss": 2.1517839431762695,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 0.6261149048805237,
+ "eval_runtime": 60.1318,
+ "eval_samples_per_second": 40.611,
+ "eval_steps_per_second": 0.649,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01396160558464,
+ "grad_norm": 0.035890862345695496,
+ "learning_rate": 1.5868592276717137e-05,
+ "loss": 3.4109573364257812,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02792321116929,
+ "grad_norm": 0.03540349006652832,
+ "learning_rate": 1.5800558134556806e-05,
+ "loss": 3.4465813636779785,
+ "step": 3314
+ },
+ {
+ "epoch": 46.04188481675393,
+ "grad_norm": 0.035230789333581924,
+ "learning_rate": 1.5732755383799315e-05,
+ "loss": 3.4250102043151855,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05584642233857,
+ "grad_norm": 0.03603345900774002,
+ "learning_rate": 1.5665184079072162e-05,
+ "loss": 3.4378719329833984,
+ "step": 3316
+ },
+ {
+ "epoch": 46.06980802792321,
+ "grad_norm": 0.0353219211101532,
+ "learning_rate": 1.5597844274816352e-05,
+ "loss": 3.447131872177124,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083769633507856,
+ "grad_norm": 0.03627607971429825,
+ "learning_rate": 1.553073602528626e-05,
+ "loss": 3.4486751556396484,
+ "step": 3318
+ },
+ {
+ "epoch": 46.0977312390925,
+ "grad_norm": 0.03613058105111122,
+ "learning_rate": 1.546385938454982e-05,
+ "loss": 3.4362406730651855,
+ "step": 3319
+ },
+ {
+ "epoch": 46.11169284467714,
+ "grad_norm": 0.03603339567780495,
+ "learning_rate": 1.5397214406488374e-05,
+ "loss": 3.4231607913970947,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12565445026178,
+ "grad_norm": 0.03566336631774902,
+ "learning_rate": 1.533080114479657e-05,
+ "loss": 3.39627742767334,
+ "step": 3321
+ },
+ {
+ "epoch": 46.139616055846425,
+ "grad_norm": 0.0368712842464447,
+ "learning_rate": 1.5264619652982352e-05,
+ "loss": 3.464022159576416,
+ "step": 3322
+ },
+ {
+ "epoch": 46.153577661431065,
+ "grad_norm": 0.0362398736178875,
+ "learning_rate": 1.5198669984366876e-05,
+ "loss": 3.4518275260925293,
+ "step": 3323
+ },
+ {
+ "epoch": 46.167539267015705,
+ "grad_norm": 0.036385323852300644,
+ "learning_rate": 1.5132952192084793e-05,
+ "loss": 3.425281524658203,
+ "step": 3324
+ },
+ {
+ "epoch": 46.181500872600346,
+ "grad_norm": 0.03549065440893173,
+ "learning_rate": 1.5067466329083533e-05,
+ "loss": 3.4173173904418945,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19546247818499,
+ "grad_norm": 0.03571503981947899,
+ "learning_rate": 1.5002212448123997e-05,
+ "loss": 3.424938201904297,
+ "step": 3326
+ },
+ {
+ "epoch": 46.20942408376963,
+ "grad_norm": 0.0356319397687912,
+ "learning_rate": 1.4937190601780052e-05,
+ "loss": 3.4231014251708984,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223385689354274,
+ "grad_norm": 0.036459170281887054,
+ "learning_rate": 1.4872400842438576e-05,
+ "loss": 3.442168951034546,
+ "step": 3328
+ },
+ {
+ "epoch": 46.23734729493892,
+ "grad_norm": 0.036033108830451965,
+ "learning_rate": 1.4807843222299557e-05,
+ "loss": 3.452871084213257,
+ "step": 3329
+ },
+ {
+ "epoch": 46.25130890052356,
+ "grad_norm": 0.03554454445838928,
+ "learning_rate": 1.4743517793375814e-05,
+ "loss": 3.4228341579437256,
+ "step": 3330
+ },
+ {
+ "epoch": 46.2652705061082,
+ "grad_norm": 0.036166585981845856,
+ "learning_rate": 1.4679424607493315e-05,
+ "loss": 3.424604892730713,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27923211169284,
+ "grad_norm": 0.03760233521461487,
+ "learning_rate": 1.4615563716290698e-05,
+ "loss": 3.424304723739624,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29319371727749,
+ "grad_norm": 0.03685798496007919,
+ "learning_rate": 1.455193517121951e-05,
+ "loss": 3.4280381202697754,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30715532286213,
+ "grad_norm": 0.03640700876712799,
+ "learning_rate": 1.448853902354421e-05,
+ "loss": 3.4119839668273926,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32111692844677,
+ "grad_norm": 0.03598141670227051,
+ "learning_rate": 1.4425375324341827e-05,
+ "loss": 3.4472408294677734,
+ "step": 3335
+ },
+ {
+ "epoch": 46.33507853403141,
+ "grad_norm": 0.03695828467607498,
+ "learning_rate": 1.4362444124502309e-05,
+ "loss": 3.4203333854675293,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34904013961606,
+ "grad_norm": 0.036167584359645844,
+ "learning_rate": 1.429974547472814e-05,
+ "loss": 3.437302589416504,
+ "step": 3337
+ },
+ {
+ "epoch": 46.3630017452007,
+ "grad_norm": 0.03652525320649147,
+ "learning_rate": 1.4237279425534442e-05,
+ "loss": 3.451777935028076,
+ "step": 3338
+ },
+ {
+ "epoch": 46.37696335078534,
+ "grad_norm": 0.03641019016504288,
+ "learning_rate": 1.4175046027249065e-05,
+ "loss": 3.44952392578125,
+ "step": 3339
+ },
+ {
+ "epoch": 46.390924956369986,
+ "grad_norm": 0.036645423620939255,
+ "learning_rate": 1.4113045330012324e-05,
+ "loss": 3.454132080078125,
+ "step": 3340
+ },
+ {
+ "epoch": 46.404886561954626,
+ "grad_norm": 0.03628278151154518,
+ "learning_rate": 1.4051277383776953e-05,
+ "loss": 3.4473910331726074,
+ "step": 3341
+ },
+ {
+ "epoch": 46.41884816753927,
+ "grad_norm": 0.03580992668867111,
+ "learning_rate": 1.398974223830839e-05,
+ "loss": 3.4273030757904053,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43280977312391,
+ "grad_norm": 0.0360463410615921,
+ "learning_rate": 1.3928439943184394e-05,
+ "loss": 3.4377493858337402,
+ "step": 3343
+ },
+ {
+ "epoch": 46.446771378708554,
+ "grad_norm": 0.036654215306043625,
+ "learning_rate": 1.3867370547795065e-05,
+ "loss": 3.397080898284912,
+ "step": 3344
+ },
+ {
+ "epoch": 46.460732984293195,
+ "grad_norm": 0.037537649273872375,
+ "learning_rate": 1.3806534101342891e-05,
+ "loss": 3.429286479949951,
+ "step": 3345
+ },
+ {
+ "epoch": 46.474694589877835,
+ "grad_norm": 0.03625643998384476,
+ "learning_rate": 1.3745930652842728e-05,
+ "loss": 3.436239004135132,
+ "step": 3346
+ },
+ {
+ "epoch": 46.488656195462475,
+ "grad_norm": 0.03664219751954079,
+ "learning_rate": 1.368556025112176e-05,
+ "loss": 3.443955898284912,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50261780104712,
+ "grad_norm": 0.036535412073135376,
+ "learning_rate": 1.362542294481918e-05,
+ "loss": 3.427825927734375,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51657940663176,
+ "grad_norm": 0.037270862609148026,
+ "learning_rate": 1.356551878238656e-05,
+ "loss": 3.43320894241333,
+ "step": 3349
+ },
+ {
+ "epoch": 46.5305410122164,
+ "grad_norm": 0.03715480491518974,
+ "learning_rate": 1.3505847812087724e-05,
+ "loss": 3.430527687072754,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544502617801044,
+ "grad_norm": 0.03792518377304077,
+ "learning_rate": 1.3446410081998384e-05,
+ "loss": 3.4130091667175293,
+ "step": 3351
+ },
+ {
+ "epoch": 46.55846422338569,
+ "grad_norm": 0.03608744218945503,
+ "learning_rate": 1.3387205640006494e-05,
+ "loss": 3.447970390319824,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57242582897033,
+ "grad_norm": 0.03714914619922638,
+ "learning_rate": 1.332823453381196e-05,
+ "loss": 3.430121660232544,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58638743455497,
+ "grad_norm": 0.03783484175801277,
+ "learning_rate": 1.3269496810926807e-05,
+ "loss": 3.4351656436920166,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60034904013962,
+ "grad_norm": 0.037746530026197433,
+ "learning_rate": 1.3210992518674848e-05,
+ "loss": 3.4475512504577637,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61431064572426,
+ "grad_norm": 0.035554710775613785,
+ "learning_rate": 1.315272170419201e-05,
+ "loss": 3.4232020378112793,
+ "step": 3356
+ },
+ {
+ "epoch": 46.6282722513089,
+ "grad_norm": 0.03769008442759514,
+ "learning_rate": 1.3094684414426073e-05,
+ "loss": 3.4355225563049316,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64223385689354,
+ "grad_norm": 0.036033954471349716,
+ "learning_rate": 1.3036880696136532e-05,
+ "loss": 3.4258928298950195,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65619546247819,
+ "grad_norm": 0.037698592990636826,
+ "learning_rate": 1.2979310595894882e-05,
+ "loss": 3.445732831954956,
+ "step": 3359
+ },
+ {
+ "epoch": 46.67015706806283,
+ "grad_norm": 0.03696145489811897,
+ "learning_rate": 1.2921974160084251e-05,
+ "loss": 3.424511194229126,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68411867364747,
+ "grad_norm": 0.03692244738340378,
+ "learning_rate": 1.2864871434899672e-05,
+ "loss": 3.42803955078125,
+ "step": 3361
+ },
+ {
+ "epoch": 46.69808027923211,
+ "grad_norm": 0.03666501119732857,
+ "learning_rate": 1.2808002466347654e-05,
+ "loss": 3.4394397735595703,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712041884816756,
+ "grad_norm": 0.03683903440833092,
+ "learning_rate": 1.2751367300246535e-05,
+ "loss": 3.458301067352295,
+ "step": 3363
+ },
+ {
+ "epoch": 46.726003490401396,
+ "grad_norm": 0.03796245530247688,
+ "learning_rate": 1.2694965982226301e-05,
+ "loss": 3.4264326095581055,
+ "step": 3364
+ },
+ {
+ "epoch": 46.73996509598604,
+ "grad_norm": 0.036369986832141876,
+ "learning_rate": 1.2638798557728402e-05,
+ "loss": 3.458315849304199,
+ "step": 3365
+ },
+ {
+ "epoch": 46.753926701570684,
+ "grad_norm": 0.037210140377283096,
+ "learning_rate": 1.2582865072006e-05,
+ "loss": 3.440868616104126,
+ "step": 3366
+ },
+ {
+ "epoch": 46.767888307155324,
+ "grad_norm": 0.03766356781125069,
+ "learning_rate": 1.2527165570123586e-05,
+ "loss": 3.425314426422119,
+ "step": 3367
+ },
+ {
+ "epoch": 46.781849912739965,
+ "grad_norm": 0.03648895397782326,
+ "learning_rate": 1.2471700096957376e-05,
+ "loss": 3.4281005859375,
+ "step": 3368
+ },
+ {
+ "epoch": 46.795811518324605,
+ "grad_norm": 0.03725716471672058,
+ "learning_rate": 1.2416468697194767e-05,
+ "loss": 3.419691562652588,
+ "step": 3369
+ },
+ {
+ "epoch": 46.80977312390925,
+ "grad_norm": 0.037754807621240616,
+ "learning_rate": 1.2361471415334795e-05,
+ "loss": 3.431955099105835,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82373472949389,
+ "grad_norm": 0.03793100640177727,
+ "learning_rate": 1.2306708295687706e-05,
+ "loss": 3.435089349746704,
+ "step": 3371
+ },
+ {
+ "epoch": 46.83769633507853,
+ "grad_norm": 0.03681453317403793,
+ "learning_rate": 1.2252179382375211e-05,
+ "loss": 3.431636333465576,
+ "step": 3372
+ },
+ {
+ "epoch": 46.85165794066317,
+ "grad_norm": 0.03768916055560112,
+ "learning_rate": 1.2197884719330281e-05,
+ "loss": 3.41259503364563,
+ "step": 3373
+ },
+ {
+ "epoch": 46.86561954624782,
+ "grad_norm": 0.03774314001202583,
+ "learning_rate": 1.2143824350297104e-05,
+ "loss": 3.4260759353637695,
+ "step": 3374
+ },
+ {
+ "epoch": 46.87958115183246,
+ "grad_norm": 0.036926914006471634,
+ "learning_rate": 1.2089998318831137e-05,
+ "loss": 3.446438789367676,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8935427574171,
+ "grad_norm": 0.03717603161931038,
+ "learning_rate": 1.2036406668299062e-05,
+ "loss": 3.435614585876465,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90750436300174,
+ "grad_norm": 0.03680265322327614,
+ "learning_rate": 1.198304944187868e-05,
+ "loss": 3.430088758468628,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92146596858639,
+ "grad_norm": 0.03788964822888374,
+ "learning_rate": 1.1929926682558995e-05,
+ "loss": 3.45151424407959,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93542757417103,
+ "grad_norm": 0.03683070093393326,
+ "learning_rate": 1.1877038433139998e-05,
+ "loss": 3.437394380569458,
+ "step": 3379
+ },
+ {
+ "epoch": 46.94938917975567,
+ "grad_norm": 0.036409493535757065,
+ "learning_rate": 1.1824384736232879e-05,
+ "loss": 3.454728603363037,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96335078534032,
+ "grad_norm": 0.03721268102526665,
+ "learning_rate": 1.1771965634259653e-05,
+ "loss": 3.4802770614624023,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97731239092496,
+ "grad_norm": 0.037396397441625595,
+ "learning_rate": 1.1719781169453536e-05,
+ "loss": 3.440497875213623,
+ "step": 3382
+ },
+ {
+ "epoch": 46.9912739965096,
+ "grad_norm": 0.03714139387011528,
+ "learning_rate": 1.1667831383858562e-05,
+ "loss": 3.4558377265930176,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.028893817216157913,
+ "learning_rate": 1.1616116319329808e-05,
+ "loss": 2.155693531036377,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 0.6268184781074524,
+ "eval_runtime": 60.4953,
+ "eval_samples_per_second": 40.367,
+ "eval_steps_per_second": 0.645,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01396160558464,
+ "grad_norm": 0.03274908661842346,
+ "learning_rate": 1.1564636017533036e-05,
+ "loss": 3.4290950298309326,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02792321116929,
+ "grad_norm": 0.033523742109537125,
+ "learning_rate": 1.151339051994508e-05,
+ "loss": 3.422708034515381,
+ "step": 3386
+ },
+ {
+ "epoch": 47.04188481675393,
+ "grad_norm": 0.034173768013715744,
+ "learning_rate": 1.146237986785356e-05,
+ "loss": 3.421980857849121,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05584642233857,
+ "grad_norm": 0.032883189618587494,
+ "learning_rate": 1.141160410235674e-05,
+ "loss": 3.4074416160583496,
+ "step": 3388
+ },
+ {
+ "epoch": 47.06980802792321,
+ "grad_norm": 0.034340377897024155,
+ "learning_rate": 1.1361063264363776e-05,
+ "loss": 3.435147762298584,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083769633507856,
+ "grad_norm": 0.03302326798439026,
+ "learning_rate": 1.1310757394594514e-05,
+ "loss": 3.4387168884277344,
+ "step": 3390
+ },
+ {
+ "epoch": 47.0977312390925,
+ "grad_norm": 0.03356143459677696,
+ "learning_rate": 1.1260686533579481e-05,
+ "loss": 3.442434072494507,
+ "step": 3391
+ },
+ {
+ "epoch": 47.11169284467714,
+ "grad_norm": 0.033129025250673294,
+ "learning_rate": 1.121085072165986e-05,
+ "loss": 3.409524440765381,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12565445026178,
+ "grad_norm": 0.03436426445841789,
+ "learning_rate": 1.1161249998987429e-05,
+ "loss": 3.422654151916504,
+ "step": 3393
+ },
+ {
+ "epoch": 47.139616055846425,
+ "grad_norm": 0.03324475511908531,
+ "learning_rate": 1.111188440552471e-05,
+ "loss": 3.440382957458496,
+ "step": 3394
+ },
+ {
+ "epoch": 47.153577661431065,
+ "grad_norm": 0.03275232017040253,
+ "learning_rate": 1.106275398104455e-05,
+ "loss": 3.4362287521362305,
+ "step": 3395
+ },
+ {
+ "epoch": 47.167539267015705,
+ "grad_norm": 0.03385675698518753,
+ "learning_rate": 1.1013858765130468e-05,
+ "loss": 3.421109437942505,
+ "step": 3396
+ },
+ {
+ "epoch": 47.181500872600346,
+ "grad_norm": 0.03292263299226761,
+ "learning_rate": 1.0965198797176513e-05,
+ "loss": 3.4364070892333984,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19546247818499,
+ "grad_norm": 0.03343760222196579,
+ "learning_rate": 1.0916774116387103e-05,
+ "loss": 3.4234514236450195,
+ "step": 3398
+ },
+ {
+ "epoch": 47.20942408376963,
+ "grad_norm": 0.03327542915940285,
+ "learning_rate": 1.0868584761777125e-05,
+ "loss": 3.4316861629486084,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223385689354274,
+ "grad_norm": 0.03442928567528725,
+ "learning_rate": 1.0820630772171869e-05,
+ "loss": 3.4337024688720703,
+ "step": 3400
+ },
+ {
+ "epoch": 47.23734729493892,
+ "grad_norm": 0.03395121544599533,
+ "learning_rate": 1.0772912186206987e-05,
+ "loss": 3.4570722579956055,
+ "step": 3401
+ },
+ {
+ "epoch": 47.25130890052356,
+ "grad_norm": 0.03375592082738876,
+ "learning_rate": 1.0725429042328542e-05,
+ "loss": 3.4370129108428955,
+ "step": 3402
+ },
+ {
+ "epoch": 47.2652705061082,
+ "grad_norm": 0.03535306826233864,
+ "learning_rate": 1.0678181378792794e-05,
+ "loss": 3.4383325576782227,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27923211169284,
+ "grad_norm": 0.0331198051571846,
+ "learning_rate": 1.0631169233666306e-05,
+ "loss": 3.4234068393707275,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29319371727749,
+ "grad_norm": 0.03407658636569977,
+ "learning_rate": 1.0584392644826016e-05,
+ "loss": 3.4324707984924316,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30715532286213,
+ "grad_norm": 0.034469254314899445,
+ "learning_rate": 1.053785164995886e-05,
+ "loss": 3.4415392875671387,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32111692844677,
+ "grad_norm": 0.03270866721868515,
+ "learning_rate": 1.0491546286562148e-05,
+ "loss": 3.4356021881103516,
+ "step": 3407
+ },
+ {
+ "epoch": 47.33507853403141,
+ "grad_norm": 0.03403601422905922,
+ "learning_rate": 1.0445476591943259e-05,
+ "loss": 3.417623519897461,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34904013961606,
+ "grad_norm": 0.03400130942463875,
+ "learning_rate": 1.0399642603219712e-05,
+ "loss": 3.4350204467773438,
+ "step": 3409
+ },
+ {
+ "epoch": 47.3630017452007,
+ "grad_norm": 0.03446796536445618,
+ "learning_rate": 1.0354044357319129e-05,
+ "loss": 3.416016101837158,
+ "step": 3410
+ },
+ {
+ "epoch": 47.37696335078534,
+ "grad_norm": 0.03428051993250847,
+ "learning_rate": 1.0308681890979173e-05,
+ "loss": 3.4321467876434326,
+ "step": 3411
+ },
+ {
+ "epoch": 47.390924956369986,
+ "grad_norm": 0.034308090806007385,
+ "learning_rate": 1.026355524074758e-05,
+ "loss": 3.4337902069091797,
+ "step": 3412
+ },
+ {
+ "epoch": 47.404886561954626,
+ "grad_norm": 0.035150643438100815,
+ "learning_rate": 1.0218664442982084e-05,
+ "loss": 3.4207749366760254,
+ "step": 3413
+ },
+ {
+ "epoch": 47.41884816753927,
+ "grad_norm": 0.03394652158021927,
+ "learning_rate": 1.0174009533850368e-05,
+ "loss": 3.4538183212280273,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43280977312391,
+ "grad_norm": 0.03471367061138153,
+ "learning_rate": 1.0129590549330119e-05,
+ "loss": 3.440670967102051,
+ "step": 3415
+ },
+ {
+ "epoch": 47.446771378708554,
+ "grad_norm": 0.03411132097244263,
+ "learning_rate": 1.0085407525208865e-05,
+ "loss": 3.4367117881774902,
+ "step": 3416
+ },
+ {
+ "epoch": 47.460732984293195,
+ "grad_norm": 0.033756449818611145,
+ "learning_rate": 1.0041460497084135e-05,
+ "loss": 3.430683135986328,
+ "step": 3417
+ },
+ {
+ "epoch": 47.474694589877835,
+ "grad_norm": 0.03421223908662796,
+ "learning_rate": 9.997749500363172e-06,
+ "loss": 3.4115402698516846,
+ "step": 3418
+ },
+ {
+ "epoch": 47.488656195462475,
+ "grad_norm": 0.034839928150177,
+ "learning_rate": 9.954274570263191e-06,
+ "loss": 3.4277877807617188,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50261780104712,
+ "grad_norm": 0.0345400907099247,
+ "learning_rate": 9.911035741811153e-06,
+ "loss": 3.4279632568359375,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51657940663176,
+ "grad_norm": 0.03440488502383232,
+ "learning_rate": 9.868033049843787e-06,
+ "loss": 3.395587205886841,
+ "step": 3421
+ },
+ {
+ "epoch": 47.5305410122164,
+ "grad_norm": 0.034132976084947586,
+ "learning_rate": 9.825266529007601e-06,
+ "loss": 3.440031051635742,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544502617801044,
+ "grad_norm": 0.033575937151908875,
+ "learning_rate": 9.782736213758782e-06,
+ "loss": 3.4644534587860107,
+ "step": 3423
+ },
+ {
+ "epoch": 47.55846422338569,
+ "grad_norm": 0.03381044790148735,
+ "learning_rate": 9.740442138363317e-06,
+ "loss": 3.432535409927368,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57242582897033,
+ "grad_norm": 0.034270722419023514,
+ "learning_rate": 9.698384336896679e-06,
+ "loss": 3.4293737411499023,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58638743455497,
+ "grad_norm": 0.03408277407288551,
+ "learning_rate": 9.656562843244114e-06,
+ "loss": 3.439113140106201,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60034904013962,
+ "grad_norm": 0.0336000993847847,
+ "learning_rate": 9.614977691100445e-06,
+ "loss": 3.4486842155456543,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61431064572426,
+ "grad_norm": 0.034019410610198975,
+ "learning_rate": 9.57362891397007e-06,
+ "loss": 3.4345788955688477,
+ "step": 3428
+ },
+ {
+ "epoch": 47.6282722513089,
+ "grad_norm": 0.035146355628967285,
+ "learning_rate": 9.532516545167004e-06,
+ "loss": 3.428173780441284,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64223385689354,
+ "grad_norm": 0.03422766923904419,
+ "learning_rate": 9.491640617814634e-06,
+ "loss": 3.4496498107910156,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65619546247819,
+ "grad_norm": 0.033867280930280685,
+ "learning_rate": 9.451001164846092e-06,
+ "loss": 3.435512065887451,
+ "step": 3431
+ },
+ {
+ "epoch": 47.67015706806283,
+ "grad_norm": 0.0348028726875782,
+ "learning_rate": 9.410598219003725e-06,
+ "loss": 3.422104835510254,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68411867364747,
+ "grad_norm": 0.03454888239502907,
+ "learning_rate": 9.370431812839556e-06,
+ "loss": 3.4214844703674316,
+ "step": 3433
+ },
+ {
+ "epoch": 47.69808027923211,
+ "grad_norm": 0.034317076206207275,
+ "learning_rate": 9.330501978714857e-06,
+ "loss": 3.4299774169921875,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712041884816756,
+ "grad_norm": 0.03441619873046875,
+ "learning_rate": 9.290808748800408e-06,
+ "loss": 3.435246467590332,
+ "step": 3435
+ },
+ {
+ "epoch": 47.726003490401396,
+ "grad_norm": 0.034184832125902176,
+ "learning_rate": 9.25135215507634e-06,
+ "loss": 3.428515911102295,
+ "step": 3436
+ },
+ {
+ "epoch": 47.73996509598604,
+ "grad_norm": 0.034739039838314056,
+ "learning_rate": 9.212132229332094e-06,
+ "loss": 3.406221389770508,
+ "step": 3437
+ },
+ {
+ "epoch": 47.753926701570684,
+ "grad_norm": 0.034154124557971954,
+ "learning_rate": 9.173149003166458e-06,
+ "loss": 3.4106082916259766,
+ "step": 3438
+ },
+ {
+ "epoch": 47.767888307155324,
+ "grad_norm": 0.034395184367895126,
+ "learning_rate": 9.134402507987533e-06,
+ "loss": 3.4354968070983887,
+ "step": 3439
+ },
+ {
+ "epoch": 47.781849912739965,
+ "grad_norm": 0.03440655767917633,
+ "learning_rate": 9.095892775012672e-06,
+ "loss": 3.424171209335327,
+ "step": 3440
+ },
+ {
+ "epoch": 47.795811518324605,
+ "grad_norm": 0.03425966203212738,
+ "learning_rate": 9.057619835268437e-06,
+ "loss": 3.407989025115967,
+ "step": 3441
+ },
+ {
+ "epoch": 47.80977312390925,
+ "grad_norm": 0.034892335534095764,
+ "learning_rate": 9.019583719590641e-06,
+ "loss": 3.419440746307373,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82373472949389,
+ "grad_norm": 0.03456826135516167,
+ "learning_rate": 8.981784458624375e-06,
+ "loss": 3.412938117980957,
+ "step": 3443
+ },
+ {
+ "epoch": 47.83769633507853,
+ "grad_norm": 0.034370701760053635,
+ "learning_rate": 8.944222082823717e-06,
+ "loss": 3.4475297927856445,
+ "step": 3444
+ },
+ {
+ "epoch": 47.85165794066317,
+ "grad_norm": 0.035071875900030136,
+ "learning_rate": 8.906896622452055e-06,
+ "loss": 3.446824550628662,
+ "step": 3445
+ },
+ {
+ "epoch": 47.86561954624782,
+ "grad_norm": 0.03523378074169159,
+ "learning_rate": 8.86980810758183e-06,
+ "loss": 3.4080722332000732,
+ "step": 3446
+ },
+ {
+ "epoch": 47.87958115183246,
+ "grad_norm": 0.03434052690863609,
+ "learning_rate": 8.832956568094596e-06,
+ "loss": 3.4339864253997803,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8935427574171,
+ "grad_norm": 0.03419225662946701,
+ "learning_rate": 8.796342033680928e-06,
+ "loss": 3.397799491882324,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90750436300174,
+ "grad_norm": 0.03464251011610031,
+ "learning_rate": 8.759964533840574e-06,
+ "loss": 3.4296271800994873,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92146596858639,
+ "grad_norm": 0.034543540328741074,
+ "learning_rate": 8.723824097882145e-06,
+ "loss": 3.421177387237549,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93542757417103,
+ "grad_norm": 0.03456886112689972,
+ "learning_rate": 8.687920754923393e-06,
+ "loss": 3.425032615661621,
+ "step": 3451
+ },
+ {
+ "epoch": 47.94938917975567,
+ "grad_norm": 0.03451840579509735,
+ "learning_rate": 8.652254533890956e-06,
+ "loss": 3.435209274291992,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96335078534032,
+ "grad_norm": 0.0347096286714077,
+ "learning_rate": 8.616825463520493e-06,
+ "loss": 3.457643508911133,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97731239092496,
+ "grad_norm": 0.03466152399778366,
+ "learning_rate": 8.581633572356576e-06,
+ "loss": 3.4021284580230713,
+ "step": 3454
+ },
+ {
+ "epoch": 47.9912739965096,
+ "grad_norm": 0.034237995743751526,
+ "learning_rate": 8.546678888752629e-06,
+ "loss": 3.4286880493164062,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.027228690683841705,
+ "learning_rate": 8.51196144087103e-06,
+ "loss": 2.1402902603149414,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 0.6273696422576904,
+ "eval_runtime": 59.3508,
+ "eval_samples_per_second": 41.145,
+ "eval_steps_per_second": 0.657,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01396160558464,
+ "grad_norm": 0.03215591609477997,
+ "learning_rate": 8.477481256683005e-06,
+ "loss": 3.3785500526428223,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02792321116929,
+ "grad_norm": 0.0311955027282238,
+ "learning_rate": 8.44323836396864e-06,
+ "loss": 3.428605079650879,
+ "step": 3458
+ },
+ {
+ "epoch": 48.04188481675393,
+ "grad_norm": 0.03140958398580551,
+ "learning_rate": 8.409232790316763e-06,
+ "loss": 3.434476852416992,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05584642233857,
+ "grad_norm": 0.03168640658259392,
+ "learning_rate": 8.375464563125094e-06,
+ "loss": 3.438292980194092,
+ "step": 3460
+ },
+ {
+ "epoch": 48.06980802792321,
+ "grad_norm": 0.031872380524873734,
+ "learning_rate": 8.341933709600102e-06,
+ "loss": 3.440722703933716,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083769633507856,
+ "grad_norm": 0.0315161794424057,
+ "learning_rate": 8.308640256756916e-06,
+ "loss": 3.4365200996398926,
+ "step": 3462
+ },
+ {
+ "epoch": 48.0977312390925,
+ "grad_norm": 0.031423915177583694,
+ "learning_rate": 8.275584231419512e-06,
+ "loss": 3.4143881797790527,
+ "step": 3463
+ },
+ {
+ "epoch": 48.11169284467714,
+ "grad_norm": 0.03173846751451492,
+ "learning_rate": 8.24276566022055e-06,
+ "loss": 3.4107308387756348,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12565445026178,
+ "grad_norm": 0.03221066668629646,
+ "learning_rate": 8.21018456960132e-06,
+ "loss": 3.409773349761963,
+ "step": 3465
+ },
+ {
+ "epoch": 48.139616055846425,
+ "grad_norm": 0.03243516385555267,
+ "learning_rate": 8.177840985811858e-06,
+ "loss": 3.448481559753418,
+ "step": 3466
+ },
+ {
+ "epoch": 48.153577661431065,
+ "grad_norm": 0.03129534423351288,
+ "learning_rate": 8.145734934910757e-06,
+ "loss": 3.4231836795806885,
+ "step": 3467
+ },
+ {
+ "epoch": 48.167539267015705,
+ "grad_norm": 0.03119921125471592,
+ "learning_rate": 8.113866442765364e-06,
+ "loss": 3.4331789016723633,
+ "step": 3468
+ },
+ {
+ "epoch": 48.181500872600346,
+ "grad_norm": 0.032179273664951324,
+ "learning_rate": 8.082235535051449e-06,
+ "loss": 3.424485921859741,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19546247818499,
+ "grad_norm": 0.0314272977411747,
+ "learning_rate": 8.050842237253533e-06,
+ "loss": 3.450509786605835,
+ "step": 3470
+ },
+ {
+ "epoch": 48.20942408376963,
+ "grad_norm": 0.031465087085962296,
+ "learning_rate": 8.019686574664564e-06,
+ "loss": 3.434774875640869,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223385689354274,
+ "grad_norm": 0.0315382219851017,
+ "learning_rate": 7.988768572386171e-06,
+ "loss": 3.439669609069824,
+ "step": 3472
+ },
+ {
+ "epoch": 48.23734729493892,
+ "grad_norm": 0.03274233266711235,
+ "learning_rate": 7.958088255328415e-06,
+ "loss": 3.4329638481140137,
+ "step": 3473
+ },
+ {
+ "epoch": 48.25130890052356,
+ "grad_norm": 0.032395631074905396,
+ "learning_rate": 7.927645648209837e-06,
+ "loss": 3.4123125076293945,
+ "step": 3474
+ },
+ {
+ "epoch": 48.2652705061082,
+ "grad_norm": 0.03218825161457062,
+ "learning_rate": 7.897440775557506e-06,
+ "loss": 3.4242851734161377,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27923211169284,
+ "grad_norm": 0.03341231867671013,
+ "learning_rate": 7.867473661706946e-06,
+ "loss": 3.442500591278076,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29319371727749,
+ "grad_norm": 0.0315360464155674,
+ "learning_rate": 7.837744330802136e-06,
+ "loss": 3.423917293548584,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30715532286213,
+ "grad_norm": 0.03201025351881981,
+ "learning_rate": 7.808252806795446e-06,
+ "loss": 3.4042210578918457,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32111692844677,
+ "grad_norm": 0.032367199659347534,
+ "learning_rate": 7.778999113447673e-06,
+ "loss": 3.418100357055664,
+ "step": 3479
+ },
+ {
+ "epoch": 48.33507853403141,
+ "grad_norm": 0.032645534723997116,
+ "learning_rate": 7.749983274327994e-06,
+ "loss": 3.4184858798980713,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34904013961606,
+ "grad_norm": 0.03230610117316246,
+ "learning_rate": 7.721205312813921e-06,
+ "loss": 3.4303946495056152,
+ "step": 3481
+ },
+ {
+ "epoch": 48.3630017452007,
+ "grad_norm": 0.03246692195534706,
+ "learning_rate": 7.69266525209132e-06,
+ "loss": 3.4086356163024902,
+ "step": 3482
+ },
+ {
+ "epoch": 48.37696335078534,
+ "grad_norm": 0.032483771443367004,
+ "learning_rate": 7.664363115154476e-06,
+ "loss": 3.3896312713623047,
+ "step": 3483
+ },
+ {
+ "epoch": 48.390924956369986,
+ "grad_norm": 0.03206900879740715,
+ "learning_rate": 7.636298924805872e-06,
+ "loss": 3.4026384353637695,
+ "step": 3484
+ },
+ {
+ "epoch": 48.404886561954626,
+ "grad_norm": 0.03220972418785095,
+ "learning_rate": 7.608472703656282e-06,
+ "loss": 3.4464452266693115,
+ "step": 3485
+ },
+ {
+ "epoch": 48.41884816753927,
+ "grad_norm": 0.03237165883183479,
+ "learning_rate": 7.580884474124833e-06,
+ "loss": 3.4372801780700684,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43280977312391,
+ "grad_norm": 0.033048976212739944,
+ "learning_rate": 7.55353425843888e-06,
+ "loss": 3.4144935607910156,
+ "step": 3487
+ },
+ {
+ "epoch": 48.446771378708554,
+ "grad_norm": 0.031966496258974075,
+ "learning_rate": 7.526422078633939e-06,
+ "loss": 3.4209954738616943,
+ "step": 3488
+ },
+ {
+ "epoch": 48.460732984293195,
+ "grad_norm": 0.032010894268751144,
+ "learning_rate": 7.4995479565538476e-06,
+ "loss": 3.4237444400787354,
+ "step": 3489
+ },
+ {
+ "epoch": 48.474694589877835,
+ "grad_norm": 0.03210063651204109,
+ "learning_rate": 7.4729119138506044e-06,
+ "loss": 3.4414892196655273,
+ "step": 3490
+ },
+ {
+ "epoch": 48.488656195462475,
+ "grad_norm": 0.03240114077925682,
+ "learning_rate": 7.446513971984366e-06,
+ "loss": 3.390763282775879,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50261780104712,
+ "grad_norm": 0.0321781150996685,
+ "learning_rate": 7.420354152223485e-06,
+ "loss": 3.454559326171875,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51657940663176,
+ "grad_norm": 0.03208549693226814,
+ "learning_rate": 7.394432475644402e-06,
+ "loss": 3.439821481704712,
+ "step": 3493
+ },
+ {
+ "epoch": 48.5305410122164,
+ "grad_norm": 0.0323772095143795,
+ "learning_rate": 7.368748963131854e-06,
+ "loss": 3.4085192680358887,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544502617801044,
+ "grad_norm": 0.03271934390068054,
+ "learning_rate": 7.343303635378537e-06,
+ "loss": 3.4317445755004883,
+ "step": 3495
+ },
+ {
+ "epoch": 48.55846422338569,
+ "grad_norm": 0.03232505917549133,
+ "learning_rate": 7.31809651288524e-06,
+ "loss": 3.408017635345459,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57242582897033,
+ "grad_norm": 0.03184394910931587,
+ "learning_rate": 7.2931276159609475e-06,
+ "loss": 3.428473949432373,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58638743455497,
+ "grad_norm": 0.03233828395605087,
+ "learning_rate": 7.268396964722602e-06,
+ "loss": 3.417452812194824,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60034904013962,
+ "grad_norm": 0.032878946512937546,
+ "learning_rate": 7.243904579095308e-06,
+ "loss": 3.3997645378112793,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61431064572426,
+ "grad_norm": 0.03272545337677002,
+ "learning_rate": 7.219650478812065e-06,
+ "loss": 3.4124348163604736,
+ "step": 3500
+ },
+ {
+ "epoch": 48.6282722513089,
+ "grad_norm": 0.03187870979309082,
+ "learning_rate": 7.195634683413996e-06,
+ "loss": 3.3969287872314453,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64223385689354,
+ "grad_norm": 0.03188911825418472,
+ "learning_rate": 7.171857212250225e-06,
+ "loss": 3.4451255798339844,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65619546247819,
+ "grad_norm": 0.032626260071992874,
+ "learning_rate": 7.148318084477766e-06,
+ "loss": 3.413586139678955,
+ "step": 3503
+ },
+ {
+ "epoch": 48.67015706806283,
+ "grad_norm": 0.03300228342413902,
+ "learning_rate": 7.125017319061728e-06,
+ "loss": 3.433492660522461,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68411867364747,
+ "grad_norm": 0.032605383545160294,
+ "learning_rate": 7.101954934775084e-06,
+ "loss": 3.4164743423461914,
+ "step": 3505
+ },
+ {
+ "epoch": 48.69808027923211,
+ "grad_norm": 0.03274476155638695,
+ "learning_rate": 7.079130950198802e-06,
+ "loss": 3.4328882694244385,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712041884816756,
+ "grad_norm": 0.03277606889605522,
+ "learning_rate": 7.05654538372171e-06,
+ "loss": 3.424349784851074,
+ "step": 3507
+ },
+ {
+ "epoch": 48.726003490401396,
+ "grad_norm": 0.03255699947476387,
+ "learning_rate": 7.034198253540665e-06,
+ "loss": 3.434109687805176,
+ "step": 3508
+ },
+ {
+ "epoch": 48.73996509598604,
+ "grad_norm": 0.0332169346511364,
+ "learning_rate": 7.012089577660255e-06,
+ "loss": 3.4454803466796875,
+ "step": 3509
+ },
+ {
+ "epoch": 48.753926701570684,
+ "grad_norm": 0.03308543562889099,
+ "learning_rate": 6.990219373893128e-06,
+ "loss": 3.4382896423339844,
+ "step": 3510
+ },
+ {
+ "epoch": 48.767888307155324,
+ "grad_norm": 0.03211184963583946,
+ "learning_rate": 6.968587659859665e-06,
+ "loss": 3.4157304763793945,
+ "step": 3511
+ },
+ {
+ "epoch": 48.781849912739965,
+ "grad_norm": 0.03240763023495674,
+ "learning_rate": 6.9471944529881376e-06,
+ "loss": 3.414407253265381,
+ "step": 3512
+ },
+ {
+ "epoch": 48.795811518324605,
+ "grad_norm": 0.032774582505226135,
+ "learning_rate": 6.926039770514686e-06,
+ "loss": 3.4374051094055176,
+ "step": 3513
+ },
+ {
+ "epoch": 48.80977312390925,
+ "grad_norm": 0.03298104181885719,
+ "learning_rate": 6.905123629483277e-06,
+ "loss": 3.4332118034362793,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82373472949389,
+ "grad_norm": 0.03287169337272644,
+ "learning_rate": 6.884446046745608e-06,
+ "loss": 3.4131298065185547,
+ "step": 3515
+ },
+ {
+ "epoch": 48.83769633507853,
+ "grad_norm": 0.03234318271279335,
+ "learning_rate": 6.8640070389613055e-06,
+ "loss": 3.4444355964660645,
+ "step": 3516
+ },
+ {
+ "epoch": 48.85165794066317,
+ "grad_norm": 0.03258844092488289,
+ "learning_rate": 6.843806622597661e-06,
+ "loss": 3.431270122528076,
+ "step": 3517
+ },
+ {
+ "epoch": 48.86561954624782,
+ "grad_norm": 0.033729180693626404,
+ "learning_rate": 6.823844813929796e-06,
+ "loss": 3.419201374053955,
+ "step": 3518
+ },
+ {
+ "epoch": 48.87958115183246,
+ "grad_norm": 0.03248931095004082,
+ "learning_rate": 6.804121629040597e-06,
+ "loss": 3.4233977794647217,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8935427574171,
+ "grad_norm": 0.0324002243578434,
+ "learning_rate": 6.7846370838206455e-06,
+ "loss": 3.4010872840881348,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90750436300174,
+ "grad_norm": 0.032746996730566025,
+ "learning_rate": 6.765391193968287e-06,
+ "loss": 3.4540910720825195,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92146596858639,
+ "grad_norm": 0.03319564461708069,
+ "learning_rate": 6.7463839749896335e-06,
+ "loss": 3.432769775390625,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93542757417103,
+ "grad_norm": 0.03328274190425873,
+ "learning_rate": 6.727615442198394e-06,
+ "loss": 3.4270334243774414,
+ "step": 3523
+ },
+ {
+ "epoch": 48.94938917975567,
+ "grad_norm": 0.03210938349366188,
+ "learning_rate": 6.709085610716107e-06,
+ "loss": 3.4090805053710938,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96335078534032,
+ "grad_norm": 0.03288741782307625,
+ "learning_rate": 6.690794495471878e-06,
+ "loss": 3.417360544204712,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97731239092496,
+ "grad_norm": 0.032983556389808655,
+ "learning_rate": 6.672742111202543e-06,
+ "loss": 3.428494453430176,
+ "step": 3526
+ },
+ {
+ "epoch": 48.9912739965096,
+ "grad_norm": 0.03385569155216217,
+ "learning_rate": 6.654928472452569e-06,
+ "loss": 3.4049391746520996,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.025613969191908836,
+ "learning_rate": 6.637353593574088e-06,
+ "loss": 2.1437878608703613,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 0.6278687119483948,
+ "eval_runtime": 60.0591,
+ "eval_samples_per_second": 40.66,
+ "eval_steps_per_second": 0.649,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01396160558464,
+ "grad_norm": 0.030343951657414436,
+ "learning_rate": 6.620017488726867e-06,
+ "loss": 3.4471704959869385,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02792321116929,
+ "grad_norm": 0.030530281364917755,
+ "learning_rate": 6.602920171878265e-06,
+ "loss": 3.406923294067383,
+ "step": 3530
+ },
+ {
+ "epoch": 49.04188481675393,
+ "grad_norm": 0.030529819428920746,
+ "learning_rate": 6.5860616568033465e-06,
+ "loss": 3.427823543548584,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05584642233857,
+ "grad_norm": 0.030696529895067215,
+ "learning_rate": 6.569441957084671e-06,
+ "loss": 3.40974760055542,
+ "step": 3532
+ },
+ {
+ "epoch": 49.06980802792321,
+ "grad_norm": 0.030391080304980278,
+ "learning_rate": 6.5530610861124e-06,
+ "loss": 3.4272167682647705,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083769633507856,
+ "grad_norm": 0.03033747524023056,
+ "learning_rate": 6.536919057084357e-06,
+ "loss": 3.4266576766967773,
+ "step": 3534
+ },
+ {
+ "epoch": 49.0977312390925,
+ "grad_norm": 0.031158700585365295,
+ "learning_rate": 6.521015883005866e-06,
+ "loss": 3.41561222076416,
+ "step": 3535
+ },
+ {
+ "epoch": 49.11169284467714,
+ "grad_norm": 0.03036365658044815,
+ "learning_rate": 6.5053515766898195e-06,
+ "loss": 3.446746587753296,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12565445026178,
+ "grad_norm": 0.030888007953763008,
+ "learning_rate": 6.489926150756674e-06,
+ "loss": 3.449531078338623,
+ "step": 3537
+ },
+ {
+ "epoch": 49.139616055846425,
+ "grad_norm": 0.03136712312698364,
+ "learning_rate": 6.4747396176344176e-06,
+ "loss": 3.423194408416748,
+ "step": 3538
+ },
+ {
+ "epoch": 49.153577661431065,
+ "grad_norm": 0.030854566022753716,
+ "learning_rate": 6.459791989558542e-06,
+ "loss": 3.4471354484558105,
+ "step": 3539
+ },
+ {
+ "epoch": 49.167539267015705,
+ "grad_norm": 0.031110677868127823,
+ "learning_rate": 6.445083278572134e-06,
+ "loss": 3.391568183898926,
+ "step": 3540
+ },
+ {
+ "epoch": 49.181500872600346,
+ "grad_norm": 0.03041466511785984,
+ "learning_rate": 6.430613496525685e-06,
+ "loss": 3.4078521728515625,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19546247818499,
+ "grad_norm": 0.030827714130282402,
+ "learning_rate": 6.416382655077248e-06,
+ "loss": 3.4390079975128174,
+ "step": 3542
+ },
+ {
+ "epoch": 49.20942408376963,
+ "grad_norm": 0.03015112690627575,
+ "learning_rate": 6.402390765692381e-06,
+ "loss": 3.4455251693725586,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223385689354274,
+ "grad_norm": 0.0314689539372921,
+ "learning_rate": 6.388637839644073e-06,
+ "loss": 3.4379444122314453,
+ "step": 3544
+ },
+ {
+ "epoch": 49.23734729493892,
+ "grad_norm": 0.03094467893242836,
+ "learning_rate": 6.375123888012812e-06,
+ "loss": 3.4084763526916504,
+ "step": 3545
+ },
+ {
+ "epoch": 49.25130890052356,
+ "grad_norm": 0.031439658254384995,
+ "learning_rate": 6.361848921686523e-06,
+ "loss": 3.42118501663208,
+ "step": 3546
+ },
+ {
+ "epoch": 49.2652705061082,
+ "grad_norm": 0.030968086794018745,
+ "learning_rate": 6.348812951360662e-06,
+ "loss": 3.4220309257507324,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27923211169284,
+ "grad_norm": 0.030803948640823364,
+ "learning_rate": 6.33601598753802e-06,
+ "loss": 3.4087119102478027,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29319371727749,
+ "grad_norm": 0.031383004039525986,
+ "learning_rate": 6.323458040528923e-06,
+ "loss": 3.4118404388427734,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30715532286213,
+ "grad_norm": 0.030414072796702385,
+ "learning_rate": 6.311139120451062e-06,
+ "loss": 3.404762029647827,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32111692844677,
+ "grad_norm": 0.030658090487122536,
+ "learning_rate": 6.299059237229567e-06,
+ "loss": 3.4134716987609863,
+ "step": 3551
+ },
+ {
+ "epoch": 49.33507853403141,
+ "grad_norm": 0.031208444386720657,
+ "learning_rate": 6.287218400596961e-06,
+ "loss": 3.3999810218811035,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34904013961606,
+ "grad_norm": 0.030635200440883636,
+ "learning_rate": 6.27561662009321e-06,
+ "loss": 3.4217896461486816,
+ "step": 3553
+ },
+ {
+ "epoch": 49.3630017452007,
+ "grad_norm": 0.031378358602523804,
+ "learning_rate": 6.264253905065675e-06,
+ "loss": 3.419522762298584,
+ "step": 3554
+ },
+ {
+ "epoch": 49.37696335078534,
+ "grad_norm": 0.031133271753787994,
+ "learning_rate": 6.2531302646690866e-06,
+ "loss": 3.4028677940368652,
+ "step": 3555
+ },
+ {
+ "epoch": 49.390924956369986,
+ "grad_norm": 0.031054602935910225,
+ "learning_rate": 6.242245707865545e-06,
+ "loss": 3.414252281188965,
+ "step": 3556
+ },
+ {
+ "epoch": 49.404886561954626,
+ "grad_norm": 0.03133461996912956,
+ "learning_rate": 6.23160024342455e-06,
+ "loss": 3.43626070022583,
+ "step": 3557
+ },
+ {
+ "epoch": 49.41884816753927,
+ "grad_norm": 0.030775003135204315,
+ "learning_rate": 6.221193879922969e-06,
+ "loss": 3.3899011611938477,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43280977312391,
+ "grad_norm": 0.03135809302330017,
+ "learning_rate": 6.211026625745005e-06,
+ "loss": 3.4367318153381348,
+ "step": 3559
+ },
+ {
+ "epoch": 49.446771378708554,
+ "grad_norm": 0.031782686710357666,
+ "learning_rate": 6.201098489082232e-06,
+ "loss": 3.432323455810547,
+ "step": 3560
+ },
+ {
+ "epoch": 49.460732984293195,
+ "grad_norm": 0.031206553801894188,
+ "learning_rate": 6.191409477933554e-06,
+ "loss": 3.419071674346924,
+ "step": 3561
+ },
+ {
+ "epoch": 49.474694589877835,
+ "grad_norm": 0.031115612015128136,
+ "learning_rate": 6.18195960010528e-06,
+ "loss": 3.4201083183288574,
+ "step": 3562
+ },
+ {
+ "epoch": 49.488656195462475,
+ "grad_norm": 0.03141428902745247,
+ "learning_rate": 6.172748863210954e-06,
+ "loss": 3.4002976417541504,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50261780104712,
+ "grad_norm": 0.030737776309251785,
+ "learning_rate": 6.163777274671552e-06,
+ "loss": 3.4197514057159424,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51657940663176,
+ "grad_norm": 0.031241273507475853,
+ "learning_rate": 6.155044841715289e-06,
+ "loss": 3.400420665740967,
+ "step": 3565
+ },
+ {
+ "epoch": 49.5305410122164,
+ "grad_norm": 0.03118479810655117,
+ "learning_rate": 6.146551571377715e-06,
+ "loss": 3.4250917434692383,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544502617801044,
+ "grad_norm": 0.030931556597352028,
+ "learning_rate": 6.1382974705017456e-06,
+ "loss": 3.4266953468322754,
+ "step": 3567
+ },
+ {
+ "epoch": 49.55846422338569,
+ "grad_norm": 0.0308473352342844,
+ "learning_rate": 6.130282545737535e-06,
+ "loss": 3.421746253967285,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57242582897033,
+ "grad_norm": 0.031132593750953674,
+ "learning_rate": 6.1225068035425725e-06,
+ "loss": 3.407823085784912,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58638743455497,
+ "grad_norm": 0.031208796426653862,
+ "learning_rate": 6.114970250181616e-06,
+ "loss": 3.4003689289093018,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60034904013962,
+ "grad_norm": 0.03152323514223099,
+ "learning_rate": 6.107672891726724e-06,
+ "loss": 3.4073328971862793,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61431064572426,
+ "grad_norm": 0.031811248511075974,
+ "learning_rate": 6.1006147340572916e-06,
+ "loss": 3.44144868850708,
+ "step": 3572
+ },
+ {
+ "epoch": 49.6282722513089,
+ "grad_norm": 0.03205077350139618,
+ "learning_rate": 6.093795782859917e-06,
+ "loss": 3.413670063018799,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64223385689354,
+ "grad_norm": 0.03199009224772453,
+ "learning_rate": 6.087216043628502e-06,
+ "loss": 3.416214942932129,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65619546247819,
+ "grad_norm": 0.031675588339567184,
+ "learning_rate": 6.0808755216642146e-06,
+ "loss": 3.4274444580078125,
+ "step": 3575
+ },
+ {
+ "epoch": 49.67015706806283,
+ "grad_norm": 0.031750451773405075,
+ "learning_rate": 6.07477422207556e-06,
+ "loss": 3.431197166442871,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68411867364747,
+ "grad_norm": 0.03186594322323799,
+ "learning_rate": 6.06891214977818e-06,
+ "loss": 3.4129209518432617,
+ "step": 3577
+ },
+ {
+ "epoch": 49.69808027923211,
+ "grad_norm": 0.03211659938097,
+ "learning_rate": 6.063289309495051e-06,
+ "loss": 3.4306721687316895,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712041884816756,
+ "grad_norm": 0.031404994428157806,
+ "learning_rate": 6.057905705756421e-06,
+ "loss": 3.4264345169067383,
+ "step": 3579
+ },
+ {
+ "epoch": 49.726003490401396,
+ "grad_norm": 0.03126844763755798,
+ "learning_rate": 6.052761342899737e-06,
+ "loss": 3.454092264175415,
+ "step": 3580
+ },
+ {
+ "epoch": 49.73996509598604,
+ "grad_norm": 0.03277238458395004,
+ "learning_rate": 6.0478562250697185e-06,
+ "loss": 3.4290542602539062,
+ "step": 3581
+ },
+ {
+ "epoch": 49.753926701570684,
+ "grad_norm": 0.03199701011180878,
+ "learning_rate": 6.043190356218355e-06,
+ "loss": 3.4044201374053955,
+ "step": 3582
+ },
+ {
+ "epoch": 49.767888307155324,
+ "grad_norm": 0.03167097270488739,
+ "learning_rate": 6.038763740104835e-06,
+ "loss": 3.4093637466430664,
+ "step": 3583
+ },
+ {
+ "epoch": 49.781849912739965,
+ "grad_norm": 0.03169747442007065,
+ "learning_rate": 6.034576380295586e-06,
+ "loss": 3.4041571617126465,
+ "step": 3584
+ },
+ {
+ "epoch": 49.795811518324605,
+ "grad_norm": 0.03176243230700493,
+ "learning_rate": 6.030628280164338e-06,
+ "loss": 3.445474624633789,
+ "step": 3585
+ },
+ {
+ "epoch": 49.80977312390925,
+ "grad_norm": 0.031467586755752563,
+ "learning_rate": 6.026919442891958e-06,
+ "loss": 3.4312257766723633,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82373472949389,
+ "grad_norm": 0.031910646706819534,
+ "learning_rate": 6.023449871466615e-06,
+ "loss": 3.392921209335327,
+ "step": 3587
+ },
+ {
+ "epoch": 49.83769633507853,
+ "grad_norm": 0.03192969784140587,
+ "learning_rate": 6.020219568683678e-06,
+ "loss": 3.43118953704834,
+ "step": 3588
+ },
+ {
+ "epoch": 49.85165794066317,
+ "grad_norm": 0.031732942909002304,
+ "learning_rate": 6.0172285371456925e-06,
+ "loss": 3.4532341957092285,
+ "step": 3589
+ },
+ {
+ "epoch": 49.86561954624782,
+ "grad_norm": 0.032593902200460434,
+ "learning_rate": 6.014476779262568e-06,
+ "loss": 3.426077365875244,
+ "step": 3590
+ },
+ {
+ "epoch": 49.87958115183246,
+ "grad_norm": 0.03203500434756279,
+ "learning_rate": 6.011964297251285e-06,
+ "loss": 3.4295597076416016,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8935427574171,
+ "grad_norm": 0.03187621384859085,
+ "learning_rate": 6.0096910931361275e-06,
+ "loss": 3.4121551513671875,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90750436300174,
+ "grad_norm": 0.031860485672950745,
+ "learning_rate": 6.007657168748551e-06,
+ "loss": 3.4054465293884277,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92146596858639,
+ "grad_norm": 0.0321529321372509,
+ "learning_rate": 6.005862525727277e-06,
+ "loss": 3.431338310241699,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93542757417103,
+ "grad_norm": 0.03187413141131401,
+ "learning_rate": 6.0043071655181986e-06,
+ "loss": 3.4150989055633545,
+ "step": 3595
+ },
+ {
+ "epoch": 49.94938917975567,
+ "grad_norm": 0.032197415828704834,
+ "learning_rate": 6.002991089374446e-06,
+ "loss": 3.4240708351135254,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96335078534032,
+ "grad_norm": 0.03229861333966255,
+ "learning_rate": 6.0019142983563875e-06,
+ "loss": 3.418485164642334,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97731239092496,
+ "grad_norm": 0.03289807215332985,
+ "learning_rate": 6.001076793331555e-06,
+ "loss": 3.4143126010894775,
+ "step": 3598
+ },
+ {
+ "epoch": 49.9912739965096,
+ "grad_norm": 0.0329769104719162,
+ "learning_rate": 6.000478574974689e-06,
+ "loss": 3.417052984237671,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.02592312917113304,
+ "learning_rate": 6.000119643767767e-06,
+ "loss": 2.14823579788208,
+ "step": 3600
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": true
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.5361471231819776e+18,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-3600/training_args.bin b/runs/l2r90-wd0033-ssl/checkpoint-3600/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..313c41733ff55284c06034a59bb885b30e40ffc8
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-3600/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7ddd0bdddba68d3e19ab33477000f3fc684fe4ee8e3580eeb8c02646e5fc27ac
+size 4856
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-72/chat_template.jinja b/runs/l2r90-wd0033-ssl/checkpoint-72/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-72/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-72/config.json b/runs/l2r90-wd0033-ssl/checkpoint-72/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-72/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-72/generation_config.json b/runs/l2r90-wd0033-ssl/checkpoint-72/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-72/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-72/model.safetensors b/runs/l2r90-wd0033-ssl/checkpoint-72/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..2daf53fce2270e6d938163a9a3e94dedc591a345
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-72/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:09600b03de9187c854f050da6662b8d5a9ddd4f72be3f8a0c562ea2b62e37ebd
+size 583356232
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-72/optimizer.pt b/runs/l2r90-wd0033-ssl/checkpoint-72/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..dc6e2ea02db99647e6e18d858f4dee9793b128f8
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-72/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1074212f2d0b2b7f12721de1cf70df4cdc689aed1efe323f7e592fea896923ac
+size 1166825338
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-72/rng_state_0.pth b/runs/l2r90-wd0033-ssl/checkpoint-72/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..e4121d856142b0c74c1be3f8922d6b03f08a6259
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-72/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ad792af33c7cfa8b15298ecc9d976ebdcdeb444ca0e704c7b0657f41ee6547eb
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-72/rng_state_1.pth b/runs/l2r90-wd0033-ssl/checkpoint-72/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..fd01aa502c5cd06d5b3af549d6b6c3499ee4f60b
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-72/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:722c924fceffd85f8ab1a5445f1ea1e6c502644b6a42e2ff6b5a9a76ea26e1fe
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-72/scheduler.pt b/runs/l2r90-wd0033-ssl/checkpoint-72/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..43cf114590f1088fd438450226c42885d843da41
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-72/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:878ec53f5c2a524fc37af578b94a9025923faefea55f621be8e5643a79915d0f
+size 1064
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-72/tokenizer.json b/runs/l2r90-wd0033-ssl/checkpoint-72/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-72/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-72/tokenizer_config.json b/runs/l2r90-wd0033-ssl/checkpoint-72/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-72/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-72/trainer_state.json b/runs/l2r90-wd0033-ssl/checkpoint-72/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..5fb4ada60cbc64a674b66feea8944e59c626939e
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-72/trainer_state.json
@@ -0,0 +1,538 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 1.0,
+ "eval_steps": 500,
+ "global_step": 72,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013961605584642234,
+ "grad_norm": 0.26733073592185974,
+ "learning_rate": 0.0,
+ "loss": 7.990724086761475,
+ "step": 1
+ },
+ {
+ "epoch": 0.027923211169284468,
+ "grad_norm": 0.26712340116500854,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 7.990110874176025,
+ "step": 2
+ },
+ {
+ "epoch": 0.041884816753926704,
+ "grad_norm": 0.2646932601928711,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 7.956935405731201,
+ "step": 3
+ },
+ {
+ "epoch": 0.055846422338568937,
+ "grad_norm": 0.24684545397758484,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 7.890539646148682,
+ "step": 4
+ },
+ {
+ "epoch": 0.06980802792321117,
+ "grad_norm": 0.21016643941402435,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 7.8146138191223145,
+ "step": 5
+ },
+ {
+ "epoch": 0.08376963350785341,
+ "grad_norm": 0.17992742359638214,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 7.724307060241699,
+ "step": 6
+ },
+ {
+ "epoch": 0.09773123909249563,
+ "grad_norm": 0.14356893301010132,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 7.645662307739258,
+ "step": 7
+ },
+ {
+ "epoch": 0.11169284467713787,
+ "grad_norm": 0.10830964893102646,
+ "learning_rate": 4.2e-05,
+ "loss": 7.582019329071045,
+ "step": 8
+ },
+ {
+ "epoch": 0.1256544502617801,
+ "grad_norm": 0.08265161514282227,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 7.5230793952941895,
+ "step": 9
+ },
+ {
+ "epoch": 0.13961605584642234,
+ "grad_norm": 0.06423710286617279,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 7.473812103271484,
+ "step": 10
+ },
+ {
+ "epoch": 0.15357766143106458,
+ "grad_norm": 0.05132246017456055,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 7.440786838531494,
+ "step": 11
+ },
+ {
+ "epoch": 0.16753926701570682,
+ "grad_norm": 0.040572404861450195,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 7.41290283203125,
+ "step": 12
+ },
+ {
+ "epoch": 0.18150087260034903,
+ "grad_norm": 0.03138705715537071,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 7.39749002456665,
+ "step": 13
+ },
+ {
+ "epoch": 0.19546247818499127,
+ "grad_norm": 0.02553008496761322,
+ "learning_rate": 7.8e-05,
+ "loss": 7.387092590332031,
+ "step": 14
+ },
+ {
+ "epoch": 0.2094240837696335,
+ "grad_norm": 0.022178977727890015,
+ "learning_rate": 8.4e-05,
+ "loss": 7.378993511199951,
+ "step": 15
+ },
+ {
+ "epoch": 0.22338568935427575,
+ "grad_norm": 0.019915293902158737,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 7.37220573425293,
+ "step": 16
+ },
+ {
+ "epoch": 0.23734729493891799,
+ "grad_norm": 0.017390629276633263,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 7.366759300231934,
+ "step": 17
+ },
+ {
+ "epoch": 0.2513089005235602,
+ "grad_norm": 0.015472771599888802,
+ "learning_rate": 0.000102,
+ "loss": 7.3652663230896,
+ "step": 18
+ },
+ {
+ "epoch": 0.26527050610820246,
+ "grad_norm": 0.013599214144051075,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 7.359714984893799,
+ "step": 19
+ },
+ {
+ "epoch": 0.2792321116928447,
+ "grad_norm": 0.012442861683666706,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 7.353302001953125,
+ "step": 20
+ },
+ {
+ "epoch": 0.2931937172774869,
+ "grad_norm": 0.010471965186297894,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 7.357417583465576,
+ "step": 21
+ },
+ {
+ "epoch": 0.30715532286212915,
+ "grad_norm": 0.009647224098443985,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 7.351763725280762,
+ "step": 22
+ },
+ {
+ "epoch": 0.32111692844677137,
+ "grad_norm": 0.008228210732340813,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 7.345434188842773,
+ "step": 23
+ },
+ {
+ "epoch": 0.33507853403141363,
+ "grad_norm": 0.007454514969140291,
+ "learning_rate": 0.000138,
+ "loss": 7.342520236968994,
+ "step": 24
+ },
+ {
+ "epoch": 0.34904013961605584,
+ "grad_norm": 0.0069856527261435986,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 7.338870048522949,
+ "step": 25
+ },
+ {
+ "epoch": 0.36300174520069806,
+ "grad_norm": 0.006091665010899305,
+ "learning_rate": 0.00015,
+ "loss": 7.328121662139893,
+ "step": 26
+ },
+ {
+ "epoch": 0.3769633507853403,
+ "grad_norm": 0.005525320768356323,
+ "learning_rate": 0.000156,
+ "loss": 7.335029602050781,
+ "step": 27
+ },
+ {
+ "epoch": 0.39092495636998253,
+ "grad_norm": 0.0061133443377912045,
+ "learning_rate": 0.000162,
+ "loss": 7.319417953491211,
+ "step": 28
+ },
+ {
+ "epoch": 0.4048865619546248,
+ "grad_norm": 0.006040909793227911,
+ "learning_rate": 0.000168,
+ "loss": 7.326376438140869,
+ "step": 29
+ },
+ {
+ "epoch": 0.418848167539267,
+ "grad_norm": 0.006003216840326786,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 7.319552421569824,
+ "step": 30
+ },
+ {
+ "epoch": 0.4328097731239092,
+ "grad_norm": 0.007096003741025925,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 7.323486328125,
+ "step": 31
+ },
+ {
+ "epoch": 0.4467713787085515,
+ "grad_norm": 0.007725459989160299,
+ "learning_rate": 0.000186,
+ "loss": 7.3190717697143555,
+ "step": 32
+ },
+ {
+ "epoch": 0.4607329842931937,
+ "grad_norm": 0.007338670082390308,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 7.317154884338379,
+ "step": 33
+ },
+ {
+ "epoch": 0.47469458987783597,
+ "grad_norm": 0.007709544617682695,
+ "learning_rate": 0.000198,
+ "loss": 7.311826705932617,
+ "step": 34
+ },
+ {
+ "epoch": 0.4886561954624782,
+ "grad_norm": 0.009249784983694553,
+ "learning_rate": 0.000204,
+ "loss": 7.322498798370361,
+ "step": 35
+ },
+ {
+ "epoch": 0.5026178010471204,
+ "grad_norm": 0.006950767710804939,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 7.314571380615234,
+ "step": 36
+ },
+ {
+ "epoch": 0.5165794066317626,
+ "grad_norm": 0.007583985570818186,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 7.309540748596191,
+ "step": 37
+ },
+ {
+ "epoch": 0.5305410122164049,
+ "grad_norm": 0.014731820672750473,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 7.307004928588867,
+ "step": 38
+ },
+ {
+ "epoch": 0.5445026178010471,
+ "grad_norm": 0.016276473179459572,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 7.299642562866211,
+ "step": 39
+ },
+ {
+ "epoch": 0.5584642233856894,
+ "grad_norm": 0.020018689334392548,
+ "learning_rate": 0.000234,
+ "loss": 7.298466682434082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5724258289703316,
+ "grad_norm": 0.014941536821424961,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 7.289968967437744,
+ "step": 41
+ },
+ {
+ "epoch": 0.5863874345549738,
+ "grad_norm": 0.018299812451004982,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 7.276564598083496,
+ "step": 42
+ },
+ {
+ "epoch": 0.6003490401396161,
+ "grad_norm": 0.01615075021982193,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.2575554847717285,
+ "step": 43
+ },
+ {
+ "epoch": 0.6143106457242583,
+ "grad_norm": 0.013063831254839897,
+ "learning_rate": 0.000258,
+ "loss": 7.253885269165039,
+ "step": 44
+ },
+ {
+ "epoch": 0.6282722513089005,
+ "grad_norm": 0.015211676247417927,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.245779037475586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6422338568935427,
+ "grad_norm": 0.02093500830233097,
+ "learning_rate": 0.00027,
+ "loss": 7.225130081176758,
+ "step": 46
+ },
+ {
+ "epoch": 0.6561954624781849,
+ "grad_norm": 0.01808053068816662,
+ "learning_rate": 0.000276,
+ "loss": 7.209654331207275,
+ "step": 47
+ },
+ {
+ "epoch": 0.6701570680628273,
+ "grad_norm": 0.02209644205868244,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.197124481201172,
+ "step": 48
+ },
+ {
+ "epoch": 0.6841186736474695,
+ "grad_norm": 0.023524878546595573,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.1841936111450195,
+ "step": 49
+ },
+ {
+ "epoch": 0.6980802792321117,
+ "grad_norm": 0.031148087233304977,
+ "learning_rate": 0.000294,
+ "loss": 7.188040733337402,
+ "step": 50
+ },
+ {
+ "epoch": 0.7120418848167539,
+ "grad_norm": 0.02931929938495159,
+ "learning_rate": 0.0003,
+ "loss": 7.17371940612793,
+ "step": 51
+ },
+ {
+ "epoch": 0.7260034904013961,
+ "grad_norm": 0.01611483097076416,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.1714324951171875,
+ "step": 52
+ },
+ {
+ "epoch": 0.7399650959860384,
+ "grad_norm": 0.02619907818734646,
+ "learning_rate": 0.000312,
+ "loss": 7.158261299133301,
+ "step": 53
+ },
+ {
+ "epoch": 0.7539267015706806,
+ "grad_norm": 0.049462560564279556,
+ "learning_rate": 0.000318,
+ "loss": 7.167076110839844,
+ "step": 54
+ },
+ {
+ "epoch": 0.7678883071553229,
+ "grad_norm": 0.09835070371627808,
+ "learning_rate": 0.000324,
+ "loss": 7.154292106628418,
+ "step": 55
+ },
+ {
+ "epoch": 0.7818499127399651,
+ "grad_norm": 0.11406510323286057,
+ "learning_rate": 0.00033,
+ "loss": 7.162016868591309,
+ "step": 56
+ },
+ {
+ "epoch": 0.7958115183246073,
+ "grad_norm": 0.027026303112506866,
+ "learning_rate": 0.000336,
+ "loss": 7.141416549682617,
+ "step": 57
+ },
+ {
+ "epoch": 0.8097731239092496,
+ "grad_norm": 0.06690579652786255,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.13959264755249,
+ "step": 58
+ },
+ {
+ "epoch": 0.8237347294938918,
+ "grad_norm": 0.020919431000947952,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.126513481140137,
+ "step": 59
+ },
+ {
+ "epoch": 0.837696335078534,
+ "grad_norm": 0.05618130415678024,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.124345779418945,
+ "step": 60
+ },
+ {
+ "epoch": 0.8516579406631762,
+ "grad_norm": 0.03402302414178848,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.100615501403809,
+ "step": 61
+ },
+ {
+ "epoch": 0.8656195462478184,
+ "grad_norm": 0.03646906837821007,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.116507530212402,
+ "step": 62
+ },
+ {
+ "epoch": 0.8795811518324608,
+ "grad_norm": 0.03076397068798542,
+ "learning_rate": 0.000372,
+ "loss": 7.097659111022949,
+ "step": 63
+ },
+ {
+ "epoch": 0.893542757417103,
+ "grad_norm": 0.03329646959900856,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.097290992736816,
+ "step": 64
+ },
+ {
+ "epoch": 0.9075043630017452,
+ "grad_norm": 0.01998630352318287,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.092268943786621,
+ "step": 65
+ },
+ {
+ "epoch": 0.9214659685863874,
+ "grad_norm": 0.03278940171003342,
+ "learning_rate": 0.00039,
+ "loss": 7.088292121887207,
+ "step": 66
+ },
+ {
+ "epoch": 0.9354275741710296,
+ "grad_norm": 0.024620421230793,
+ "learning_rate": 0.000396,
+ "loss": 7.078650951385498,
+ "step": 67
+ },
+ {
+ "epoch": 0.9493891797556719,
+ "grad_norm": 0.03750383481383324,
+ "learning_rate": 0.000402,
+ "loss": 7.052778244018555,
+ "step": 68
+ },
+ {
+ "epoch": 0.9633507853403142,
+ "grad_norm": 0.07558075338602066,
+ "learning_rate": 0.000408,
+ "loss": 7.045991897583008,
+ "step": 69
+ },
+ {
+ "epoch": 0.9773123909249564,
+ "grad_norm": 0.16021482646465302,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.023152828216553,
+ "step": 70
+ },
+ {
+ "epoch": 0.9912739965095986,
+ "grad_norm": 0.24981893599033356,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.067540645599365,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.10255520790815353,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 4.425640106201172,
+ "step": 72
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.072294246363955e+16,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-72/training_args.bin b/runs/l2r90-wd0033-ssl/checkpoint-72/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..313c41733ff55284c06034a59bb885b30e40ffc8
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-72/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7ddd0bdddba68d3e19ab33477000f3fc684fe4ee8e3580eeb8c02646e5fc27ac
+size 4856
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-720/chat_template.jinja b/runs/l2r90-wd0033-ssl/checkpoint-720/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-720/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-720/config.json b/runs/l2r90-wd0033-ssl/checkpoint-720/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-720/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-720/generation_config.json b/runs/l2r90-wd0033-ssl/checkpoint-720/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-720/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-720/model.safetensors b/runs/l2r90-wd0033-ssl/checkpoint-720/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..47bcbc6095b8ad5f17b12a5642cd67f5ec700e1f
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-720/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:be7f017fc879ac1659ea67607ddc8c72e2d5fd46d060f435a4aba466b97a8b9a
+size 583356232
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-720/optimizer.pt b/runs/l2r90-wd0033-ssl/checkpoint-720/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..12fe29713b78de62c3d1347357b341f345872d57
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-720/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0c07be5921fe189acea9975b794fe86b7e6e319079f3b728c30afa1688a66e94
+size 1166825338
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-720/rng_state_0.pth b/runs/l2r90-wd0033-ssl/checkpoint-720/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..720b63ca1f4a04e71bacdf1cd0560aac038befdf
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-720/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b5f503f84a11ebab6b89c76dd5062d5bcc4482c7a8ef4ef01ca5406e03fe687e
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-720/rng_state_1.pth b/runs/l2r90-wd0033-ssl/checkpoint-720/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..faf0769a6fedd36bab92e6e1fd133a397fe314ec
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-720/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5d843339757c923787d44d858edc141226ba1a0b234f88942b96af25298311aa
+size 14512
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-720/scheduler.pt b/runs/l2r90-wd0033-ssl/checkpoint-720/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..53133f8d1ca0990636972b84579aec1212febb2a
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-720/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ea248f5cdc3b9f2d6793617acdaa11d93bab34c0a87ea94ea6aaabc148225da6
+size 1064
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-720/tokenizer.json b/runs/l2r90-wd0033-ssl/checkpoint-720/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-720/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-720/tokenizer_config.json b/runs/l2r90-wd0033-ssl/checkpoint-720/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-720/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-720/trainer_state.json b/runs/l2r90-wd0033-ssl/checkpoint-720/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..2e7a8be1bdde580c7863636617487c6fe5fd0f9e
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-720/trainer_state.json
@@ -0,0 +1,5146 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 10.0,
+ "eval_steps": 500,
+ "global_step": 720,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013961605584642234,
+ "grad_norm": 0.26733073592185974,
+ "learning_rate": 0.0,
+ "loss": 7.990724086761475,
+ "step": 1
+ },
+ {
+ "epoch": 0.027923211169284468,
+ "grad_norm": 0.26712340116500854,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 7.990110874176025,
+ "step": 2
+ },
+ {
+ "epoch": 0.041884816753926704,
+ "grad_norm": 0.2646932601928711,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 7.956935405731201,
+ "step": 3
+ },
+ {
+ "epoch": 0.055846422338568937,
+ "grad_norm": 0.24684545397758484,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 7.890539646148682,
+ "step": 4
+ },
+ {
+ "epoch": 0.06980802792321117,
+ "grad_norm": 0.21016643941402435,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 7.8146138191223145,
+ "step": 5
+ },
+ {
+ "epoch": 0.08376963350785341,
+ "grad_norm": 0.17992742359638214,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 7.724307060241699,
+ "step": 6
+ },
+ {
+ "epoch": 0.09773123909249563,
+ "grad_norm": 0.14356893301010132,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 7.645662307739258,
+ "step": 7
+ },
+ {
+ "epoch": 0.11169284467713787,
+ "grad_norm": 0.10830964893102646,
+ "learning_rate": 4.2e-05,
+ "loss": 7.582019329071045,
+ "step": 8
+ },
+ {
+ "epoch": 0.1256544502617801,
+ "grad_norm": 0.08265161514282227,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 7.5230793952941895,
+ "step": 9
+ },
+ {
+ "epoch": 0.13961605584642234,
+ "grad_norm": 0.06423710286617279,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 7.473812103271484,
+ "step": 10
+ },
+ {
+ "epoch": 0.15357766143106458,
+ "grad_norm": 0.05132246017456055,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 7.440786838531494,
+ "step": 11
+ },
+ {
+ "epoch": 0.16753926701570682,
+ "grad_norm": 0.040572404861450195,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 7.41290283203125,
+ "step": 12
+ },
+ {
+ "epoch": 0.18150087260034903,
+ "grad_norm": 0.03138705715537071,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 7.39749002456665,
+ "step": 13
+ },
+ {
+ "epoch": 0.19546247818499127,
+ "grad_norm": 0.02553008496761322,
+ "learning_rate": 7.8e-05,
+ "loss": 7.387092590332031,
+ "step": 14
+ },
+ {
+ "epoch": 0.2094240837696335,
+ "grad_norm": 0.022178977727890015,
+ "learning_rate": 8.4e-05,
+ "loss": 7.378993511199951,
+ "step": 15
+ },
+ {
+ "epoch": 0.22338568935427575,
+ "grad_norm": 0.019915293902158737,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 7.37220573425293,
+ "step": 16
+ },
+ {
+ "epoch": 0.23734729493891799,
+ "grad_norm": 0.017390629276633263,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 7.366759300231934,
+ "step": 17
+ },
+ {
+ "epoch": 0.2513089005235602,
+ "grad_norm": 0.015472771599888802,
+ "learning_rate": 0.000102,
+ "loss": 7.3652663230896,
+ "step": 18
+ },
+ {
+ "epoch": 0.26527050610820246,
+ "grad_norm": 0.013599214144051075,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 7.359714984893799,
+ "step": 19
+ },
+ {
+ "epoch": 0.2792321116928447,
+ "grad_norm": 0.012442861683666706,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 7.353302001953125,
+ "step": 20
+ },
+ {
+ "epoch": 0.2931937172774869,
+ "grad_norm": 0.010471965186297894,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 7.357417583465576,
+ "step": 21
+ },
+ {
+ "epoch": 0.30715532286212915,
+ "grad_norm": 0.009647224098443985,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 7.351763725280762,
+ "step": 22
+ },
+ {
+ "epoch": 0.32111692844677137,
+ "grad_norm": 0.008228210732340813,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 7.345434188842773,
+ "step": 23
+ },
+ {
+ "epoch": 0.33507853403141363,
+ "grad_norm": 0.007454514969140291,
+ "learning_rate": 0.000138,
+ "loss": 7.342520236968994,
+ "step": 24
+ },
+ {
+ "epoch": 0.34904013961605584,
+ "grad_norm": 0.0069856527261435986,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 7.338870048522949,
+ "step": 25
+ },
+ {
+ "epoch": 0.36300174520069806,
+ "grad_norm": 0.006091665010899305,
+ "learning_rate": 0.00015,
+ "loss": 7.328121662139893,
+ "step": 26
+ },
+ {
+ "epoch": 0.3769633507853403,
+ "grad_norm": 0.005525320768356323,
+ "learning_rate": 0.000156,
+ "loss": 7.335029602050781,
+ "step": 27
+ },
+ {
+ "epoch": 0.39092495636998253,
+ "grad_norm": 0.0061133443377912045,
+ "learning_rate": 0.000162,
+ "loss": 7.319417953491211,
+ "step": 28
+ },
+ {
+ "epoch": 0.4048865619546248,
+ "grad_norm": 0.006040909793227911,
+ "learning_rate": 0.000168,
+ "loss": 7.326376438140869,
+ "step": 29
+ },
+ {
+ "epoch": 0.418848167539267,
+ "grad_norm": 0.006003216840326786,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 7.319552421569824,
+ "step": 30
+ },
+ {
+ "epoch": 0.4328097731239092,
+ "grad_norm": 0.007096003741025925,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 7.323486328125,
+ "step": 31
+ },
+ {
+ "epoch": 0.4467713787085515,
+ "grad_norm": 0.007725459989160299,
+ "learning_rate": 0.000186,
+ "loss": 7.3190717697143555,
+ "step": 32
+ },
+ {
+ "epoch": 0.4607329842931937,
+ "grad_norm": 0.007338670082390308,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 7.317154884338379,
+ "step": 33
+ },
+ {
+ "epoch": 0.47469458987783597,
+ "grad_norm": 0.007709544617682695,
+ "learning_rate": 0.000198,
+ "loss": 7.311826705932617,
+ "step": 34
+ },
+ {
+ "epoch": 0.4886561954624782,
+ "grad_norm": 0.009249784983694553,
+ "learning_rate": 0.000204,
+ "loss": 7.322498798370361,
+ "step": 35
+ },
+ {
+ "epoch": 0.5026178010471204,
+ "grad_norm": 0.006950767710804939,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 7.314571380615234,
+ "step": 36
+ },
+ {
+ "epoch": 0.5165794066317626,
+ "grad_norm": 0.007583985570818186,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 7.309540748596191,
+ "step": 37
+ },
+ {
+ "epoch": 0.5305410122164049,
+ "grad_norm": 0.014731820672750473,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 7.307004928588867,
+ "step": 38
+ },
+ {
+ "epoch": 0.5445026178010471,
+ "grad_norm": 0.016276473179459572,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 7.299642562866211,
+ "step": 39
+ },
+ {
+ "epoch": 0.5584642233856894,
+ "grad_norm": 0.020018689334392548,
+ "learning_rate": 0.000234,
+ "loss": 7.298466682434082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5724258289703316,
+ "grad_norm": 0.014941536821424961,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 7.289968967437744,
+ "step": 41
+ },
+ {
+ "epoch": 0.5863874345549738,
+ "grad_norm": 0.018299812451004982,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 7.276564598083496,
+ "step": 42
+ },
+ {
+ "epoch": 0.6003490401396161,
+ "grad_norm": 0.01615075021982193,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 7.2575554847717285,
+ "step": 43
+ },
+ {
+ "epoch": 0.6143106457242583,
+ "grad_norm": 0.013063831254839897,
+ "learning_rate": 0.000258,
+ "loss": 7.253885269165039,
+ "step": 44
+ },
+ {
+ "epoch": 0.6282722513089005,
+ "grad_norm": 0.015211676247417927,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 7.245779037475586,
+ "step": 45
+ },
+ {
+ "epoch": 0.6422338568935427,
+ "grad_norm": 0.02093500830233097,
+ "learning_rate": 0.00027,
+ "loss": 7.225130081176758,
+ "step": 46
+ },
+ {
+ "epoch": 0.6561954624781849,
+ "grad_norm": 0.01808053068816662,
+ "learning_rate": 0.000276,
+ "loss": 7.209654331207275,
+ "step": 47
+ },
+ {
+ "epoch": 0.6701570680628273,
+ "grad_norm": 0.02209644205868244,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 7.197124481201172,
+ "step": 48
+ },
+ {
+ "epoch": 0.6841186736474695,
+ "grad_norm": 0.023524878546595573,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 7.1841936111450195,
+ "step": 49
+ },
+ {
+ "epoch": 0.6980802792321117,
+ "grad_norm": 0.031148087233304977,
+ "learning_rate": 0.000294,
+ "loss": 7.188040733337402,
+ "step": 50
+ },
+ {
+ "epoch": 0.7120418848167539,
+ "grad_norm": 0.02931929938495159,
+ "learning_rate": 0.0003,
+ "loss": 7.17371940612793,
+ "step": 51
+ },
+ {
+ "epoch": 0.7260034904013961,
+ "grad_norm": 0.01611483097076416,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.1714324951171875,
+ "step": 52
+ },
+ {
+ "epoch": 0.7399650959860384,
+ "grad_norm": 0.02619907818734646,
+ "learning_rate": 0.000312,
+ "loss": 7.158261299133301,
+ "step": 53
+ },
+ {
+ "epoch": 0.7539267015706806,
+ "grad_norm": 0.049462560564279556,
+ "learning_rate": 0.000318,
+ "loss": 7.167076110839844,
+ "step": 54
+ },
+ {
+ "epoch": 0.7678883071553229,
+ "grad_norm": 0.09835070371627808,
+ "learning_rate": 0.000324,
+ "loss": 7.154292106628418,
+ "step": 55
+ },
+ {
+ "epoch": 0.7818499127399651,
+ "grad_norm": 0.11406510323286057,
+ "learning_rate": 0.00033,
+ "loss": 7.162016868591309,
+ "step": 56
+ },
+ {
+ "epoch": 0.7958115183246073,
+ "grad_norm": 0.027026303112506866,
+ "learning_rate": 0.000336,
+ "loss": 7.141416549682617,
+ "step": 57
+ },
+ {
+ "epoch": 0.8097731239092496,
+ "grad_norm": 0.06690579652786255,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.13959264755249,
+ "step": 58
+ },
+ {
+ "epoch": 0.8237347294938918,
+ "grad_norm": 0.020919431000947952,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.126513481140137,
+ "step": 59
+ },
+ {
+ "epoch": 0.837696335078534,
+ "grad_norm": 0.05618130415678024,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.124345779418945,
+ "step": 60
+ },
+ {
+ "epoch": 0.8516579406631762,
+ "grad_norm": 0.03402302414178848,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.100615501403809,
+ "step": 61
+ },
+ {
+ "epoch": 0.8656195462478184,
+ "grad_norm": 0.03646906837821007,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.116507530212402,
+ "step": 62
+ },
+ {
+ "epoch": 0.8795811518324608,
+ "grad_norm": 0.03076397068798542,
+ "learning_rate": 0.000372,
+ "loss": 7.097659111022949,
+ "step": 63
+ },
+ {
+ "epoch": 0.893542757417103,
+ "grad_norm": 0.03329646959900856,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.097290992736816,
+ "step": 64
+ },
+ {
+ "epoch": 0.9075043630017452,
+ "grad_norm": 0.01998630352318287,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.092268943786621,
+ "step": 65
+ },
+ {
+ "epoch": 0.9214659685863874,
+ "grad_norm": 0.03278940171003342,
+ "learning_rate": 0.00039,
+ "loss": 7.088292121887207,
+ "step": 66
+ },
+ {
+ "epoch": 0.9354275741710296,
+ "grad_norm": 0.024620421230793,
+ "learning_rate": 0.000396,
+ "loss": 7.078650951385498,
+ "step": 67
+ },
+ {
+ "epoch": 0.9493891797556719,
+ "grad_norm": 0.03750383481383324,
+ "learning_rate": 0.000402,
+ "loss": 7.052778244018555,
+ "step": 68
+ },
+ {
+ "epoch": 0.9633507853403142,
+ "grad_norm": 0.07558075338602066,
+ "learning_rate": 0.000408,
+ "loss": 7.045991897583008,
+ "step": 69
+ },
+ {
+ "epoch": 0.9773123909249564,
+ "grad_norm": 0.16021482646465302,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.023152828216553,
+ "step": 70
+ },
+ {
+ "epoch": 0.9912739965095986,
+ "grad_norm": 0.24981893599033356,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.067540645599365,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.10255520790815353,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 4.425640106201172,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8861758708953857,
+ "eval_runtime": 58.2699,
+ "eval_samples_per_second": 41.908,
+ "eval_steps_per_second": 0.669,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139616055846423,
+ "grad_norm": 0.13374797999858856,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.073373317718506,
+ "step": 73
+ },
+ {
+ "epoch": 1.0279232111692844,
+ "grad_norm": 0.05079387128353119,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.0243306159973145,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418848167539267,
+ "grad_norm": 0.10056068748235703,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.014476299285889,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558464223385688,
+ "grad_norm": 0.06995758414268494,
+ "learning_rate": 0.00045,
+ "loss": 7.011931419372559,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698080279232112,
+ "grad_norm": 0.0483892522752285,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 6.9822797775268555,
+ "step": 77
+ },
+ {
+ "epoch": 1.0837696335078535,
+ "grad_norm": 0.08408170193433762,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 6.964348793029785,
+ "step": 78
+ },
+ {
+ "epoch": 1.0977312390924956,
+ "grad_norm": 0.08027710020542145,
+ "learning_rate": 0.000468,
+ "loss": 6.96852445602417,
+ "step": 79
+ },
+ {
+ "epoch": 1.111692844677138,
+ "grad_norm": 0.05647159367799759,
+ "learning_rate": 0.000474,
+ "loss": 6.944484710693359,
+ "step": 80
+ },
+ {
+ "epoch": 1.12565445026178,
+ "grad_norm": 0.07464330643415451,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 6.9246745109558105,
+ "step": 81
+ },
+ {
+ "epoch": 1.1396160558464223,
+ "grad_norm": 0.04451674595475197,
+ "learning_rate": 0.000486,
+ "loss": 6.895403861999512,
+ "step": 82
+ },
+ {
+ "epoch": 1.1535776614310647,
+ "grad_norm": 0.05820256471633911,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 6.887338638305664,
+ "step": 83
+ },
+ {
+ "epoch": 1.1675392670157068,
+ "grad_norm": 0.02925296127796173,
+ "learning_rate": 0.000498,
+ "loss": 6.857797622680664,
+ "step": 84
+ },
+ {
+ "epoch": 1.181500872600349,
+ "grad_norm": 0.05096987634897232,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 6.854822158813477,
+ "step": 85
+ },
+ {
+ "epoch": 1.1954624781849912,
+ "grad_norm": 0.03483197093009949,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 6.840914726257324,
+ "step": 86
+ },
+ {
+ "epoch": 1.2094240837696335,
+ "grad_norm": 0.03323264792561531,
+ "learning_rate": 0.000516,
+ "loss": 6.834497451782227,
+ "step": 87
+ },
+ {
+ "epoch": 1.2233856893542758,
+ "grad_norm": 0.03290868178009987,
+ "learning_rate": 0.000522,
+ "loss": 6.820746898651123,
+ "step": 88
+ },
+ {
+ "epoch": 1.237347294938918,
+ "grad_norm": 0.025398844853043556,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 6.795746803283691,
+ "step": 89
+ },
+ {
+ "epoch": 1.2513089005235603,
+ "grad_norm": 0.030700596049427986,
+ "learning_rate": 0.000534,
+ "loss": 6.7867889404296875,
+ "step": 90
+ },
+ {
+ "epoch": 1.2652705061082026,
+ "grad_norm": 0.019589420408010483,
+ "learning_rate": 0.00054,
+ "loss": 6.77443790435791,
+ "step": 91
+ },
+ {
+ "epoch": 1.2792321116928447,
+ "grad_norm": 0.03313920646905899,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 6.762271881103516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2931937172774868,
+ "grad_norm": 0.025389468297362328,
+ "learning_rate": 0.000552,
+ "loss": 6.750258445739746,
+ "step": 93
+ },
+ {
+ "epoch": 1.307155322862129,
+ "grad_norm": 0.040296655148267746,
+ "learning_rate": 0.000558,
+ "loss": 6.729808807373047,
+ "step": 94
+ },
+ {
+ "epoch": 1.3211169284467714,
+ "grad_norm": 0.04735933244228363,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 6.737982749938965,
+ "step": 95
+ },
+ {
+ "epoch": 1.3350785340314135,
+ "grad_norm": 0.05529680848121643,
+ "learning_rate": 0.00057,
+ "loss": 6.735714912414551,
+ "step": 96
+ },
+ {
+ "epoch": 1.3490401396160558,
+ "grad_norm": 0.07148554176092148,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 6.708859920501709,
+ "step": 97
+ },
+ {
+ "epoch": 1.3630017452006982,
+ "grad_norm": 0.059708479791879654,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 6.708860397338867,
+ "step": 98
+ },
+ {
+ "epoch": 1.3769633507853403,
+ "grad_norm": 0.027111195027828217,
+ "learning_rate": 0.000588,
+ "loss": 6.686685085296631,
+ "step": 99
+ },
+ {
+ "epoch": 1.3909249563699826,
+ "grad_norm": 0.0500967763364315,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 6.676325798034668,
+ "step": 100
+ },
+ {
+ "epoch": 1.404886561954625,
+ "grad_norm": 0.044565796852111816,
+ "learning_rate": 0.0006,
+ "loss": 6.674530029296875,
+ "step": 101
+ },
+ {
+ "epoch": 1.418848167539267,
+ "grad_norm": 0.03608180209994316,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 6.653741359710693,
+ "step": 102
+ },
+ {
+ "epoch": 1.4328097731239091,
+ "grad_norm": 0.030394982546567917,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 6.643152236938477,
+ "step": 103
+ },
+ {
+ "epoch": 1.4467713787085514,
+ "grad_norm": 0.040185656398534775,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 6.628686904907227,
+ "step": 104
+ },
+ {
+ "epoch": 1.4607329842931938,
+ "grad_norm": 0.021049322560429573,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 6.610864639282227,
+ "step": 105
+ },
+ {
+ "epoch": 1.4746945898778359,
+ "grad_norm": 0.04164277762174606,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 6.602880954742432,
+ "step": 106
+ },
+ {
+ "epoch": 1.4886561954624782,
+ "grad_norm": 0.0389864556491375,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 6.599456787109375,
+ "step": 107
+ },
+ {
+ "epoch": 1.5026178010471205,
+ "grad_norm": 0.05805973336100578,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 6.588700294494629,
+ "step": 108
+ },
+ {
+ "epoch": 1.5165794066317626,
+ "grad_norm": 0.08825036883354187,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 6.587496757507324,
+ "step": 109
+ },
+ {
+ "epoch": 1.530541012216405,
+ "grad_norm": 0.12521523237228394,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 6.626501083374023,
+ "step": 110
+ },
+ {
+ "epoch": 1.5445026178010473,
+ "grad_norm": 0.07429645210504532,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 6.590241432189941,
+ "step": 111
+ },
+ {
+ "epoch": 1.5584642233856894,
+ "grad_norm": 0.06463009119033813,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 6.578714370727539,
+ "step": 112
+ },
+ {
+ "epoch": 1.5724258289703315,
+ "grad_norm": 0.07508203387260437,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 6.561270713806152,
+ "step": 113
+ },
+ {
+ "epoch": 1.5863874345549738,
+ "grad_norm": 0.04125197231769562,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 6.558821678161621,
+ "step": 114
+ },
+ {
+ "epoch": 1.600349040139616,
+ "grad_norm": 0.05457386001944542,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.541611671447754,
+ "step": 115
+ },
+ {
+ "epoch": 1.6143106457242582,
+ "grad_norm": 0.03684385493397713,
+ "learning_rate": 0.000599973080557108,
+ "loss": 6.547738075256348,
+ "step": 116
+ },
+ {
+ "epoch": 1.6282722513089005,
+ "grad_norm": 0.05888746306300163,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.529491424560547,
+ "step": 117
+ },
+ {
+ "epoch": 1.6422338568935428,
+ "grad_norm": 0.056668370962142944,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.519444942474365,
+ "step": 118
+ },
+ {
+ "epoch": 1.656195462478185,
+ "grad_norm": 0.09123896062374115,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.522984981536865,
+ "step": 119
+ },
+ {
+ "epoch": 1.6701570680628273,
+ "grad_norm": 0.12380023300647736,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.515375137329102,
+ "step": 120
+ },
+ {
+ "epoch": 1.6841186736474696,
+ "grad_norm": 0.06953322887420654,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.500491142272949,
+ "step": 121
+ },
+ {
+ "epoch": 1.6980802792321117,
+ "grad_norm": 0.08613869547843933,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.5097198486328125,
+ "step": 122
+ },
+ {
+ "epoch": 1.7120418848167538,
+ "grad_norm": 0.08296383172273636,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.487521171569824,
+ "step": 123
+ },
+ {
+ "epoch": 1.7260034904013961,
+ "grad_norm": 0.05744783207774162,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.485134124755859,
+ "step": 124
+ },
+ {
+ "epoch": 1.7399650959860384,
+ "grad_norm": 0.04108566418290138,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.466182708740234,
+ "step": 125
+ },
+ {
+ "epoch": 1.7539267015706805,
+ "grad_norm": 0.05374931916594505,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.479011535644531,
+ "step": 126
+ },
+ {
+ "epoch": 1.7678883071553229,
+ "grad_norm": 0.03867914527654648,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.465902328491211,
+ "step": 127
+ },
+ {
+ "epoch": 1.7818499127399652,
+ "grad_norm": 0.05477374792098999,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.453082084655762,
+ "step": 128
+ },
+ {
+ "epoch": 1.7958115183246073,
+ "grad_norm": 0.06508298963308334,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.446089744567871,
+ "step": 129
+ },
+ {
+ "epoch": 1.8097731239092496,
+ "grad_norm": 0.06388624757528305,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.451748847961426,
+ "step": 130
+ },
+ {
+ "epoch": 1.823734729493892,
+ "grad_norm": 0.04223188757896423,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.4492387771606445,
+ "step": 131
+ },
+ {
+ "epoch": 1.837696335078534,
+ "grad_norm": 0.04528658092021942,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.409390926361084,
+ "step": 132
+ },
+ {
+ "epoch": 1.8516579406631761,
+ "grad_norm": 0.04962967336177826,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.415774345397949,
+ "step": 133
+ },
+ {
+ "epoch": 1.8656195462478184,
+ "grad_norm": 0.031944390386343,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.411528587341309,
+ "step": 134
+ },
+ {
+ "epoch": 1.8795811518324608,
+ "grad_norm": 0.028411149978637695,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.392329216003418,
+ "step": 135
+ },
+ {
+ "epoch": 1.8935427574171029,
+ "grad_norm": 0.03516148403286934,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.392238616943359,
+ "step": 136
+ },
+ {
+ "epoch": 1.9075043630017452,
+ "grad_norm": 0.027800120413303375,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.39179801940918,
+ "step": 137
+ },
+ {
+ "epoch": 1.9214659685863875,
+ "grad_norm": 0.02362987957894802,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.378960609436035,
+ "step": 138
+ },
+ {
+ "epoch": 1.9354275741710296,
+ "grad_norm": 0.033999308943748474,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.36802339553833,
+ "step": 139
+ },
+ {
+ "epoch": 1.949389179755672,
+ "grad_norm": 0.03290879726409912,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.366308689117432,
+ "step": 140
+ },
+ {
+ "epoch": 1.9633507853403143,
+ "grad_norm": 0.0286843404173851,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.350399971008301,
+ "step": 141
+ },
+ {
+ "epoch": 1.9773123909249564,
+ "grad_norm": 0.027145877480506897,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.33477783203125,
+ "step": 142
+ },
+ {
+ "epoch": 1.9912739965095985,
+ "grad_norm": 0.03530878946185112,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.322859764099121,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.028133150190114975,
+ "learning_rate": 0.000599778806120077,
+ "loss": 3.955288887023926,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7927101254463196,
+ "eval_runtime": 60.8372,
+ "eval_samples_per_second": 40.14,
+ "eval_steps_per_second": 0.641,
+ "step": 144
+ },
+ {
+ "epoch": 2.013961605584642,
+ "grad_norm": 0.03054557554423809,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.311975479125977,
+ "step": 145
+ },
+ {
+ "epoch": 2.0279232111692846,
+ "grad_norm": 0.05406459793448448,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.3030853271484375,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418848167539267,
+ "grad_norm": 0.10911019891500473,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.329056739807129,
+ "step": 147
+ },
+ {
+ "epoch": 2.055846422338569,
+ "grad_norm": 0.13948315382003784,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.362421035766602,
+ "step": 148
+ },
+ {
+ "epoch": 2.0698080279232114,
+ "grad_norm": 0.13176991045475006,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.3888139724731445,
+ "step": 149
+ },
+ {
+ "epoch": 2.0837696335078535,
+ "grad_norm": 0.09615885466337204,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.35154914855957,
+ "step": 150
+ },
+ {
+ "epoch": 2.0977312390924956,
+ "grad_norm": 0.06462561339139938,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.335790634155273,
+ "step": 151
+ },
+ {
+ "epoch": 2.1116928446771377,
+ "grad_norm": 0.05186621472239494,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.324396133422852,
+ "step": 152
+ },
+ {
+ "epoch": 2.1256544502617802,
+ "grad_norm": 0.06095981225371361,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.318549156188965,
+ "step": 153
+ },
+ {
+ "epoch": 2.1396160558464223,
+ "grad_norm": 0.036121442914009094,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.300607681274414,
+ "step": 154
+ },
+ {
+ "epoch": 2.1535776614310644,
+ "grad_norm": 0.055716633796691895,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.307400703430176,
+ "step": 155
+ },
+ {
+ "epoch": 2.167539267015707,
+ "grad_norm": 0.037626396864652634,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.287364959716797,
+ "step": 156
+ },
+ {
+ "epoch": 2.181500872600349,
+ "grad_norm": 0.04329098388552666,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.28244686126709,
+ "step": 157
+ },
+ {
+ "epoch": 2.195462478184991,
+ "grad_norm": 0.031868450343608856,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.2602410316467285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2094240837696333,
+ "grad_norm": 0.03647344186902046,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.263988971710205,
+ "step": 159
+ },
+ {
+ "epoch": 2.223385689354276,
+ "grad_norm": 0.03141765296459198,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.263205051422119,
+ "step": 160
+ },
+ {
+ "epoch": 2.237347294938918,
+ "grad_norm": 0.028253229334950447,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.253107070922852,
+ "step": 161
+ },
+ {
+ "epoch": 2.25130890052356,
+ "grad_norm": 0.03608971834182739,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.2571702003479,
+ "step": 162
+ },
+ {
+ "epoch": 2.2652705061082026,
+ "grad_norm": 0.02644292451441288,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.241732597351074,
+ "step": 163
+ },
+ {
+ "epoch": 2.2792321116928447,
+ "grad_norm": 0.028604455292224884,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.230289459228516,
+ "step": 164
+ },
+ {
+ "epoch": 2.2931937172774868,
+ "grad_norm": 0.023704614490270615,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.226101398468018,
+ "step": 165
+ },
+ {
+ "epoch": 2.3071553228621293,
+ "grad_norm": 0.020691027864813805,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.2096405029296875,
+ "step": 166
+ },
+ {
+ "epoch": 2.3211169284467714,
+ "grad_norm": 0.025598427280783653,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.22075080871582,
+ "step": 167
+ },
+ {
+ "epoch": 2.3350785340314135,
+ "grad_norm": 0.030546002089977264,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.216886520385742,
+ "step": 168
+ },
+ {
+ "epoch": 2.349040139616056,
+ "grad_norm": 0.026066245511174202,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.2022504806518555,
+ "step": 169
+ },
+ {
+ "epoch": 2.363001745200698,
+ "grad_norm": 0.023802239447832108,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.189509391784668,
+ "step": 170
+ },
+ {
+ "epoch": 2.3769633507853403,
+ "grad_norm": 0.029805002734065056,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.186277389526367,
+ "step": 171
+ },
+ {
+ "epoch": 2.3909249563699824,
+ "grad_norm": 0.043327830731868744,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.164876937866211,
+ "step": 172
+ },
+ {
+ "epoch": 2.404886561954625,
+ "grad_norm": 0.058787211775779724,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.179936408996582,
+ "step": 173
+ },
+ {
+ "epoch": 2.418848167539267,
+ "grad_norm": 0.0787397176027298,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.1830244064331055,
+ "step": 174
+ },
+ {
+ "epoch": 2.432809773123909,
+ "grad_norm": 0.0899854376912117,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.190343856811523,
+ "step": 175
+ },
+ {
+ "epoch": 2.4467713787085517,
+ "grad_norm": 0.06520707905292511,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.185057163238525,
+ "step": 176
+ },
+ {
+ "epoch": 2.4607329842931938,
+ "grad_norm": 0.1002504974603653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.171279430389404,
+ "step": 177
+ },
+ {
+ "epoch": 2.474694589877836,
+ "grad_norm": 0.09591004997491837,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.174382209777832,
+ "step": 178
+ },
+ {
+ "epoch": 2.488656195462478,
+ "grad_norm": 0.12127121537923813,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.185717582702637,
+ "step": 179
+ },
+ {
+ "epoch": 2.5026178010471205,
+ "grad_norm": 0.09640590846538544,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.196681976318359,
+ "step": 180
+ },
+ {
+ "epoch": 2.5165794066317626,
+ "grad_norm": 0.048341453075408936,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.158774375915527,
+ "step": 181
+ },
+ {
+ "epoch": 2.530541012216405,
+ "grad_norm": 0.08206334710121155,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.17445707321167,
+ "step": 182
+ },
+ {
+ "epoch": 2.5445026178010473,
+ "grad_norm": 0.038934968411922455,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.170527458190918,
+ "step": 183
+ },
+ {
+ "epoch": 2.5584642233856894,
+ "grad_norm": 0.060987457633018494,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.1501994132995605,
+ "step": 184
+ },
+ {
+ "epoch": 2.5724258289703315,
+ "grad_norm": 0.033269546926021576,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.144186019897461,
+ "step": 185
+ },
+ {
+ "epoch": 2.5863874345549736,
+ "grad_norm": 0.046137236058712006,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.1236677169799805,
+ "step": 186
+ },
+ {
+ "epoch": 2.600349040139616,
+ "grad_norm": 0.038265109062194824,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.121979236602783,
+ "step": 187
+ },
+ {
+ "epoch": 2.614310645724258,
+ "grad_norm": 0.039128001779317856,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.124156951904297,
+ "step": 188
+ },
+ {
+ "epoch": 2.6282722513089007,
+ "grad_norm": 0.025303540751338005,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.108036994934082,
+ "step": 189
+ },
+ {
+ "epoch": 2.642233856893543,
+ "grad_norm": 0.04441308230161667,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.1211161613464355,
+ "step": 190
+ },
+ {
+ "epoch": 2.656195462478185,
+ "grad_norm": 0.03083532303571701,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.094561576843262,
+ "step": 191
+ },
+ {
+ "epoch": 2.670157068062827,
+ "grad_norm": 0.02951296605169773,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.091608047485352,
+ "step": 192
+ },
+ {
+ "epoch": 2.6841186736474696,
+ "grad_norm": 0.02385987527668476,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.087080001831055,
+ "step": 193
+ },
+ {
+ "epoch": 2.6980802792321117,
+ "grad_norm": 0.03315780311822891,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.088132381439209,
+ "step": 194
+ },
+ {
+ "epoch": 2.712041884816754,
+ "grad_norm": 0.02410835586488247,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.081460952758789,
+ "step": 195
+ },
+ {
+ "epoch": 2.7260034904013963,
+ "grad_norm": 0.033069487661123276,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.0696539878845215,
+ "step": 196
+ },
+ {
+ "epoch": 2.7399650959860384,
+ "grad_norm": 0.031288471072912216,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.047612190246582,
+ "step": 197
+ },
+ {
+ "epoch": 2.7539267015706805,
+ "grad_norm": 0.03399677947163582,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.019026756286621,
+ "step": 198
+ },
+ {
+ "epoch": 2.7678883071553226,
+ "grad_norm": 0.031948331743478775,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.050133228302002,
+ "step": 199
+ },
+ {
+ "epoch": 2.781849912739965,
+ "grad_norm": 0.049206435680389404,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.045332431793213,
+ "step": 200
+ },
+ {
+ "epoch": 2.7958115183246073,
+ "grad_norm": 0.07528849691152573,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.059202194213867,
+ "step": 201
+ },
+ {
+ "epoch": 2.80977312390925,
+ "grad_norm": 0.11815470457077026,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.077486038208008,
+ "step": 202
+ },
+ {
+ "epoch": 2.823734729493892,
+ "grad_norm": 0.0867348462343216,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.045670032501221,
+ "step": 203
+ },
+ {
+ "epoch": 2.837696335078534,
+ "grad_norm": 0.05716705322265625,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.034373760223389,
+ "step": 204
+ },
+ {
+ "epoch": 2.851657940663176,
+ "grad_norm": 0.05944284051656723,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.049911975860596,
+ "step": 205
+ },
+ {
+ "epoch": 2.8656195462478182,
+ "grad_norm": 0.03220470994710922,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.039861679077148,
+ "step": 206
+ },
+ {
+ "epoch": 2.8795811518324608,
+ "grad_norm": 0.057081177830696106,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.010714530944824,
+ "step": 207
+ },
+ {
+ "epoch": 2.893542757417103,
+ "grad_norm": 0.04037516191601753,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.018877983093262,
+ "step": 208
+ },
+ {
+ "epoch": 2.9075043630017454,
+ "grad_norm": 0.036492880433797836,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.009493827819824,
+ "step": 209
+ },
+ {
+ "epoch": 2.9214659685863875,
+ "grad_norm": 0.04785023257136345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.0140862464904785,
+ "step": 210
+ },
+ {
+ "epoch": 2.9354275741710296,
+ "grad_norm": 0.07721036672592163,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.030359268188477,
+ "step": 211
+ },
+ {
+ "epoch": 2.9493891797556717,
+ "grad_norm": 0.14220190048217773,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.045989990234375,
+ "step": 212
+ },
+ {
+ "epoch": 2.9633507853403143,
+ "grad_norm": 0.10265158116817474,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.043968200683594,
+ "step": 213
+ },
+ {
+ "epoch": 2.9773123909249564,
+ "grad_norm": 0.051398951560258865,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.022278785705566,
+ "step": 214
+ },
+ {
+ "epoch": 2.9912739965095985,
+ "grad_norm": 0.06044618785381317,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.017296314239502,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.03284750506281853,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 3.7544798851013184,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7531244158744812,
+ "eval_runtime": 60.3986,
+ "eval_samples_per_second": 40.431,
+ "eval_steps_per_second": 0.646,
+ "step": 216
+ },
+ {
+ "epoch": 3.013961605584642,
+ "grad_norm": 0.03895840048789978,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 5.9985857009887695,
+ "step": 217
+ },
+ {
+ "epoch": 3.0279232111692846,
+ "grad_norm": 0.04554992541670799,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 5.984710693359375,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418848167539267,
+ "grad_norm": 0.03359116613864899,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 5.984078407287598,
+ "step": 219
+ },
+ {
+ "epoch": 3.055846422338569,
+ "grad_norm": 0.035171881318092346,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 5.979424476623535,
+ "step": 220
+ },
+ {
+ "epoch": 3.0698080279232114,
+ "grad_norm": 0.032774072140455246,
+ "learning_rate": 0.000598278794687186,
+ "loss": 5.94467830657959,
+ "step": 221
+ },
+ {
+ "epoch": 3.0837696335078535,
+ "grad_norm": 0.032842908054590225,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 5.95986270904541,
+ "step": 222
+ },
+ {
+ "epoch": 3.0977312390924956,
+ "grad_norm": 0.02870488539338112,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 5.951186656951904,
+ "step": 223
+ },
+ {
+ "epoch": 3.1116928446771377,
+ "grad_norm": 0.030777374282479286,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 5.942493438720703,
+ "step": 224
+ },
+ {
+ "epoch": 3.1256544502617802,
+ "grad_norm": 0.026439018547534943,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 5.938275337219238,
+ "step": 225
+ },
+ {
+ "epoch": 3.1396160558464223,
+ "grad_norm": 0.028689472004771233,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 5.931982040405273,
+ "step": 226
+ },
+ {
+ "epoch": 3.1535776614310644,
+ "grad_norm": 0.028100362047553062,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 5.919692039489746,
+ "step": 227
+ },
+ {
+ "epoch": 3.167539267015707,
+ "grad_norm": 0.02407177910208702,
+ "learning_rate": 0.00059807235435179,
+ "loss": 5.909029006958008,
+ "step": 228
+ },
+ {
+ "epoch": 3.181500872600349,
+ "grad_norm": 0.0237748846411705,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 5.9131598472595215,
+ "step": 229
+ },
+ {
+ "epoch": 3.195462478184991,
+ "grad_norm": 0.022418556734919548,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 5.909396171569824,
+ "step": 230
+ },
+ {
+ "epoch": 3.2094240837696333,
+ "grad_norm": 0.024656571447849274,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 5.885243892669678,
+ "step": 231
+ },
+ {
+ "epoch": 3.223385689354276,
+ "grad_norm": 0.02136118710041046,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 5.898958206176758,
+ "step": 232
+ },
+ {
+ "epoch": 3.237347294938918,
+ "grad_norm": 0.025290116667747498,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 5.886986255645752,
+ "step": 233
+ },
+ {
+ "epoch": 3.25130890052356,
+ "grad_norm": 0.02656404674053192,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 5.897647857666016,
+ "step": 234
+ },
+ {
+ "epoch": 3.2652705061082026,
+ "grad_norm": 0.019188793376088142,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 5.875659942626953,
+ "step": 235
+ },
+ {
+ "epoch": 3.2792321116928447,
+ "grad_norm": 0.02308926172554493,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 5.859380722045898,
+ "step": 236
+ },
+ {
+ "epoch": 3.2931937172774868,
+ "grad_norm": 0.029775725677609444,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 5.873961448669434,
+ "step": 237
+ },
+ {
+ "epoch": 3.3071553228621293,
+ "grad_norm": 0.03875259310007095,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 5.863171100616455,
+ "step": 238
+ },
+ {
+ "epoch": 3.3211169284467714,
+ "grad_norm": 0.052446842193603516,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 5.886547088623047,
+ "step": 239
+ },
+ {
+ "epoch": 3.3350785340314135,
+ "grad_norm": 0.05583837628364563,
+ "learning_rate": 0.000597691359743243,
+ "loss": 5.8849287033081055,
+ "step": 240
+ },
+ {
+ "epoch": 3.349040139616056,
+ "grad_norm": 0.034729067236185074,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 5.863419532775879,
+ "step": 241
+ },
+ {
+ "epoch": 3.363001745200698,
+ "grad_norm": 0.03590013459324837,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 5.852169990539551,
+ "step": 242
+ },
+ {
+ "epoch": 3.3769633507853403,
+ "grad_norm": 0.035041507333517075,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 5.836690902709961,
+ "step": 243
+ },
+ {
+ "epoch": 3.3909249563699824,
+ "grad_norm": 0.02857939340174198,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 5.836485862731934,
+ "step": 244
+ },
+ {
+ "epoch": 3.404886561954625,
+ "grad_norm": 0.03176642581820488,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 5.860719203948975,
+ "step": 245
+ },
+ {
+ "epoch": 3.418848167539267,
+ "grad_norm": 0.026105748489499092,
+ "learning_rate": 0.000597488038559129,
+ "loss": 5.8354949951171875,
+ "step": 246
+ },
+ {
+ "epoch": 3.432809773123909,
+ "grad_norm": 0.029094981029629707,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 5.846247673034668,
+ "step": 247
+ },
+ {
+ "epoch": 3.4467713787085517,
+ "grad_norm": 0.02388785220682621,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 5.837607383728027,
+ "step": 248
+ },
+ {
+ "epoch": 3.4607329842931938,
+ "grad_norm": 0.03094472736120224,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 5.832090377807617,
+ "step": 249
+ },
+ {
+ "epoch": 3.474694589877836,
+ "grad_norm": 0.03227658197283745,
+ "learning_rate": 0.000597347745466109,
+ "loss": 5.83722448348999,
+ "step": 250
+ },
+ {
+ "epoch": 3.488656195462478,
+ "grad_norm": 0.042813923209905624,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 5.851722717285156,
+ "step": 251
+ },
+ {
+ "epoch": 3.5026178010471205,
+ "grad_norm": 0.05825525149703026,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.795719146728516,
+ "step": 252
+ },
+ {
+ "epoch": 3.5165794066317626,
+ "grad_norm": 0.0867447778582573,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.813750267028809,
+ "step": 253
+ },
+ {
+ "epoch": 3.530541012216405,
+ "grad_norm": 0.09192710369825363,
+ "learning_rate": 0.000597203657966319,
+ "loss": 5.823488235473633,
+ "step": 254
+ },
+ {
+ "epoch": 3.5445026178010473,
+ "grad_norm": 0.06069248169660568,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 5.814342021942139,
+ "step": 255
+ },
+ {
+ "epoch": 3.5584642233856894,
+ "grad_norm": 0.06259952485561371,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 5.834272384643555,
+ "step": 256
+ },
+ {
+ "epoch": 3.5724258289703315,
+ "grad_norm": 0.0513639897108078,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.807951927185059,
+ "step": 257
+ },
+ {
+ "epoch": 3.5863874345549736,
+ "grad_norm": 0.03327798470854759,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.790238857269287,
+ "step": 258
+ },
+ {
+ "epoch": 3.600349040139616,
+ "grad_norm": 0.03963128477334976,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.7956695556640625,
+ "step": 259
+ },
+ {
+ "epoch": 3.614310645724258,
+ "grad_norm": 0.031153453513979912,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 5.7874531745910645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6282722513089007,
+ "grad_norm": 0.031419433653354645,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 5.779650688171387,
+ "step": 261
+ },
+ {
+ "epoch": 3.642233856893543,
+ "grad_norm": 0.0288684144616127,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.776577949523926,
+ "step": 262
+ },
+ {
+ "epoch": 3.656195462478185,
+ "grad_norm": 0.028315918520092964,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.778571128845215,
+ "step": 263
+ },
+ {
+ "epoch": 3.670157068062827,
+ "grad_norm": 0.024977795779705048,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 5.776470184326172,
+ "step": 264
+ },
+ {
+ "epoch": 3.6841186736474696,
+ "grad_norm": 0.026822632178664207,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.755108833312988,
+ "step": 265
+ },
+ {
+ "epoch": 3.6980802792321117,
+ "grad_norm": 0.02738630585372448,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.74041748046875,
+ "step": 266
+ },
+ {
+ "epoch": 3.712041884816754,
+ "grad_norm": 0.02361280657351017,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.764027118682861,
+ "step": 267
+ },
+ {
+ "epoch": 3.7260034904013963,
+ "grad_norm": 0.029926009476184845,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.780390739440918,
+ "step": 268
+ },
+ {
+ "epoch": 3.7399650959860384,
+ "grad_norm": 0.03519507870078087,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.758105278015137,
+ "step": 269
+ },
+ {
+ "epoch": 3.7539267015706805,
+ "grad_norm": 0.030114633962512016,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.757230758666992,
+ "step": 270
+ },
+ {
+ "epoch": 3.7678883071553226,
+ "grad_norm": 0.02467748522758484,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.7615461349487305,
+ "step": 271
+ },
+ {
+ "epoch": 3.781849912739965,
+ "grad_norm": 0.03425183147192001,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.749448776245117,
+ "step": 272
+ },
+ {
+ "epoch": 3.7958115183246073,
+ "grad_norm": 0.046307582408189774,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.7356977462768555,
+ "step": 273
+ },
+ {
+ "epoch": 3.80977312390925,
+ "grad_norm": 0.054102856665849686,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.741869926452637,
+ "step": 274
+ },
+ {
+ "epoch": 3.823734729493892,
+ "grad_norm": 0.07026689499616623,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.721103668212891,
+ "step": 275
+ },
+ {
+ "epoch": 3.837696335078534,
+ "grad_norm": 0.0733039602637291,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.731414794921875,
+ "step": 276
+ },
+ {
+ "epoch": 3.851657940663176,
+ "grad_norm": 0.05342920497059822,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.734922885894775,
+ "step": 277
+ },
+ {
+ "epoch": 3.8656195462478182,
+ "grad_norm": 0.0504680834710598,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.732746124267578,
+ "step": 278
+ },
+ {
+ "epoch": 3.8795811518324608,
+ "grad_norm": 0.038991592824459076,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.719473838806152,
+ "step": 279
+ },
+ {
+ "epoch": 3.893542757417103,
+ "grad_norm": 0.04434404894709587,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.709481239318848,
+ "step": 280
+ },
+ {
+ "epoch": 3.9075043630017454,
+ "grad_norm": 0.047285184264183044,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.693531036376953,
+ "step": 281
+ },
+ {
+ "epoch": 3.9214659685863875,
+ "grad_norm": 0.045081138610839844,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.728959083557129,
+ "step": 282
+ },
+ {
+ "epoch": 3.9354275741710296,
+ "grad_norm": 0.025133663788437843,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.719696998596191,
+ "step": 283
+ },
+ {
+ "epoch": 3.9493891797556717,
+ "grad_norm": 0.03353596478700638,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.700235366821289,
+ "step": 284
+ },
+ {
+ "epoch": 3.9633507853403143,
+ "grad_norm": 0.035162411630153656,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.684836387634277,
+ "step": 285
+ },
+ {
+ "epoch": 3.9773123909249564,
+ "grad_norm": 0.04025450348854065,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.692097187042236,
+ "step": 286
+ },
+ {
+ "epoch": 3.9912739965095985,
+ "grad_norm": 0.04995397478342056,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.6766862869262695,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.031619708985090256,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 3.5519747734069824,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.714687705039978,
+ "eval_runtime": 59.637,
+ "eval_samples_per_second": 40.948,
+ "eval_steps_per_second": 0.654,
+ "step": 288
+ },
+ {
+ "epoch": 4.013961605584642,
+ "grad_norm": 0.036328598856925964,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.666294097900391,
+ "step": 289
+ },
+ {
+ "epoch": 4.027923211169284,
+ "grad_norm": 0.07326336950063705,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.6879167556762695,
+ "step": 290
+ },
+ {
+ "epoch": 4.041884816753926,
+ "grad_norm": 0.07947526127099991,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.674883842468262,
+ "step": 291
+ },
+ {
+ "epoch": 4.055846422338569,
+ "grad_norm": 0.04599885269999504,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.688086986541748,
+ "step": 292
+ },
+ {
+ "epoch": 4.069808027923211,
+ "grad_norm": 0.0465150885283947,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.649483680725098,
+ "step": 293
+ },
+ {
+ "epoch": 4.0837696335078535,
+ "grad_norm": 0.038542453199625015,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.653055667877197,
+ "step": 294
+ },
+ {
+ "epoch": 4.097731239092496,
+ "grad_norm": 0.03659164905548096,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.65955924987793,
+ "step": 295
+ },
+ {
+ "epoch": 4.111692844677138,
+ "grad_norm": 0.03824038803577423,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.650853157043457,
+ "step": 296
+ },
+ {
+ "epoch": 4.12565445026178,
+ "grad_norm": 0.03591204807162285,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.643499374389648,
+ "step": 297
+ },
+ {
+ "epoch": 4.139616055846423,
+ "grad_norm": 0.03269682824611664,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.657600402832031,
+ "step": 298
+ },
+ {
+ "epoch": 4.153577661431065,
+ "grad_norm": 0.03343537449836731,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.6352081298828125,
+ "step": 299
+ },
+ {
+ "epoch": 4.167539267015707,
+ "grad_norm": 0.034283801913261414,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.626757621765137,
+ "step": 300
+ },
+ {
+ "epoch": 4.181500872600349,
+ "grad_norm": 0.034614723175764084,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.6316680908203125,
+ "step": 301
+ },
+ {
+ "epoch": 4.195462478184991,
+ "grad_norm": 0.0340128131210804,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.648160934448242,
+ "step": 302
+ },
+ {
+ "epoch": 4.209424083769633,
+ "grad_norm": 0.03068280965089798,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.619738578796387,
+ "step": 303
+ },
+ {
+ "epoch": 4.223385689354275,
+ "grad_norm": 0.030269058421254158,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.609238624572754,
+ "step": 304
+ },
+ {
+ "epoch": 4.237347294938918,
+ "grad_norm": 0.03612440824508667,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.602194786071777,
+ "step": 305
+ },
+ {
+ "epoch": 4.2513089005235605,
+ "grad_norm": 0.044292811304330826,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.61159610748291,
+ "step": 306
+ },
+ {
+ "epoch": 4.265270506108203,
+ "grad_norm": 0.04449291527271271,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.6274237632751465,
+ "step": 307
+ },
+ {
+ "epoch": 4.279232111692845,
+ "grad_norm": 0.05014493316411972,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.589034080505371,
+ "step": 308
+ },
+ {
+ "epoch": 4.293193717277487,
+ "grad_norm": 0.05286206305027008,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.600770950317383,
+ "step": 309
+ },
+ {
+ "epoch": 4.307155322862129,
+ "grad_norm": 0.048459846526384354,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.5796098709106445,
+ "step": 310
+ },
+ {
+ "epoch": 4.321116928446771,
+ "grad_norm": 0.05329663306474686,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.590407371520996,
+ "step": 311
+ },
+ {
+ "epoch": 4.335078534031414,
+ "grad_norm": 0.052595119923353195,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.583693504333496,
+ "step": 312
+ },
+ {
+ "epoch": 4.349040139616056,
+ "grad_norm": 0.05164601653814316,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.603281021118164,
+ "step": 313
+ },
+ {
+ "epoch": 4.363001745200698,
+ "grad_norm": 0.04657137766480446,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.591344833374023,
+ "step": 314
+ },
+ {
+ "epoch": 4.37696335078534,
+ "grad_norm": 0.03804599866271019,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.580578804016113,
+ "step": 315
+ },
+ {
+ "epoch": 4.390924956369982,
+ "grad_norm": 0.03835320845246315,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.597836971282959,
+ "step": 316
+ },
+ {
+ "epoch": 4.4048865619546245,
+ "grad_norm": 0.034059133380651474,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.567502021789551,
+ "step": 317
+ },
+ {
+ "epoch": 4.418848167539267,
+ "grad_norm": 0.03350386768579483,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.552022933959961,
+ "step": 318
+ },
+ {
+ "epoch": 4.43280977312391,
+ "grad_norm": 0.032574303448200226,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.559177875518799,
+ "step": 319
+ },
+ {
+ "epoch": 4.446771378708552,
+ "grad_norm": 0.030893854796886444,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.549288272857666,
+ "step": 320
+ },
+ {
+ "epoch": 4.460732984293194,
+ "grad_norm": 0.03255036100745201,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.5400390625,
+ "step": 321
+ },
+ {
+ "epoch": 4.474694589877836,
+ "grad_norm": 0.03009258769452572,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.5328474044799805,
+ "step": 322
+ },
+ {
+ "epoch": 4.488656195462478,
+ "grad_norm": 0.03178320452570915,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.5594482421875,
+ "step": 323
+ },
+ {
+ "epoch": 4.50261780104712,
+ "grad_norm": 0.03360465541481972,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.5402727127075195,
+ "step": 324
+ },
+ {
+ "epoch": 4.516579406631763,
+ "grad_norm": 0.04275049269199371,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.534474849700928,
+ "step": 325
+ },
+ {
+ "epoch": 4.530541012216405,
+ "grad_norm": 0.04140641540288925,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.5324859619140625,
+ "step": 326
+ },
+ {
+ "epoch": 4.544502617801047,
+ "grad_norm": 0.04769204929471016,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.4919352531433105,
+ "step": 327
+ },
+ {
+ "epoch": 4.558464223385689,
+ "grad_norm": 0.04402285814285278,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.520858287811279,
+ "step": 328
+ },
+ {
+ "epoch": 4.5724258289703315,
+ "grad_norm": 0.038526229560375214,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.523778438568115,
+ "step": 329
+ },
+ {
+ "epoch": 4.5863874345549736,
+ "grad_norm": 0.05422377213835716,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.517973899841309,
+ "step": 330
+ },
+ {
+ "epoch": 4.600349040139616,
+ "grad_norm": 0.07236721366643906,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.548083782196045,
+ "step": 331
+ },
+ {
+ "epoch": 4.614310645724259,
+ "grad_norm": 0.06861061602830887,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.535577774047852,
+ "step": 332
+ },
+ {
+ "epoch": 4.628272251308901,
+ "grad_norm": 0.07148340344429016,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.54155158996582,
+ "step": 333
+ },
+ {
+ "epoch": 4.642233856893543,
+ "grad_norm": 0.07838302850723267,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.552852630615234,
+ "step": 334
+ },
+ {
+ "epoch": 4.656195462478185,
+ "grad_norm": 0.10946758091449738,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.555066108703613,
+ "step": 335
+ },
+ {
+ "epoch": 4.670157068062827,
+ "grad_norm": 0.10143530368804932,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.584932804107666,
+ "step": 336
+ },
+ {
+ "epoch": 4.684118673647469,
+ "grad_norm": 0.08926606178283691,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.570462226867676,
+ "step": 337
+ },
+ {
+ "epoch": 4.698080279232112,
+ "grad_norm": 0.07060826569795609,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.5716753005981445,
+ "step": 338
+ },
+ {
+ "epoch": 4.712041884816754,
+ "grad_norm": 0.0418456569314003,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.556496620178223,
+ "step": 339
+ },
+ {
+ "epoch": 4.726003490401396,
+ "grad_norm": 0.05952411890029907,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.567801475524902,
+ "step": 340
+ },
+ {
+ "epoch": 4.739965095986038,
+ "grad_norm": 0.042999885976314545,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.537403106689453,
+ "step": 341
+ },
+ {
+ "epoch": 4.7539267015706805,
+ "grad_norm": 0.04426449537277222,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.552387237548828,
+ "step": 342
+ },
+ {
+ "epoch": 4.767888307155323,
+ "grad_norm": 0.040835145860910416,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.53051233291626,
+ "step": 343
+ },
+ {
+ "epoch": 4.781849912739965,
+ "grad_norm": 0.03819291293621063,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.531174659729004,
+ "step": 344
+ },
+ {
+ "epoch": 4.795811518324607,
+ "grad_norm": 0.030088599771261215,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.499328136444092,
+ "step": 345
+ },
+ {
+ "epoch": 4.80977312390925,
+ "grad_norm": 0.029984362423419952,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.504540920257568,
+ "step": 346
+ },
+ {
+ "epoch": 4.823734729493892,
+ "grad_norm": 0.027402915060520172,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.50899076461792,
+ "step": 347
+ },
+ {
+ "epoch": 4.837696335078534,
+ "grad_norm": 0.025096213445067406,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.480479717254639,
+ "step": 348
+ },
+ {
+ "epoch": 4.851657940663176,
+ "grad_norm": 0.026828132569789886,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.501729965209961,
+ "step": 349
+ },
+ {
+ "epoch": 4.865619546247818,
+ "grad_norm": 0.023715168237686157,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.466860771179199,
+ "step": 350
+ },
+ {
+ "epoch": 4.879581151832461,
+ "grad_norm": 0.024503640830516815,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.491827964782715,
+ "step": 351
+ },
+ {
+ "epoch": 4.893542757417103,
+ "grad_norm": 0.020891105756163597,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.478384971618652,
+ "step": 352
+ },
+ {
+ "epoch": 4.907504363001745,
+ "grad_norm": 0.021373765543103218,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.4612135887146,
+ "step": 353
+ },
+ {
+ "epoch": 4.9214659685863875,
+ "grad_norm": 0.021773340180516243,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.4547319412231445,
+ "step": 354
+ },
+ {
+ "epoch": 4.93542757417103,
+ "grad_norm": 0.019637586548924446,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.431741714477539,
+ "step": 355
+ },
+ {
+ "epoch": 4.949389179755672,
+ "grad_norm": 0.023108696565032005,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.448612213134766,
+ "step": 356
+ },
+ {
+ "epoch": 4.963350785340314,
+ "grad_norm": 0.021935103461146355,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.447112083435059,
+ "step": 357
+ },
+ {
+ "epoch": 4.977312390924956,
+ "grad_norm": 0.023260753601789474,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.436091423034668,
+ "step": 358
+ },
+ {
+ "epoch": 4.991273996509599,
+ "grad_norm": 0.021355057135224342,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.408421993255615,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.014992325566709042,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 3.4004223346710205,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6834986805915833,
+ "eval_runtime": 60.4148,
+ "eval_samples_per_second": 40.421,
+ "eval_steps_per_second": 0.646,
+ "step": 360
+ },
+ {
+ "epoch": 5.013961605584642,
+ "grad_norm": 0.02471218630671501,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.396238803863525,
+ "step": 361
+ },
+ {
+ "epoch": 5.027923211169284,
+ "grad_norm": 0.027672411873936653,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.4048919677734375,
+ "step": 362
+ },
+ {
+ "epoch": 5.041884816753926,
+ "grad_norm": 0.031560689210891724,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.399665832519531,
+ "step": 363
+ },
+ {
+ "epoch": 5.055846422338569,
+ "grad_norm": 0.03459061309695244,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.396360397338867,
+ "step": 364
+ },
+ {
+ "epoch": 5.069808027923211,
+ "grad_norm": 0.04040960222482681,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.405309677124023,
+ "step": 365
+ },
+ {
+ "epoch": 5.0837696335078535,
+ "grad_norm": 0.05444670096039772,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.408894062042236,
+ "step": 366
+ },
+ {
+ "epoch": 5.097731239092496,
+ "grad_norm": 0.05816745012998581,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.403566360473633,
+ "step": 367
+ },
+ {
+ "epoch": 5.111692844677138,
+ "grad_norm": 0.06330069154500961,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.395961761474609,
+ "step": 368
+ },
+ {
+ "epoch": 5.12565445026178,
+ "grad_norm": 0.07458917796611786,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.400968551635742,
+ "step": 369
+ },
+ {
+ "epoch": 5.139616055846423,
+ "grad_norm": 0.06231356039643288,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.412624359130859,
+ "step": 370
+ },
+ {
+ "epoch": 5.153577661431065,
+ "grad_norm": 0.05212346464395523,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.394794464111328,
+ "step": 371
+ },
+ {
+ "epoch": 5.167539267015707,
+ "grad_norm": 0.050003375858068466,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.380125045776367,
+ "step": 372
+ },
+ {
+ "epoch": 5.181500872600349,
+ "grad_norm": 0.048393815755844116,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.390806198120117,
+ "step": 373
+ },
+ {
+ "epoch": 5.195462478184991,
+ "grad_norm": 0.04200794920325279,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.387467384338379,
+ "step": 374
+ },
+ {
+ "epoch": 5.209424083769633,
+ "grad_norm": 0.04931448772549629,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.37552547454834,
+ "step": 375
+ },
+ {
+ "epoch": 5.223385689354275,
+ "grad_norm": 0.03851857781410217,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.373126983642578,
+ "step": 376
+ },
+ {
+ "epoch": 5.237347294938918,
+ "grad_norm": 0.03458717465400696,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.346826553344727,
+ "step": 377
+ },
+ {
+ "epoch": 5.2513089005235605,
+ "grad_norm": 0.031107677146792412,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.353341102600098,
+ "step": 378
+ },
+ {
+ "epoch": 5.265270506108203,
+ "grad_norm": 0.03077058307826519,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.358076095581055,
+ "step": 379
+ },
+ {
+ "epoch": 5.279232111692845,
+ "grad_norm": 0.028170401230454445,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.354019641876221,
+ "step": 380
+ },
+ {
+ "epoch": 5.293193717277487,
+ "grad_norm": 0.02871990203857422,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.368258476257324,
+ "step": 381
+ },
+ {
+ "epoch": 5.307155322862129,
+ "grad_norm": 0.03046584501862526,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.342505931854248,
+ "step": 382
+ },
+ {
+ "epoch": 5.321116928446771,
+ "grad_norm": 0.02781563811004162,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.339411735534668,
+ "step": 383
+ },
+ {
+ "epoch": 5.335078534031414,
+ "grad_norm": 0.028825288638472557,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.341948509216309,
+ "step": 384
+ },
+ {
+ "epoch": 5.349040139616056,
+ "grad_norm": 0.026684079319238663,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.340881824493408,
+ "step": 385
+ },
+ {
+ "epoch": 5.363001745200698,
+ "grad_norm": 0.02429167740046978,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.342733383178711,
+ "step": 386
+ },
+ {
+ "epoch": 5.37696335078534,
+ "grad_norm": 0.02855239436030388,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.313878536224365,
+ "step": 387
+ },
+ {
+ "epoch": 5.390924956369982,
+ "grad_norm": 0.029651548713445663,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.312665939331055,
+ "step": 388
+ },
+ {
+ "epoch": 5.4048865619546245,
+ "grad_norm": 0.027191132307052612,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3086838722229,
+ "step": 389
+ },
+ {
+ "epoch": 5.418848167539267,
+ "grad_norm": 0.026719525456428528,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.305359840393066,
+ "step": 390
+ },
+ {
+ "epoch": 5.43280977312391,
+ "grad_norm": 0.026763100177049637,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.3164381980896,
+ "step": 391
+ },
+ {
+ "epoch": 5.446771378708552,
+ "grad_norm": 0.03125109151005745,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.3083953857421875,
+ "step": 392
+ },
+ {
+ "epoch": 5.460732984293194,
+ "grad_norm": 0.0394827201962471,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.275996208190918,
+ "step": 393
+ },
+ {
+ "epoch": 5.474694589877836,
+ "grad_norm": 0.040579475462436676,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.292995929718018,
+ "step": 394
+ },
+ {
+ "epoch": 5.488656195462478,
+ "grad_norm": 0.037223003804683685,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.298526763916016,
+ "step": 395
+ },
+ {
+ "epoch": 5.50261780104712,
+ "grad_norm": 0.04207765311002731,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.287985801696777,
+ "step": 396
+ },
+ {
+ "epoch": 5.516579406631763,
+ "grad_norm": 0.048299964517354965,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.244608402252197,
+ "step": 397
+ },
+ {
+ "epoch": 5.530541012216405,
+ "grad_norm": 0.04650307446718216,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.285453796386719,
+ "step": 398
+ },
+ {
+ "epoch": 5.544502617801047,
+ "grad_norm": 0.045339711010456085,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.312187671661377,
+ "step": 399
+ },
+ {
+ "epoch": 5.558464223385689,
+ "grad_norm": 0.041067708283662796,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.292692184448242,
+ "step": 400
+ },
+ {
+ "epoch": 5.5724258289703315,
+ "grad_norm": 0.05455026403069496,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.253364562988281,
+ "step": 401
+ },
+ {
+ "epoch": 5.5863874345549736,
+ "grad_norm": 0.06660915911197662,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.262646675109863,
+ "step": 402
+ },
+ {
+ "epoch": 5.600349040139616,
+ "grad_norm": 0.06634793430566788,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.274322986602783,
+ "step": 403
+ },
+ {
+ "epoch": 5.614310645724259,
+ "grad_norm": 0.08996330946683884,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.287951469421387,
+ "step": 404
+ },
+ {
+ "epoch": 5.628272251308901,
+ "grad_norm": 0.08526470512151718,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.3007097244262695,
+ "step": 405
+ },
+ {
+ "epoch": 5.642233856893543,
+ "grad_norm": 0.08766836673021317,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.332538604736328,
+ "step": 406
+ },
+ {
+ "epoch": 5.656195462478185,
+ "grad_norm": 0.07055048644542694,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.296169281005859,
+ "step": 407
+ },
+ {
+ "epoch": 5.670157068062827,
+ "grad_norm": 0.07520575076341629,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.309683322906494,
+ "step": 408
+ },
+ {
+ "epoch": 5.684118673647469,
+ "grad_norm": 0.06115756183862686,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.311301231384277,
+ "step": 409
+ },
+ {
+ "epoch": 5.698080279232112,
+ "grad_norm": 0.05862593650817871,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.28066349029541,
+ "step": 410
+ },
+ {
+ "epoch": 5.712041884816754,
+ "grad_norm": 0.06169601157307625,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.286840438842773,
+ "step": 411
+ },
+ {
+ "epoch": 5.726003490401396,
+ "grad_norm": 0.05270719155669212,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.302594184875488,
+ "step": 412
+ },
+ {
+ "epoch": 5.739965095986038,
+ "grad_norm": 0.03912634402513504,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.262563705444336,
+ "step": 413
+ },
+ {
+ "epoch": 5.7539267015706805,
+ "grad_norm": 0.03215811401605606,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.269310474395752,
+ "step": 414
+ },
+ {
+ "epoch": 5.767888307155323,
+ "grad_norm": 0.03236516937613487,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.268766403198242,
+ "step": 415
+ },
+ {
+ "epoch": 5.781849912739965,
+ "grad_norm": 0.027429452165961266,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.283801555633545,
+ "step": 416
+ },
+ {
+ "epoch": 5.795811518324607,
+ "grad_norm": 0.02990538813173771,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.22291374206543,
+ "step": 417
+ },
+ {
+ "epoch": 5.80977312390925,
+ "grad_norm": 0.030410146340727806,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.2281599044799805,
+ "step": 418
+ },
+ {
+ "epoch": 5.823734729493892,
+ "grad_norm": 0.030068187043070793,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.250294208526611,
+ "step": 419
+ },
+ {
+ "epoch": 5.837696335078534,
+ "grad_norm": 0.02800513058900833,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.226508617401123,
+ "step": 420
+ },
+ {
+ "epoch": 5.851657940663176,
+ "grad_norm": 0.023823332041502,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.247838020324707,
+ "step": 421
+ },
+ {
+ "epoch": 5.865619546247818,
+ "grad_norm": 0.024974249303340912,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.2392473220825195,
+ "step": 422
+ },
+ {
+ "epoch": 5.879581151832461,
+ "grad_norm": 0.023320546373724937,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.209543704986572,
+ "step": 423
+ },
+ {
+ "epoch": 5.893542757417103,
+ "grad_norm": 0.024201491847634315,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.219725608825684,
+ "step": 424
+ },
+ {
+ "epoch": 5.907504363001745,
+ "grad_norm": 0.021649666130542755,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.203141212463379,
+ "step": 425
+ },
+ {
+ "epoch": 5.9214659685863875,
+ "grad_norm": 0.0248002577573061,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.179058074951172,
+ "step": 426
+ },
+ {
+ "epoch": 5.93542757417103,
+ "grad_norm": 0.028549587354063988,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.202126502990723,
+ "step": 427
+ },
+ {
+ "epoch": 5.949389179755672,
+ "grad_norm": 0.0331057533621788,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.208107948303223,
+ "step": 428
+ },
+ {
+ "epoch": 5.963350785340314,
+ "grad_norm": 0.031843364238739014,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.184683799743652,
+ "step": 429
+ },
+ {
+ "epoch": 5.977312390924956,
+ "grad_norm": 0.024927377700805664,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.180486679077148,
+ "step": 430
+ },
+ {
+ "epoch": 5.991273996509599,
+ "grad_norm": 0.026685267686843872,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.186746597290039,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.018545351922512054,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 3.217043876647949,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.653471827507019,
+ "eval_runtime": 61.1586,
+ "eval_samples_per_second": 39.929,
+ "eval_steps_per_second": 0.638,
+ "step": 432
+ },
+ {
+ "epoch": 6.013961605584642,
+ "grad_norm": 0.025291118770837784,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.156534194946289,
+ "step": 433
+ },
+ {
+ "epoch": 6.027923211169284,
+ "grad_norm": 0.03774752840399742,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.144041538238525,
+ "step": 434
+ },
+ {
+ "epoch": 6.041884816753926,
+ "grad_norm": 0.04244692996144295,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.158862113952637,
+ "step": 435
+ },
+ {
+ "epoch": 6.055846422338569,
+ "grad_norm": 0.039782389998435974,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.166165351867676,
+ "step": 436
+ },
+ {
+ "epoch": 6.069808027923211,
+ "grad_norm": 0.05489896610379219,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.155766487121582,
+ "step": 437
+ },
+ {
+ "epoch": 6.0837696335078535,
+ "grad_norm": 0.059844326227903366,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.153733253479004,
+ "step": 438
+ },
+ {
+ "epoch": 6.097731239092496,
+ "grad_norm": 0.05539938807487488,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.130112648010254,
+ "step": 439
+ },
+ {
+ "epoch": 6.111692844677138,
+ "grad_norm": 0.05283183231949806,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.136910915374756,
+ "step": 440
+ },
+ {
+ "epoch": 6.12565445026178,
+ "grad_norm": 0.054407261312007904,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.165216445922852,
+ "step": 441
+ },
+ {
+ "epoch": 6.139616055846423,
+ "grad_norm": 0.0509788915514946,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.140888214111328,
+ "step": 442
+ },
+ {
+ "epoch": 6.153577661431065,
+ "grad_norm": 0.05018748342990875,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.13115930557251,
+ "step": 443
+ },
+ {
+ "epoch": 6.167539267015707,
+ "grad_norm": 0.047012437134981155,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.146507263183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181500872600349,
+ "grad_norm": 0.04943335801362991,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.135608673095703,
+ "step": 445
+ },
+ {
+ "epoch": 6.195462478184991,
+ "grad_norm": 0.05304260924458504,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.136394023895264,
+ "step": 446
+ },
+ {
+ "epoch": 6.209424083769633,
+ "grad_norm": 0.04554047808051109,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.124661922454834,
+ "step": 447
+ },
+ {
+ "epoch": 6.223385689354275,
+ "grad_norm": 0.040672432631254196,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.129773139953613,
+ "step": 448
+ },
+ {
+ "epoch": 6.237347294938918,
+ "grad_norm": 0.03442879393696785,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.128028392791748,
+ "step": 449
+ },
+ {
+ "epoch": 6.2513089005235605,
+ "grad_norm": 0.03927934914827347,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.092485427856445,
+ "step": 450
+ },
+ {
+ "epoch": 6.265270506108203,
+ "grad_norm": 0.029380977153778076,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.077669143676758,
+ "step": 451
+ },
+ {
+ "epoch": 6.279232111692845,
+ "grad_norm": 0.03331821784377098,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.138857841491699,
+ "step": 452
+ },
+ {
+ "epoch": 6.293193717277487,
+ "grad_norm": 0.02939118817448616,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.082937240600586,
+ "step": 453
+ },
+ {
+ "epoch": 6.307155322862129,
+ "grad_norm": 0.02730412967503071,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.093183517456055,
+ "step": 454
+ },
+ {
+ "epoch": 6.321116928446771,
+ "grad_norm": 0.02964780665934086,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.1028151512146,
+ "step": 455
+ },
+ {
+ "epoch": 6.335078534031414,
+ "grad_norm": 0.026350129395723343,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.066362380981445,
+ "step": 456
+ },
+ {
+ "epoch": 6.349040139616056,
+ "grad_norm": 0.027010295540094376,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.070037841796875,
+ "step": 457
+ },
+ {
+ "epoch": 6.363001745200698,
+ "grad_norm": 0.02906128577888012,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.107708930969238,
+ "step": 458
+ },
+ {
+ "epoch": 6.37696335078534,
+ "grad_norm": 0.028619443997740746,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.073182582855225,
+ "step": 459
+ },
+ {
+ "epoch": 6.390924956369982,
+ "grad_norm": 0.029981834813952446,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.076602935791016,
+ "step": 460
+ },
+ {
+ "epoch": 6.4048865619546245,
+ "grad_norm": 0.03255423903465271,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.063301086425781,
+ "step": 461
+ },
+ {
+ "epoch": 6.418848167539267,
+ "grad_norm": 0.03863866627216339,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.041812896728516,
+ "step": 462
+ },
+ {
+ "epoch": 6.43280977312391,
+ "grad_norm": 0.03860306739807129,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.06298828125,
+ "step": 463
+ },
+ {
+ "epoch": 6.446771378708552,
+ "grad_norm": 0.03378435596823692,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.0434370040893555,
+ "step": 464
+ },
+ {
+ "epoch": 6.460732984293194,
+ "grad_norm": 0.03276759013533592,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.036187171936035,
+ "step": 465
+ },
+ {
+ "epoch": 6.474694589877836,
+ "grad_norm": 0.04130513221025467,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.039098739624023,
+ "step": 466
+ },
+ {
+ "epoch": 6.488656195462478,
+ "grad_norm": 0.05389940366148949,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.062798976898193,
+ "step": 467
+ },
+ {
+ "epoch": 6.50261780104712,
+ "grad_norm": 0.05881704390048981,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.067028522491455,
+ "step": 468
+ },
+ {
+ "epoch": 6.516579406631763,
+ "grad_norm": 0.061897799372673035,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.056947231292725,
+ "step": 469
+ },
+ {
+ "epoch": 6.530541012216405,
+ "grad_norm": 0.06549323350191116,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.089262008666992,
+ "step": 470
+ },
+ {
+ "epoch": 6.544502617801047,
+ "grad_norm": 0.048082590103149414,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.055979251861572,
+ "step": 471
+ },
+ {
+ "epoch": 6.558464223385689,
+ "grad_norm": 0.03768027946352959,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.0660600662231445,
+ "step": 472
+ },
+ {
+ "epoch": 6.5724258289703315,
+ "grad_norm": 0.043922603130340576,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 5.036745071411133,
+ "step": 473
+ },
+ {
+ "epoch": 6.5863874345549736,
+ "grad_norm": 0.044848836958408356,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.073650360107422,
+ "step": 474
+ },
+ {
+ "epoch": 6.600349040139616,
+ "grad_norm": 0.05841026455163956,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.053045272827148,
+ "step": 475
+ },
+ {
+ "epoch": 6.614310645724259,
+ "grad_norm": 0.0650017261505127,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 5.052206039428711,
+ "step": 476
+ },
+ {
+ "epoch": 6.628272251308901,
+ "grad_norm": 0.048979002982378006,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 5.054277420043945,
+ "step": 477
+ },
+ {
+ "epoch": 6.642233856893543,
+ "grad_norm": 0.05085799843072891,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 5.063844680786133,
+ "step": 478
+ },
+ {
+ "epoch": 6.656195462478185,
+ "grad_norm": 0.04336017370223999,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 5.045474052429199,
+ "step": 479
+ },
+ {
+ "epoch": 6.670157068062827,
+ "grad_norm": 0.04113380238413811,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 5.031338214874268,
+ "step": 480
+ },
+ {
+ "epoch": 6.684118673647469,
+ "grad_norm": 0.03199538215994835,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 5.030025959014893,
+ "step": 481
+ },
+ {
+ "epoch": 6.698080279232112,
+ "grad_norm": 0.03043113648891449,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 5.033163070678711,
+ "step": 482
+ },
+ {
+ "epoch": 6.712041884816754,
+ "grad_norm": 0.031694378703832626,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 5.004739761352539,
+ "step": 483
+ },
+ {
+ "epoch": 6.726003490401396,
+ "grad_norm": 0.030243348330259323,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 5.012566089630127,
+ "step": 484
+ },
+ {
+ "epoch": 6.739965095986038,
+ "grad_norm": 0.02937416359782219,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 5.039803504943848,
+ "step": 485
+ },
+ {
+ "epoch": 6.7539267015706805,
+ "grad_norm": 0.029616443440318108,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 5.026117324829102,
+ "step": 486
+ },
+ {
+ "epoch": 6.767888307155323,
+ "grad_norm": 0.029567131772637367,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 5.006569862365723,
+ "step": 487
+ },
+ {
+ "epoch": 6.781849912739965,
+ "grad_norm": 0.029443370178341866,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.9941487312316895,
+ "step": 488
+ },
+ {
+ "epoch": 6.795811518324607,
+ "grad_norm": 0.03434352949261665,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.976431846618652,
+ "step": 489
+ },
+ {
+ "epoch": 6.80977312390925,
+ "grad_norm": 0.03845234215259552,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.993168354034424,
+ "step": 490
+ },
+ {
+ "epoch": 6.823734729493892,
+ "grad_norm": 0.05058208107948303,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.030162334442139,
+ "step": 491
+ },
+ {
+ "epoch": 6.837696335078534,
+ "grad_norm": 0.05961710214614868,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 5.007053375244141,
+ "step": 492
+ },
+ {
+ "epoch": 6.851657940663176,
+ "grad_norm": 0.056259628385305405,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 5.020356178283691,
+ "step": 493
+ },
+ {
+ "epoch": 6.865619546247818,
+ "grad_norm": 0.048628658056259155,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 5.037652969360352,
+ "step": 494
+ },
+ {
+ "epoch": 6.879581151832461,
+ "grad_norm": 0.040836453437805176,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.990390777587891,
+ "step": 495
+ },
+ {
+ "epoch": 6.893542757417103,
+ "grad_norm": 0.04112066701054573,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 5.017822742462158,
+ "step": 496
+ },
+ {
+ "epoch": 6.907504363001745,
+ "grad_norm": 0.03956829756498337,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.993971824645996,
+ "step": 497
+ },
+ {
+ "epoch": 6.9214659685863875,
+ "grad_norm": 0.03887917473912239,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.982889175415039,
+ "step": 498
+ },
+ {
+ "epoch": 6.93542757417103,
+ "grad_norm": 0.031012587249279022,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.961717128753662,
+ "step": 499
+ },
+ {
+ "epoch": 6.949389179755672,
+ "grad_norm": 0.03443489596247673,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.9904046058654785,
+ "step": 500
+ },
+ {
+ "epoch": 6.963350785340314,
+ "grad_norm": 0.035175010561943054,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.983619213104248,
+ "step": 501
+ },
+ {
+ "epoch": 6.977312390924956,
+ "grad_norm": 0.031724050641059875,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 5.000522613525391,
+ "step": 502
+ },
+ {
+ "epoch": 6.991273996509599,
+ "grad_norm": 0.032037995755672455,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.947064399719238,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.01883271522819996,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 3.112166166305542,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6296830177307129,
+ "eval_runtime": 59.5495,
+ "eval_samples_per_second": 41.008,
+ "eval_steps_per_second": 0.655,
+ "step": 504
+ },
+ {
+ "epoch": 7.013961605584642,
+ "grad_norm": 0.031091563403606415,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.919820785522461,
+ "step": 505
+ },
+ {
+ "epoch": 7.027923211169284,
+ "grad_norm": 0.04074421525001526,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.945765972137451,
+ "step": 506
+ },
+ {
+ "epoch": 7.041884816753926,
+ "grad_norm": 0.04225052148103714,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.935208797454834,
+ "step": 507
+ },
+ {
+ "epoch": 7.055846422338569,
+ "grad_norm": 0.0399051196873188,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.951314926147461,
+ "step": 508
+ },
+ {
+ "epoch": 7.069808027923211,
+ "grad_norm": 0.04785890504717827,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.92843770980835,
+ "step": 509
+ },
+ {
+ "epoch": 7.0837696335078535,
+ "grad_norm": 0.04922301694750786,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.901020050048828,
+ "step": 510
+ },
+ {
+ "epoch": 7.097731239092496,
+ "grad_norm": 0.05307295173406601,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.961151599884033,
+ "step": 511
+ },
+ {
+ "epoch": 7.111692844677138,
+ "grad_norm": 0.07003362476825714,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.9427876472473145,
+ "step": 512
+ },
+ {
+ "epoch": 7.12565445026178,
+ "grad_norm": 0.06498776376247406,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.947838306427002,
+ "step": 513
+ },
+ {
+ "epoch": 7.139616055846423,
+ "grad_norm": 0.058723628520965576,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.938859939575195,
+ "step": 514
+ },
+ {
+ "epoch": 7.153577661431065,
+ "grad_norm": 0.06590746343135834,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.962221145629883,
+ "step": 515
+ },
+ {
+ "epoch": 7.167539267015707,
+ "grad_norm": 0.052714601159095764,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.931011199951172,
+ "step": 516
+ },
+ {
+ "epoch": 7.181500872600349,
+ "grad_norm": 0.04296804964542389,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.93586540222168,
+ "step": 517
+ },
+ {
+ "epoch": 7.195462478184991,
+ "grad_norm": 0.03699173778295517,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.910763263702393,
+ "step": 518
+ },
+ {
+ "epoch": 7.209424083769633,
+ "grad_norm": 0.0317913293838501,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.9256439208984375,
+ "step": 519
+ },
+ {
+ "epoch": 7.223385689354275,
+ "grad_norm": 0.030313925817608833,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.938329696655273,
+ "step": 520
+ },
+ {
+ "epoch": 7.237347294938918,
+ "grad_norm": 0.02879229187965393,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.924966812133789,
+ "step": 521
+ },
+ {
+ "epoch": 7.2513089005235605,
+ "grad_norm": 0.02689451351761818,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.9079742431640625,
+ "step": 522
+ },
+ {
+ "epoch": 7.265270506108203,
+ "grad_norm": 0.028403930366039276,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.913238525390625,
+ "step": 523
+ },
+ {
+ "epoch": 7.279232111692845,
+ "grad_norm": 0.025912854820489883,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.896944046020508,
+ "step": 524
+ },
+ {
+ "epoch": 7.293193717277487,
+ "grad_norm": 0.024899760261178017,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.884920120239258,
+ "step": 525
+ },
+ {
+ "epoch": 7.307155322862129,
+ "grad_norm": 0.026912573724985123,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.9181060791015625,
+ "step": 526
+ },
+ {
+ "epoch": 7.321116928446771,
+ "grad_norm": 0.029539721086621284,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.914867401123047,
+ "step": 527
+ },
+ {
+ "epoch": 7.335078534031414,
+ "grad_norm": 0.028527872636914253,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.907217979431152,
+ "step": 528
+ },
+ {
+ "epoch": 7.349040139616056,
+ "grad_norm": 0.025879016146063805,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.905783176422119,
+ "step": 529
+ },
+ {
+ "epoch": 7.363001745200698,
+ "grad_norm": 0.026395199820399284,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.880505561828613,
+ "step": 530
+ },
+ {
+ "epoch": 7.37696335078534,
+ "grad_norm": 0.025072351098060608,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.870449066162109,
+ "step": 531
+ },
+ {
+ "epoch": 7.390924956369982,
+ "grad_norm": 0.02751019410789013,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.888952255249023,
+ "step": 532
+ },
+ {
+ "epoch": 7.4048865619546245,
+ "grad_norm": 0.025810658931732178,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.894552230834961,
+ "step": 533
+ },
+ {
+ "epoch": 7.418848167539267,
+ "grad_norm": 0.02399870567023754,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.8768229484558105,
+ "step": 534
+ },
+ {
+ "epoch": 7.43280977312391,
+ "grad_norm": 0.024639368057250977,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.9030914306640625,
+ "step": 535
+ },
+ {
+ "epoch": 7.446771378708552,
+ "grad_norm": 0.02543291077017784,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.86467170715332,
+ "step": 536
+ },
+ {
+ "epoch": 7.460732984293194,
+ "grad_norm": 0.02702544629573822,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.888283729553223,
+ "step": 537
+ },
+ {
+ "epoch": 7.474694589877836,
+ "grad_norm": 0.028904743492603302,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.866081237792969,
+ "step": 538
+ },
+ {
+ "epoch": 7.488656195462478,
+ "grad_norm": 0.03055848367512226,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.890739440917969,
+ "step": 539
+ },
+ {
+ "epoch": 7.50261780104712,
+ "grad_norm": 0.03410084918141365,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.862918853759766,
+ "step": 540
+ },
+ {
+ "epoch": 7.516579406631763,
+ "grad_norm": 0.03738495334982872,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.864728927612305,
+ "step": 541
+ },
+ {
+ "epoch": 7.530541012216405,
+ "grad_norm": 0.037646882236003876,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.865185737609863,
+ "step": 542
+ },
+ {
+ "epoch": 7.544502617801047,
+ "grad_norm": 0.03418618440628052,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.866114139556885,
+ "step": 543
+ },
+ {
+ "epoch": 7.558464223385689,
+ "grad_norm": 0.036924004554748535,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.883186340332031,
+ "step": 544
+ },
+ {
+ "epoch": 7.5724258289703315,
+ "grad_norm": 0.03590598329901695,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.863725662231445,
+ "step": 545
+ },
+ {
+ "epoch": 7.5863874345549736,
+ "grad_norm": 0.037363454699516296,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.841030120849609,
+ "step": 546
+ },
+ {
+ "epoch": 7.600349040139616,
+ "grad_norm": 0.046537864953279495,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.873837471008301,
+ "step": 547
+ },
+ {
+ "epoch": 7.614310645724259,
+ "grad_norm": 0.057226769626140594,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.865679740905762,
+ "step": 548
+ },
+ {
+ "epoch": 7.628272251308901,
+ "grad_norm": 0.06175961345434189,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.888645648956299,
+ "step": 549
+ },
+ {
+ "epoch": 7.642233856893543,
+ "grad_norm": 0.05063694715499878,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.877828598022461,
+ "step": 550
+ },
+ {
+ "epoch": 7.656195462478185,
+ "grad_norm": 0.03950628638267517,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.88637638092041,
+ "step": 551
+ },
+ {
+ "epoch": 7.670157068062827,
+ "grad_norm": 0.04183337092399597,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.8597917556762695,
+ "step": 552
+ },
+ {
+ "epoch": 7.684118673647469,
+ "grad_norm": 0.04642999917268753,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.880380153656006,
+ "step": 553
+ },
+ {
+ "epoch": 7.698080279232112,
+ "grad_norm": 0.05333106964826584,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.853932857513428,
+ "step": 554
+ },
+ {
+ "epoch": 7.712041884816754,
+ "grad_norm": 0.04764975979924202,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.870235443115234,
+ "step": 555
+ },
+ {
+ "epoch": 7.726003490401396,
+ "grad_norm": 0.043072231113910675,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.822697639465332,
+ "step": 556
+ },
+ {
+ "epoch": 7.739965095986038,
+ "grad_norm": 0.03913857042789459,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.849332809448242,
+ "step": 557
+ },
+ {
+ "epoch": 7.7539267015706805,
+ "grad_norm": 0.034836407750844955,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.842280387878418,
+ "step": 558
+ },
+ {
+ "epoch": 7.767888307155323,
+ "grad_norm": 0.033344294875860214,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.837811470031738,
+ "step": 559
+ },
+ {
+ "epoch": 7.781849912739965,
+ "grad_norm": 0.03342701122164726,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.876040458679199,
+ "step": 560
+ },
+ {
+ "epoch": 7.795811518324607,
+ "grad_norm": 0.034638747572898865,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.830340385437012,
+ "step": 561
+ },
+ {
+ "epoch": 7.80977312390925,
+ "grad_norm": 0.03358135372400284,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.828607559204102,
+ "step": 562
+ },
+ {
+ "epoch": 7.823734729493892,
+ "grad_norm": 0.036001041531562805,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.827905654907227,
+ "step": 563
+ },
+ {
+ "epoch": 7.837696335078534,
+ "grad_norm": 0.038079630583524704,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.866955280303955,
+ "step": 564
+ },
+ {
+ "epoch": 7.851657940663176,
+ "grad_norm": 0.03728470206260681,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.828039169311523,
+ "step": 565
+ },
+ {
+ "epoch": 7.865619546247818,
+ "grad_norm": 0.03920375928282738,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.841631889343262,
+ "step": 566
+ },
+ {
+ "epoch": 7.879581151832461,
+ "grad_norm": 0.032353926450014114,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.861110210418701,
+ "step": 567
+ },
+ {
+ "epoch": 7.893542757417103,
+ "grad_norm": 0.03241422027349472,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.818441390991211,
+ "step": 568
+ },
+ {
+ "epoch": 7.907504363001745,
+ "grad_norm": 0.03879080340266228,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.823749542236328,
+ "step": 569
+ },
+ {
+ "epoch": 7.9214659685863875,
+ "grad_norm": 0.03486498072743416,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.830697059631348,
+ "step": 570
+ },
+ {
+ "epoch": 7.93542757417103,
+ "grad_norm": 0.03279713913798332,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.815736770629883,
+ "step": 571
+ },
+ {
+ "epoch": 7.949389179755672,
+ "grad_norm": 0.03313925117254257,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.8364033699035645,
+ "step": 572
+ },
+ {
+ "epoch": 7.963350785340314,
+ "grad_norm": 0.03405062481760979,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.816010475158691,
+ "step": 573
+ },
+ {
+ "epoch": 7.977312390924956,
+ "grad_norm": 0.03259236365556717,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.813413619995117,
+ "step": 574
+ },
+ {
+ "epoch": 7.991273996509599,
+ "grad_norm": 0.030408985912799835,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.833840370178223,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.022680889815092087,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 2.99565052986145,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6126086115837097,
+ "eval_runtime": 60.2662,
+ "eval_samples_per_second": 40.52,
+ "eval_steps_per_second": 0.647,
+ "step": 576
+ },
+ {
+ "epoch": 8.013961605584642,
+ "grad_norm": 0.029732126742601395,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.7816314697265625,
+ "step": 577
+ },
+ {
+ "epoch": 8.027923211169284,
+ "grad_norm": 0.04357453063130379,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.7575225830078125,
+ "step": 578
+ },
+ {
+ "epoch": 8.041884816753926,
+ "grad_norm": 0.051903221756219864,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.802519798278809,
+ "step": 579
+ },
+ {
+ "epoch": 8.055846422338568,
+ "grad_norm": 0.0539076030254364,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.799004554748535,
+ "step": 580
+ },
+ {
+ "epoch": 8.06980802792321,
+ "grad_norm": 0.043642807751894,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.7896528244018555,
+ "step": 581
+ },
+ {
+ "epoch": 8.083769633507853,
+ "grad_norm": 0.0469554178416729,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.765430927276611,
+ "step": 582
+ },
+ {
+ "epoch": 8.097731239092496,
+ "grad_norm": 0.042060062289237976,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.784132957458496,
+ "step": 583
+ },
+ {
+ "epoch": 8.111692844677139,
+ "grad_norm": 0.035101912915706635,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.767234802246094,
+ "step": 584
+ },
+ {
+ "epoch": 8.12565445026178,
+ "grad_norm": 0.042576879262924194,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.762052536010742,
+ "step": 585
+ },
+ {
+ "epoch": 8.139616055846423,
+ "grad_norm": 0.04231419041752815,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.752150535583496,
+ "step": 586
+ },
+ {
+ "epoch": 8.153577661431065,
+ "grad_norm": 0.046912599354982376,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.771881103515625,
+ "step": 587
+ },
+ {
+ "epoch": 8.167539267015707,
+ "grad_norm": 0.051271144300699234,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.773284912109375,
+ "step": 588
+ },
+ {
+ "epoch": 8.181500872600349,
+ "grad_norm": 0.04655958339571953,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.769917964935303,
+ "step": 589
+ },
+ {
+ "epoch": 8.195462478184991,
+ "grad_norm": 0.04463687539100647,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.77938175201416,
+ "step": 590
+ },
+ {
+ "epoch": 8.209424083769633,
+ "grad_norm": 0.047959886491298676,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.785962104797363,
+ "step": 591
+ },
+ {
+ "epoch": 8.223385689354275,
+ "grad_norm": 0.04414986073970795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.786281585693359,
+ "step": 592
+ },
+ {
+ "epoch": 8.237347294938917,
+ "grad_norm": 0.044150348752737045,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.777070045471191,
+ "step": 593
+ },
+ {
+ "epoch": 8.25130890052356,
+ "grad_norm": 0.037002332508563995,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.77547550201416,
+ "step": 594
+ },
+ {
+ "epoch": 8.265270506108202,
+ "grad_norm": 0.03417367860674858,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.782101154327393,
+ "step": 595
+ },
+ {
+ "epoch": 8.279232111692846,
+ "grad_norm": 0.03069210797548294,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.745595455169678,
+ "step": 596
+ },
+ {
+ "epoch": 8.293193717277488,
+ "grad_norm": 0.03228718787431717,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.7663655281066895,
+ "step": 597
+ },
+ {
+ "epoch": 8.30715532286213,
+ "grad_norm": 0.028603695333003998,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.767274856567383,
+ "step": 598
+ },
+ {
+ "epoch": 8.321116928446772,
+ "grad_norm": 0.03098377026617527,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.7853684425354,
+ "step": 599
+ },
+ {
+ "epoch": 8.335078534031414,
+ "grad_norm": 0.029242439195513725,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.738029956817627,
+ "step": 600
+ },
+ {
+ "epoch": 8.349040139616056,
+ "grad_norm": 0.02751081995666027,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.76801872253418,
+ "step": 601
+ },
+ {
+ "epoch": 8.363001745200698,
+ "grad_norm": 0.02838125079870224,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.743884086608887,
+ "step": 602
+ },
+ {
+ "epoch": 8.37696335078534,
+ "grad_norm": 0.029203062877058983,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.746668815612793,
+ "step": 603
+ },
+ {
+ "epoch": 8.390924956369982,
+ "grad_norm": 0.02679944597184658,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.760764122009277,
+ "step": 604
+ },
+ {
+ "epoch": 8.404886561954624,
+ "grad_norm": 0.028955848887562752,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.7353668212890625,
+ "step": 605
+ },
+ {
+ "epoch": 8.418848167539267,
+ "grad_norm": 0.02880851738154888,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.766837120056152,
+ "step": 606
+ },
+ {
+ "epoch": 8.432809773123909,
+ "grad_norm": 0.02495497092604637,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.78959321975708,
+ "step": 607
+ },
+ {
+ "epoch": 8.44677137870855,
+ "grad_norm": 0.026064753532409668,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.736433029174805,
+ "step": 608
+ },
+ {
+ "epoch": 8.460732984293193,
+ "grad_norm": 0.027566388249397278,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.782382965087891,
+ "step": 609
+ },
+ {
+ "epoch": 8.474694589877837,
+ "grad_norm": 0.02967359498143196,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.735286712646484,
+ "step": 610
+ },
+ {
+ "epoch": 8.488656195462479,
+ "grad_norm": 0.03079676441848278,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.725196838378906,
+ "step": 611
+ },
+ {
+ "epoch": 8.502617801047121,
+ "grad_norm": 0.032531846314668655,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.719333648681641,
+ "step": 612
+ },
+ {
+ "epoch": 8.516579406631763,
+ "grad_norm": 0.03120352514088154,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.723845958709717,
+ "step": 613
+ },
+ {
+ "epoch": 8.530541012216405,
+ "grad_norm": 0.03328424692153931,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.7669782638549805,
+ "step": 614
+ },
+ {
+ "epoch": 8.544502617801047,
+ "grad_norm": 0.035601649433374405,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.714925765991211,
+ "step": 615
+ },
+ {
+ "epoch": 8.55846422338569,
+ "grad_norm": 0.03675796464085579,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.715317726135254,
+ "step": 616
+ },
+ {
+ "epoch": 8.572425828970331,
+ "grad_norm": 0.03701988235116005,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.724606990814209,
+ "step": 617
+ },
+ {
+ "epoch": 8.586387434554974,
+ "grad_norm": 0.036719705909490585,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.746044635772705,
+ "step": 618
+ },
+ {
+ "epoch": 8.600349040139616,
+ "grad_norm": 0.034756604582071304,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.74885368347168,
+ "step": 619
+ },
+ {
+ "epoch": 8.614310645724258,
+ "grad_norm": 0.03237922117114067,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.721858501434326,
+ "step": 620
+ },
+ {
+ "epoch": 8.6282722513089,
+ "grad_norm": 0.028420142829418182,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.733814239501953,
+ "step": 621
+ },
+ {
+ "epoch": 8.642233856893542,
+ "grad_norm": 0.026203038170933723,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.726088523864746,
+ "step": 622
+ },
+ {
+ "epoch": 8.656195462478184,
+ "grad_norm": 0.02627231739461422,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.727700233459473,
+ "step": 623
+ },
+ {
+ "epoch": 8.670157068062828,
+ "grad_norm": 0.02920261025428772,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.731454849243164,
+ "step": 624
+ },
+ {
+ "epoch": 8.68411867364747,
+ "grad_norm": 0.02792907878756523,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.734154224395752,
+ "step": 625
+ },
+ {
+ "epoch": 8.698080279232112,
+ "grad_norm": 0.026770325377583504,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.715337753295898,
+ "step": 626
+ },
+ {
+ "epoch": 8.712041884816754,
+ "grad_norm": 0.027346540242433548,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.716512680053711,
+ "step": 627
+ },
+ {
+ "epoch": 8.726003490401396,
+ "grad_norm": 0.031965773552656174,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.7072038650512695,
+ "step": 628
+ },
+ {
+ "epoch": 8.739965095986038,
+ "grad_norm": 0.03708025813102722,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.749785900115967,
+ "step": 629
+ },
+ {
+ "epoch": 8.75392670157068,
+ "grad_norm": 0.04846490919589996,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.735586166381836,
+ "step": 630
+ },
+ {
+ "epoch": 8.767888307155323,
+ "grad_norm": 0.053352758288383484,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.720524787902832,
+ "step": 631
+ },
+ {
+ "epoch": 8.781849912739965,
+ "grad_norm": 0.040559202432632446,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.719901084899902,
+ "step": 632
+ },
+ {
+ "epoch": 8.795811518324607,
+ "grad_norm": 0.03333032503724098,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.708139419555664,
+ "step": 633
+ },
+ {
+ "epoch": 8.809773123909249,
+ "grad_norm": 0.030532941222190857,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.728184223175049,
+ "step": 634
+ },
+ {
+ "epoch": 8.823734729493891,
+ "grad_norm": 0.03337649628520012,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.726020336151123,
+ "step": 635
+ },
+ {
+ "epoch": 8.837696335078533,
+ "grad_norm": 0.03352045267820358,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.7196855545043945,
+ "step": 636
+ },
+ {
+ "epoch": 8.851657940663177,
+ "grad_norm": 0.03662872314453125,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.732823371887207,
+ "step": 637
+ },
+ {
+ "epoch": 8.86561954624782,
+ "grad_norm": 0.037085745483636856,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.728443145751953,
+ "step": 638
+ },
+ {
+ "epoch": 8.879581151832461,
+ "grad_norm": 0.036635689437389374,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.713988780975342,
+ "step": 639
+ },
+ {
+ "epoch": 8.893542757417103,
+ "grad_norm": 0.03719659522175789,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.715950965881348,
+ "step": 640
+ },
+ {
+ "epoch": 8.907504363001745,
+ "grad_norm": 0.03719159588217735,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.7124924659729,
+ "step": 641
+ },
+ {
+ "epoch": 8.921465968586388,
+ "grad_norm": 0.04275721684098244,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.708404064178467,
+ "step": 642
+ },
+ {
+ "epoch": 8.93542757417103,
+ "grad_norm": 0.04204598814249039,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.715372085571289,
+ "step": 643
+ },
+ {
+ "epoch": 8.949389179755672,
+ "grad_norm": 0.03938259929418564,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.729745864868164,
+ "step": 644
+ },
+ {
+ "epoch": 8.963350785340314,
+ "grad_norm": 0.036226850003004074,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.733749866485596,
+ "step": 645
+ },
+ {
+ "epoch": 8.977312390924956,
+ "grad_norm": 0.03542874753475189,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.706825256347656,
+ "step": 646
+ },
+ {
+ "epoch": 8.991273996509598,
+ "grad_norm": 0.029055381193757057,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.718567848205566,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.01990380510687828,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 2.943183183670044,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.601699709892273,
+ "eval_runtime": 60.0221,
+ "eval_samples_per_second": 40.685,
+ "eval_steps_per_second": 0.65,
+ "step": 648
+ },
+ {
+ "epoch": 9.013961605584642,
+ "grad_norm": 0.030904127284884453,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.660327911376953,
+ "step": 649
+ },
+ {
+ "epoch": 9.027923211169284,
+ "grad_norm": 0.0441313236951828,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.701565742492676,
+ "step": 650
+ },
+ {
+ "epoch": 9.041884816753926,
+ "grad_norm": 0.047547727823257446,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.666440486907959,
+ "step": 651
+ },
+ {
+ "epoch": 9.055846422338568,
+ "grad_norm": 0.04549463838338852,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.660548210144043,
+ "step": 652
+ },
+ {
+ "epoch": 9.06980802792321,
+ "grad_norm": 0.050346389412879944,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.640229225158691,
+ "step": 653
+ },
+ {
+ "epoch": 9.083769633507853,
+ "grad_norm": 0.058505453169345856,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.673881530761719,
+ "step": 654
+ },
+ {
+ "epoch": 9.097731239092496,
+ "grad_norm": 0.051989227533340454,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.693943977355957,
+ "step": 655
+ },
+ {
+ "epoch": 9.111692844677139,
+ "grad_norm": 0.052564945071935654,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.644315242767334,
+ "step": 656
+ },
+ {
+ "epoch": 9.12565445026178,
+ "grad_norm": 0.05647841840982437,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.681768417358398,
+ "step": 657
+ },
+ {
+ "epoch": 9.139616055846423,
+ "grad_norm": 0.058640867471694946,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.676457405090332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153577661431065,
+ "grad_norm": 0.05386025458574295,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.659588813781738,
+ "step": 659
+ },
+ {
+ "epoch": 9.167539267015707,
+ "grad_norm": 0.048699021339416504,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.680031776428223,
+ "step": 660
+ },
+ {
+ "epoch": 9.181500872600349,
+ "grad_norm": 0.04374176263809204,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.672671318054199,
+ "step": 661
+ },
+ {
+ "epoch": 9.195462478184991,
+ "grad_norm": 0.042337436228990555,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.673431873321533,
+ "step": 662
+ },
+ {
+ "epoch": 9.209424083769633,
+ "grad_norm": 0.04564661160111427,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.632595062255859,
+ "step": 663
+ },
+ {
+ "epoch": 9.223385689354275,
+ "grad_norm": 0.04404835030436516,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.678891181945801,
+ "step": 664
+ },
+ {
+ "epoch": 9.237347294938917,
+ "grad_norm": 0.040642790496349335,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.633418560028076,
+ "step": 665
+ },
+ {
+ "epoch": 9.25130890052356,
+ "grad_norm": 0.036492060869932175,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.670109748840332,
+ "step": 666
+ },
+ {
+ "epoch": 9.265270506108202,
+ "grad_norm": 0.03735879808664322,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.608570098876953,
+ "step": 667
+ },
+ {
+ "epoch": 9.279232111692846,
+ "grad_norm": 0.034563641995191574,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.632505416870117,
+ "step": 668
+ },
+ {
+ "epoch": 9.293193717277488,
+ "grad_norm": 0.03522484749555588,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.6475629806518555,
+ "step": 669
+ },
+ {
+ "epoch": 9.30715532286213,
+ "grad_norm": 0.03216472268104553,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.627873420715332,
+ "step": 670
+ },
+ {
+ "epoch": 9.321116928446772,
+ "grad_norm": 0.029155759140849113,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.652706146240234,
+ "step": 671
+ },
+ {
+ "epoch": 9.335078534031414,
+ "grad_norm": 0.031060541048645973,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.644545078277588,
+ "step": 672
+ },
+ {
+ "epoch": 9.349040139616056,
+ "grad_norm": 0.034979186952114105,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.642645359039307,
+ "step": 673
+ },
+ {
+ "epoch": 9.363001745200698,
+ "grad_norm": 0.035258810967206955,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.647383689880371,
+ "step": 674
+ },
+ {
+ "epoch": 9.37696335078534,
+ "grad_norm": 0.031552575528621674,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.667779445648193,
+ "step": 675
+ },
+ {
+ "epoch": 9.390924956369982,
+ "grad_norm": 0.03193951025605202,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.63625431060791,
+ "step": 676
+ },
+ {
+ "epoch": 9.404886561954624,
+ "grad_norm": 0.029811711981892586,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.658000946044922,
+ "step": 677
+ },
+ {
+ "epoch": 9.418848167539267,
+ "grad_norm": 0.028464138507843018,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.6545515060424805,
+ "step": 678
+ },
+ {
+ "epoch": 9.432809773123909,
+ "grad_norm": 0.0315735749900341,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.641153335571289,
+ "step": 679
+ },
+ {
+ "epoch": 9.44677137870855,
+ "grad_norm": 0.033412106335163116,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.626931667327881,
+ "step": 680
+ },
+ {
+ "epoch": 9.460732984293193,
+ "grad_norm": 0.03757626190781593,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.671254634857178,
+ "step": 681
+ },
+ {
+ "epoch": 9.474694589877837,
+ "grad_norm": 0.03910793736577034,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.651954174041748,
+ "step": 682
+ },
+ {
+ "epoch": 9.488656195462479,
+ "grad_norm": 0.04015549272298813,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.613324165344238,
+ "step": 683
+ },
+ {
+ "epoch": 9.502617801047121,
+ "grad_norm": 0.040689144283533096,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.639243125915527,
+ "step": 684
+ },
+ {
+ "epoch": 9.516579406631763,
+ "grad_norm": 0.04075361415743828,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.649250030517578,
+ "step": 685
+ },
+ {
+ "epoch": 9.530541012216405,
+ "grad_norm": 0.03914607688784599,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.66444730758667,
+ "step": 686
+ },
+ {
+ "epoch": 9.544502617801047,
+ "grad_norm": 0.033253733068704605,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.647335052490234,
+ "step": 687
+ },
+ {
+ "epoch": 9.55846422338569,
+ "grad_norm": 0.03179008886218071,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.6727800369262695,
+ "step": 688
+ },
+ {
+ "epoch": 9.572425828970331,
+ "grad_norm": 0.03477025404572487,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.655778884887695,
+ "step": 689
+ },
+ {
+ "epoch": 9.586387434554974,
+ "grad_norm": 0.030094800516963005,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.680851936340332,
+ "step": 690
+ },
+ {
+ "epoch": 9.600349040139616,
+ "grad_norm": 0.030153103172779083,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.654638290405273,
+ "step": 691
+ },
+ {
+ "epoch": 9.614310645724258,
+ "grad_norm": 0.029009278863668442,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.625916481018066,
+ "step": 692
+ },
+ {
+ "epoch": 9.6282722513089,
+ "grad_norm": 0.027248440310359,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.656370162963867,
+ "step": 693
+ },
+ {
+ "epoch": 9.642233856893542,
+ "grad_norm": 0.027727462351322174,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.631473541259766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656195462478184,
+ "grad_norm": 0.026710307225584984,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.636524200439453,
+ "step": 695
+ },
+ {
+ "epoch": 9.670157068062828,
+ "grad_norm": 0.02726181596517563,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.629701614379883,
+ "step": 696
+ },
+ {
+ "epoch": 9.68411867364747,
+ "grad_norm": 0.027146652340888977,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.640240669250488,
+ "step": 697
+ },
+ {
+ "epoch": 9.698080279232112,
+ "grad_norm": 0.027165887877345085,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.64462947845459,
+ "step": 698
+ },
+ {
+ "epoch": 9.712041884816754,
+ "grad_norm": 0.03025982715189457,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.6271820068359375,
+ "step": 699
+ },
+ {
+ "epoch": 9.726003490401396,
+ "grad_norm": 0.02837025746703148,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.629697799682617,
+ "step": 700
+ },
+ {
+ "epoch": 9.739965095986038,
+ "grad_norm": 0.026777995750308037,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.646146774291992,
+ "step": 701
+ },
+ {
+ "epoch": 9.75392670157068,
+ "grad_norm": 0.024626802653074265,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.641671657562256,
+ "step": 702
+ },
+ {
+ "epoch": 9.767888307155323,
+ "grad_norm": 0.025329895317554474,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.626345157623291,
+ "step": 703
+ },
+ {
+ "epoch": 9.781849912739965,
+ "grad_norm": 0.027804844081401825,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.621457099914551,
+ "step": 704
+ },
+ {
+ "epoch": 9.795811518324607,
+ "grad_norm": 0.028763437643647194,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.6124114990234375,
+ "step": 705
+ },
+ {
+ "epoch": 9.809773123909249,
+ "grad_norm": 0.030769916251301765,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.604399681091309,
+ "step": 706
+ },
+ {
+ "epoch": 9.823734729493891,
+ "grad_norm": 0.03060949593782425,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.640804290771484,
+ "step": 707
+ },
+ {
+ "epoch": 9.837696335078533,
+ "grad_norm": 0.029832778498530388,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.593376159667969,
+ "step": 708
+ },
+ {
+ "epoch": 9.851657940663177,
+ "grad_norm": 0.02952776849269867,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.619533538818359,
+ "step": 709
+ },
+ {
+ "epoch": 9.86561954624782,
+ "grad_norm": 0.02894212119281292,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.640516757965088,
+ "step": 710
+ },
+ {
+ "epoch": 9.879581151832461,
+ "grad_norm": 0.026435624808073044,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.6342997550964355,
+ "step": 711
+ },
+ {
+ "epoch": 9.893542757417103,
+ "grad_norm": 0.02591666206717491,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.630884170532227,
+ "step": 712
+ },
+ {
+ "epoch": 9.907504363001745,
+ "grad_norm": 0.02702309936285019,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.632085800170898,
+ "step": 713
+ },
+ {
+ "epoch": 9.921465968586388,
+ "grad_norm": 0.026766657829284668,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.6231231689453125,
+ "step": 714
+ },
+ {
+ "epoch": 9.93542757417103,
+ "grad_norm": 0.026666559278964996,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.601619243621826,
+ "step": 715
+ },
+ {
+ "epoch": 9.949389179755672,
+ "grad_norm": 0.026801064610481262,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.641724109649658,
+ "step": 716
+ },
+ {
+ "epoch": 9.963350785340314,
+ "grad_norm": 0.03152798116207123,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.642511367797852,
+ "step": 717
+ },
+ {
+ "epoch": 9.977312390924956,
+ "grad_norm": 0.035406943410634995,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.635779857635498,
+ "step": 718
+ },
+ {
+ "epoch": 9.991273996509598,
+ "grad_norm": 0.037504684180021286,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.629463195800781,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.02263174019753933,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 2.876767635345459,
+ "step": 720
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.072294246363955e+17,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-wd0033-ssl/checkpoint-720/training_args.bin b/runs/l2r90-wd0033-ssl/checkpoint-720/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..313c41733ff55284c06034a59bb885b30e40ffc8
--- /dev/null
+++ b/runs/l2r90-wd0033-ssl/checkpoint-720/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7ddd0bdddba68d3e19ab33477000f3fc684fe4ee8e3580eeb8c02646e5fc27ac
+size 4856