diff --git a/.gitattributes b/.gitattributes
index 030a8fcd11af1bf9daec869768e4f2cb3629fa2d..517cc278bfb10732bfccc9ac530f98efa1485385 100644
--- a/.gitattributes
+++ b/.gitattributes
@@ -140,3 +140,14 @@ runs/baseline-constantlr/checkpoint-360/tokenizer.json filter=lfs diff=lfs merge
runs/baseline-constantlr/checkpoint-3600/tokenizer.json filter=lfs diff=lfs merge=lfs -text
runs/baseline-constantlr/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lfs -text
runs/baseline-constantlr/checkpoint-720/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-ssl/checkpoint-1080/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-ssl/checkpoint-1440/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-ssl/checkpoint-1800/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-ssl/checkpoint-2160/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-ssl/checkpoint-2520/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-ssl/checkpoint-2880/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-ssl/checkpoint-3240/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-ssl/checkpoint-360/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-ssl/checkpoint-3600/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-ssl/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-ssl/checkpoint-720/tokenizer.json filter=lfs diff=lfs merge=lfs -text
diff --git a/runs/l2r90-ssl/checkpoint-1080/chat_template.jinja b/runs/l2r90-ssl/checkpoint-1080/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1080/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-ssl/checkpoint-1080/config.json b/runs/l2r90-ssl/checkpoint-1080/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1080/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-ssl/checkpoint-1080/generation_config.json b/runs/l2r90-ssl/checkpoint-1080/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1080/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-ssl/checkpoint-1080/model.safetensors b/runs/l2r90-ssl/checkpoint-1080/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..998e9e0cb499518445550bfc15dc537959a68281
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1080/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:24cf0796f227bf28eeb0c7d88f036e96c67f6aef0076a93210e2fb717a9caa41
+size 583356232
diff --git a/runs/l2r90-ssl/checkpoint-1080/optimizer.pt b/runs/l2r90-ssl/checkpoint-1080/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..fc72ca8df0823eadd5a8a3f3882bac51909c3942
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1080/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:45714834e797c76e5abd2c03a2a25967ea1d89706b50e602251996513c77936d
+size 1166827898
diff --git a/runs/l2r90-ssl/checkpoint-1080/rng_state_0.pth b/runs/l2r90-ssl/checkpoint-1080/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..160f7ae1cb1606549bb698b305ff378b278ef22e
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1080/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5f0d1a15fe1b6cfacdd30b06f0d68190dda9ff32b569eca1fe544ac9ab4f199b
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-1080/rng_state_1.pth b/runs/l2r90-ssl/checkpoint-1080/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..f6b2c33e0836c72d73d69110cd09ef4ad0592710
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1080/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:74058d7903c987c8077a57fa316ecf9556b937704686301014c187fcf601afa3
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-1080/scheduler.pt b/runs/l2r90-ssl/checkpoint-1080/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..6460bf7d5e8aadff08654c5c446813d18e721d62
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1080/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bd8f1c3c3cf1427d3265ecb044cf76510643402bb89ce35df8f605375b066d14
+size 1064
diff --git a/runs/l2r90-ssl/checkpoint-1080/tokenizer.json b/runs/l2r90-ssl/checkpoint-1080/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1080/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-ssl/checkpoint-1080/tokenizer_config.json b/runs/l2r90-ssl/checkpoint-1080/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1080/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-1080/trainer_state.json b/runs/l2r90-ssl/checkpoint-1080/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..69a171f58d0848d3b58a1fd2baeeec2fecff3254
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1080/trainer_state.json
@@ -0,0 +1,7698 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 15.0,
+ "eval_steps": 500,
+ "global_step": 1080,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.2588357925415039,
+ "learning_rate": 0.0,
+ "loss": 3.9982247352600098,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.25455471873283386,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.998683452606201,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.2555367052555084,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.9821574687957764,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.2383691370487213,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.9531917572021484,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2023179680109024,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.9148876667022705,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.17691479623317719,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.875638961791992,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.14372293651103973,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.836796283721924,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1074419692158699,
+ "learning_rate": 4.2e-05,
+ "loss": 3.8036489486694336,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.08204353600740433,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.774808406829834,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.0632256343960762,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.7503671646118164,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.047985248267650604,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.7345473766326904,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.03675895184278488,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.721525192260742,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.02908993512392044,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.713454246520996,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.02443614974617958,
+ "learning_rate": 7.8e-05,
+ "loss": 3.7078704833984375,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02109951712191105,
+ "learning_rate": 8.4e-05,
+ "loss": 3.7034401893615723,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.018616363406181335,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.7009854316711426,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.016399752348661423,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.697768211364746,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.014256286434829235,
+ "learning_rate": 0.000102,
+ "loss": 3.6974306106567383,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.013020666316151619,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.694406509399414,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.011879238300025463,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.691807270050049,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.010290366597473621,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.6937897205352783,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.009482803754508495,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.6902384757995605,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.008425266481935978,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.686815023422241,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.007676936220377684,
+ "learning_rate": 0.000138,
+ "loss": 3.685192823410034,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.007491032127290964,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.6839442253112793,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.007428477052599192,
+ "learning_rate": 0.00015,
+ "loss": 3.6783409118652344,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.006965892389416695,
+ "learning_rate": 0.000156,
+ "loss": 3.6813433170318604,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.008106694556772709,
+ "learning_rate": 0.000162,
+ "loss": 3.6741726398468018,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.0076780058443546295,
+ "learning_rate": 0.000168,
+ "loss": 3.67673921585083,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.007904314436018467,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.6724467277526855,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.008422375656664371,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.673013925552368,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00811714306473732,
+ "learning_rate": 0.000186,
+ "loss": 3.6698389053344727,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.009063836187124252,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.668990135192871,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.01878332532942295,
+ "learning_rate": 0.000198,
+ "loss": 3.6658849716186523,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.01812782697379589,
+ "learning_rate": 0.000204,
+ "loss": 3.6697864532470703,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.020587686449289322,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.662757158279419,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.01515690516680479,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.6558732986450195,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.02510508894920349,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.6495113372802734,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.020487826317548752,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.638803243637085,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.017446046695113182,
+ "learning_rate": 0.000234,
+ "loss": 3.633307456970215,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.030667584389448166,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.6277506351470947,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.023789944127202034,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.6189804077148438,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.02660694345831871,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.610522747039795,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.023814303800463676,
+ "learning_rate": 0.000258,
+ "loss": 3.615910291671753,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.021863561123609543,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.612008571624756,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03705216199159622,
+ "learning_rate": 0.00027,
+ "loss": 3.6066887378692627,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.024143919348716736,
+ "learning_rate": 0.000276,
+ "loss": 3.5987331867218018,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.03251276910305023,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.595367431640625,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06619632989168167,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.59018611907959,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.042714253067970276,
+ "learning_rate": 0.000294,
+ "loss": 3.58981990814209,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.018893254920840263,
+ "learning_rate": 0.0003,
+ "loss": 3.578472375869751,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0749685987830162,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.5828654766082764,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.12142857164144516,
+ "learning_rate": 0.000312,
+ "loss": 3.5840892791748047,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.06944665312767029,
+ "learning_rate": 0.000318,
+ "loss": 3.580961227416992,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.04545000568032265,
+ "learning_rate": 0.000324,
+ "loss": 3.5674569606781006,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.061073049902915955,
+ "learning_rate": 0.00033,
+ "loss": 3.569991111755371,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.04133498668670654,
+ "learning_rate": 0.000336,
+ "loss": 3.5657706260681152,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.055982090532779694,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.559979200363159,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.05259208381175995,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.558039426803589,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.029976001009345055,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.5491275787353516,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.04998327046632767,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.5406670570373535,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.03230690956115723,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.5391530990600586,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.06102178618311882,
+ "learning_rate": 0.000372,
+ "loss": 3.5311684608459473,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.031715717166662216,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.518899440765381,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.06841139495372772,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.514200210571289,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.10323675721883774,
+ "learning_rate": 0.00039,
+ "loss": 3.513646125793457,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.1448228508234024,
+ "learning_rate": 0.000396,
+ "loss": 3.5110087394714355,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.12129563093185425,
+ "learning_rate": 0.000402,
+ "loss": 3.501497507095337,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.030224842950701714,
+ "learning_rate": 0.000408,
+ "loss": 3.4870352745056152,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.07492101192474365,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.482234001159668,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.03942760080099106,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.4746525287628174,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.04967200756072998,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.604069232940674,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8671882152557373,
+ "eval_runtime": 42.6602,
+ "eval_samples_per_second": 57.243,
+ "eval_steps_per_second": 0.469,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.05385325849056244,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.462052822113037,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.04870929941534996,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.459104537963867,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.08701689541339874,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.4533209800720215,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.12219952791929245,
+ "learning_rate": 0.00045,
+ "loss": 3.4558796882629395,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.0823654904961586,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.443049192428589,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.12765446305274963,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.4373486042022705,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.14215047657489777,
+ "learning_rate": 0.000468,
+ "loss": 3.449429988861084,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.03649887070059776,
+ "learning_rate": 0.000474,
+ "loss": 3.42877197265625,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09551949054002762,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.4352874755859375,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.030483486130833626,
+ "learning_rate": 0.000486,
+ "loss": 3.4125874042510986,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.0745147094130516,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.4153835773468018,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.06620422005653381,
+ "learning_rate": 0.000498,
+ "loss": 3.4053854942321777,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.06098725646734238,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.404346466064453,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.047781601548194885,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.3991761207580566,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.04409855231642723,
+ "learning_rate": 0.000516,
+ "loss": 3.396144390106201,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.04355994611978531,
+ "learning_rate": 0.000522,
+ "loss": 3.3928911685943604,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.0670008510351181,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.3831443786621094,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.0812622532248497,
+ "learning_rate": 0.000534,
+ "loss": 3.382413148880005,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.0632157102227211,
+ "learning_rate": 0.00054,
+ "loss": 3.3784892559051514,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.028686825186014175,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.3673207759857178,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.034818850457668304,
+ "learning_rate": 0.000552,
+ "loss": 3.3648972511291504,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.04509204626083374,
+ "learning_rate": 0.000558,
+ "loss": 3.35561203956604,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.054211195558309555,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.3612167835235596,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.03752049803733826,
+ "learning_rate": 0.00057,
+ "loss": 3.357103109359741,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.03620241954922676,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.3469700813293457,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.02454194985330105,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.3429036140441895,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.041030462831258774,
+ "learning_rate": 0.000588,
+ "loss": 3.338247299194336,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.03266424685716629,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.332688331604004,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.04730091616511345,
+ "learning_rate": 0.0006,
+ "loss": 3.335902452468872,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.065738245844841,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.3289852142333984,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.09062394499778748,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.3278567790985107,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.1482805460691452,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.332685947418213,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.20066532492637634,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.363180637359619,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.11500677466392517,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.3308753967285156,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.11754138022661209,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.3383703231811523,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.09672239422798157,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.3267624378204346,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.07190799713134766,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.317737579345703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.0815449133515358,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.3198161125183105,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09612327069044113,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.315814971923828,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.14795289933681488,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.317605495452881,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06667589396238327,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.3127357959747314,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.058537885546684265,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.319807767868042,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.06428155303001404,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.297541618347168,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.03881775587797165,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.297539710998535,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.061801470816135406,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.290297508239746,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.07642851769924164,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.284428596496582,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08123870939016342,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.2820234298706055,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07174684852361679,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.2697296142578125,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.04030010476708412,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.264328956604004,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.05662153288722038,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.2699337005615234,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.03761111572384834,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.255704879760742,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.043027132749557495,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.2579030990600586,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.044409338384866714,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.248547077178955,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.0239730142056942,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.249307632446289,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03628705441951752,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.2451741695404053,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.029564054682850838,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.238445997238159,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.032036975026130676,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.2320892810821533,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.030085794627666473,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.231468915939331,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.025232303887605667,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.232199192047119,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.027053389698266983,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.210745334625244,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.028464114293456078,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.210076332092285,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04616783559322357,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.2103490829467773,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.039081938564777374,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.202662467956543,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.03428936004638672,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.1997475624084473,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.04351917654275894,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.19844913482666,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.0840945914387703,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.2013347148895264,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.36558300256729126,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.26310396194458,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.09432770311832428,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.2881736755371094,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.061970267444849014,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.2840280532836914,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.05112256482243538,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.2632040977478027,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06797203421592712,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.254606246948242,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.051026541739702225,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.429918050765991,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.8117228746414185,
+ "eval_runtime": 42.8572,
+ "eval_samples_per_second": 56.98,
+ "eval_steps_per_second": 0.467,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.0608622208237648,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.2348179817199707,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.05581629276275635,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.22896146774292,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.09162136912345886,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.2343664169311523,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.13750384747982025,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.2507717609405518,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.0872517079114914,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.2440056800842285,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.07161762565374374,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.229494571685791,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04082738608121872,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.2183239459991455,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.04972497373819351,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.218075752258301,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.049254074692726135,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.213379383087158,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.056671254336833954,
+ "learning_rate": 0.000599663984012462,
+ "loss": 3.209568977355957,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.05541909113526344,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.207760810852051,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.03819117695093155,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 3.199732780456543,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.0529443584382534,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.1945419311523438,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.043010979890823364,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.185936212539673,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.07311078161001205,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.189894914627075,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06110258400440216,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 3.1911773681640625,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.06880149245262146,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 3.187437057495117,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.051326874643564224,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 3.1826558113098145,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.03363822400569916,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 3.175527572631836,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.10268989205360413,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 3.172367572784424,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.09723010659217834,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 3.1873984336853027,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.04129317030310631,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 3.165799617767334,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.05292229354381561,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 3.171780824661255,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.03873143345117569,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 3.1680874824523926,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.047018859535455704,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 3.161402940750122,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.03942955285310745,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 3.1530532836914062,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.039801858365535736,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 3.1465344429016113,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.03795155510306358,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 3.1364030838012695,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.0342850424349308,
+ "learning_rate": 0.000599379982511273,
+ "loss": 3.1391525268554688,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.03243280574679375,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 3.1331253051757812,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.029679352417588234,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 3.1346120834350586,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.030230293050408363,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 3.132615804672241,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.02845083735883236,
+ "learning_rate": 0.000599309205504528,
+ "loss": 3.1242547035217285,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.03208369016647339,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 3.119110584259033,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.040268607437610626,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 3.113833427429199,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.06951889395713806,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 3.1186914443969727,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.08879823237657547,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 3.1140987873077393,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07374681532382965,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 3.1147828102111816,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.038190409541130066,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 3.1161253452301025,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.03297388181090355,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 3.1006083488464355,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.05905177444219589,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 3.106564521789551,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.037567321211099625,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 3.0884456634521484,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.02725658379495144,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 3.09000825881958,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.04272424429655075,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 3.090528964996338,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.027587950229644775,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 3.085131883621216,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.03282274678349495,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 3.0880112648010254,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.04064026474952698,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 3.079259157180786,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.023374097421765327,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 3.0747299194335938,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.025003870949149132,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 3.0721259117126465,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.030445197597146034,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 3.0733323097229004,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.025935636833310127,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 3.070709228515625,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.021337421610951424,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 3.0627551078796387,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.040481697767972946,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 3.0529394149780273,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.047608256340026855,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 3.041020631790161,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.03356955572962761,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 3.0532307624816895,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.02101682312786579,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 3.0492985248565674,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.030098581686615944,
+ "learning_rate": 0.000598804365316586,
+ "loss": 3.0527801513671875,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.040166422724723816,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 3.0505762100219727,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.051579318940639496,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 3.0422110557556152,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.06485334783792496,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 3.0435101985931396,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.055590901523828506,
+ "learning_rate": 0.000598706872384039,
+ "loss": 3.0448739528656006,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.03183475881814957,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 3.0416858196258545,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.03995629772543907,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 3.0255980491638184,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.03907827287912369,
+ "learning_rate": 0.000598631251036868,
+ "loss": 3.032339096069336,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.027681387960910797,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 3.025439739227295,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.03470184653997421,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 3.02883243560791,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.031407181173563004,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 3.0289883613586426,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.0371522456407547,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 3.013608932495117,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.03651345893740654,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 3.0120253562927246,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.06004595756530762,
+ "learning_rate": 0.000598473577921366,
+ "loss": 3.017958402633667,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.09979518502950668,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 3.023041248321533,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.10138462483882904,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.275729179382324,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7571707367897034,
+ "eval_runtime": 42.8283,
+ "eval_samples_per_second": 57.018,
+ "eval_steps_per_second": 0.467,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.04982660710811615,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 3.016209602355957,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.06532539427280426,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 3.015043020248413,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.04369623214006424,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 3.0096778869628906,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.05180002748966217,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 3.0121543407440186,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.0440233014523983,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.9937305450439453,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.05645671859383583,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 3.0023865699768066,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.03990716487169266,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 3.0003232955932617,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.03818608075380325,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.994137763977051,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.035388972610235214,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.993170738220215,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.03162284567952156,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.9922614097595215,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.029497336596250534,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.9852986335754395,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.02586926706135273,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.9798672199249268,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.027355050668120384,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.985158920288086,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.0341978520154953,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.982581615447998,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.030727429315447807,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.9701411724090576,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.04429870471358299,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.9789254665374756,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.05775132030248642,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.9758143424987793,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.05635962635278702,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.981480121612549,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.03639192506670952,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.968752861022949,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.028748342767357826,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.9584972858428955,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.035653006285429,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.967545747756958,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.03697885945439339,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.9578099250793457,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.03518654406070709,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.9680328369140625,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.034762222319841385,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.9670262336730957,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03813016042113304,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.957522392272949,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.04372904822230339,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.9540514945983887,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.06646446138620377,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.951028823852539,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.07831092923879623,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.95898699760437,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.05779588967561722,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.965884208679199,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.04414296895265579,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.9510560035705566,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.04797815904021263,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.955803394317627,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.04482145980000496,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.954059362411499,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.037104446440935135,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.94792103767395,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.03616701066493988,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.9500932693481445,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.03882797807455063,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.9558305740356445,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.04356204345822334,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.9237029552459717,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.04535257816314697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.925853729248047,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.04519587755203247,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.924984931945801,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.055922407656908035,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.932561159133911,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.06330762803554535,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.9388322830200195,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.0517115443944931,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.9272942543029785,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.057369694113731384,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.9201431274414062,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.05687880516052246,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.922886371612549,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.055525410920381546,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.9211201667785645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04503652825951576,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.9164583683013916,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.03385302051901817,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.9145395755767822,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03564384952187538,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.916003704071045,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.04188712686300278,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.915839433670044,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03657370060682297,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.903120279312134,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.03126626834273338,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.895824432373047,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.04196290299296379,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.9093422889709473,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.1706087589263916,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.9183316230773926,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.05810421332716942,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.911245107650757,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.061330314725637436,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.9159231185913086,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.06062281131744385,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.9211535453796387,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.05590800568461418,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.9140496253967285,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.04496239125728607,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.9040920734405518,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.03835335746407509,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.9026548862457275,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.05137286335229874,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.889634132385254,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.04091101884841919,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.886838674545288,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.04076292738318443,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.892918109893799,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.045437734574079514,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.893305778503418,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.041114673018455505,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.88517165184021,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.03600773215293884,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.8779795169830322,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.04311220720410347,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.87247896194458,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.04818756878376007,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.889103412628174,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.053460877388715744,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.8899683952331543,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.06763825565576553,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.8823537826538086,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.07304801046848297,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.877340316772461,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.061337269842624664,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.8820111751556396,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.04660911113023758,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.868886709213257,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.053929202258586884,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.159337043762207,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.721762478351593,
+ "eval_runtime": 42.7977,
+ "eval_samples_per_second": 57.059,
+ "eval_steps_per_second": 0.467,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.04259369149804115,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.859579086303711,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.044425517320632935,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.8681139945983887,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.04301869496703148,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.853911876678467,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.04574296995997429,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.8724427223205566,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.050247322767972946,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.8528172969818115,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.046683840453624725,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.853405714035034,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.05966169387102127,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.8605127334594727,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.08935574442148209,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.862339973449707,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.08019179105758667,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.8655967712402344,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.04753892868757248,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.8657939434051514,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06868117302656174,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.861976146697998,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.07040470093488693,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.8562135696411133,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.05683179199695587,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.8576629161834717,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.04918001592159271,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.861943244934082,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.04921915754675865,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.8495378494262695,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.037651967257261276,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.8428618907928467,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.03522057458758354,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.8372769355773926,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.03111186809837818,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.8377151489257812,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.03906773030757904,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.846196174621582,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.02546556107699871,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.8288016319274902,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.03119821846485138,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.8334617614746094,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.030583377927541733,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.8209362030029297,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.029693076387047768,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.8267486095428467,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.028978517279028893,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.819244861602783,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.027871523052453995,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.8289647102355957,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.029287064447999,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.8235747814178467,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.024990517646074295,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.819016218185425,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.030980689451098442,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.828155040740967,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03721801936626434,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.813418388366699,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03738148882985115,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.808328866958618,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.03580411896109581,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.812070846557617,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.039036594331264496,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.807159423828125,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.044666968286037445,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.799992561340332,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.055589817464351654,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.804142951965332,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.06139020994305611,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.8182730674743652,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.07769318670034409,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.8166518211364746,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.06521332263946533,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.8142249584198,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.09083619713783264,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.8166933059692383,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.09737467020750046,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.8111555576324463,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.07148188352584839,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.8168413639068604,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.055516090244054794,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.8183553218841553,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.05338391661643982,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.8127028942108154,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.040785472840070724,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.8124775886535645,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.037135496735572815,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.8065357208251953,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.032455578446388245,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.798459053039551,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.03465662896633148,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.798649549484253,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.032026033848524094,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.7906997203826904,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.026019956916570663,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.791969060897827,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.030298635363578796,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.784593105316162,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.028147678822278976,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.782454013824463,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.023933131247758865,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.787851333618164,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.026789527386426926,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.788067579269409,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.024554884061217308,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.777033567428589,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.028679484501481056,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.786342144012451,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.026410218328237534,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.7790610790252686,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.027413051575422287,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.778944969177246,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.02908242493867874,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.763850212097168,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.029632994905114174,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.7672982215881348,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.027458716183900833,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.7702178955078125,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.03263958916068077,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.7581803798675537,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04427146911621094,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.768141508102417,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.07142394036054611,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.75819730758667,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.08833584934473038,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.782716751098633,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.08168833702802658,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.7763240337371826,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07253874838352203,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.7717690467834473,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.08186902105808258,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.767303466796875,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.05854064226150513,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.761544704437256,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.04602917283773422,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.7669177055358887,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.03920970857143402,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.766085624694824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.04551195353269577,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.7623450756073,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.040861159563064575,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.7471840381622314,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.026907291263341904,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 2.0710244178771973,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6934371590614319,
+ "eval_runtime": 42.6015,
+ "eval_samples_per_second": 57.322,
+ "eval_steps_per_second": 0.469,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.03801899403333664,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.739086627960205,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.03707238286733627,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.741638422012329,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.0387551411986351,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.740777015686035,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.028010359033942223,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.7353010177612305,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.026474518701434135,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.73842453956604,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.02920856513082981,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.736295461654663,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.026074763387441635,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.7295751571655273,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.026862047612667084,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.723040819168091,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.026300370693206787,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.7199485301971436,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.026030950248241425,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.7276217937469482,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.028342435136437416,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.7193045616149902,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.032135091722011566,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.7143118381500244,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.03068603202700615,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.7191050052642822,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.035828713327646255,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.718085765838623,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.03343002125620842,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.707131862640381,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.035051796585321426,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.7091879844665527,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.03395913913846016,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6980862617492676,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.034198079258203506,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.699704647064209,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.0322839580476284,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.702791213989258,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.041531339287757874,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.7018051147460938,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.040806882083415985,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.7096598148345947,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.04156940430402756,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6956775188446045,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.05608797073364258,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.70127272605896,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0729481428861618,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.708817720413208,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.0905836969614029,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.718085289001465,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.08345075696706772,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.7235910892486572,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.08599896728992462,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.7178783416748047,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.07301691174507141,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.714029312133789,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.08231015503406525,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.7196826934814453,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.08286412060260773,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.7176082134246826,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.08590883016586304,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.7242164611816406,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.07531397044658661,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.7225024700164795,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.04711190238595009,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.695887327194214,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.05076327919960022,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.7029781341552734,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.04495669901371002,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.7057900428771973,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.04312257468700409,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.697537899017334,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.043580975383520126,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.67315673828125,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.03360886126756668,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.6883413791656494,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.030360296368598938,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.700927257537842,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.0310810673981905,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.6917803287506104,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.02703232690691948,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.6667580604553223,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.02756045199930668,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.6702821254730225,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.02611083723604679,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.669790744781494,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.02615324780344963,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.6672821044921875,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.026000119745731354,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.6671881675720215,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.02882794849574566,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.67586088180542,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03327213600277901,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.6618690490722656,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.032920945435762405,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.665444850921631,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.02812962979078293,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.6680707931518555,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.025886602699756622,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.652898073196411,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.025222789496183395,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.6560399532318115,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.023131893947720528,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.66127347946167,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.02513662725687027,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.6463732719421387,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.027243830263614655,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.6507630348205566,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.025437965989112854,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.6485800743103027,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.02770340070128441,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.657241106033325,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.03594067320227623,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.6281137466430664,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04486451297998428,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.6340909004211426,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.05102410912513733,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.6463394165039062,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04514491558074951,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.636554002761841,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04469453915953636,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.6453680992126465,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.05421523004770279,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.646010398864746,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04779462888836861,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.6315622329711914,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.06697724014520645,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.6416730880737305,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.07520416378974915,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.641632318496704,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.055692676454782486,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.6289877891540527,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.04539895057678223,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.6363251209259033,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.04829065501689911,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.641608476638794,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.041314542293548584,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.6269643306732178,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.04100198298692703,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.6243174076080322,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.04105962812900543,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.626585006713867,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.03172122687101364,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9550899267196655,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6614077091217041,
+ "eval_runtime": 42.9009,
+ "eval_samples_per_second": 56.922,
+ "eval_steps_per_second": 0.466,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.03258531913161278,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.612126588821411,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.040250636637210846,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.6025328636169434,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.034234337508678436,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.6085972785949707,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.03635786101222038,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.6126222610473633,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.039404451847076416,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.6035327911376953,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.038504838943481445,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5985798835754395,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.03744236007332802,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5873115062713623,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.034184787422418594,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.5899314880371094,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.03766429424285889,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.602734327316284,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.03741588443517685,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.592207193374634,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.03187853842973709,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5850372314453125,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.038447409868240356,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.592654228210449,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.04046263173222542,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.5881710052490234,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.03872602432966232,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5881471633911133,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.03960775211453438,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5826945304870605,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.033132150769233704,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.584038734436035,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.029504230245947838,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.584357261657715,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.03665253147482872,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5677168369293213,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.037208545953035355,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.560795783996582,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.038244519382715225,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.591458559036255,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.043307431042194366,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.5642588138580322,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.055119726806879044,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.57377290725708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.053323689848184586,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5767312049865723,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.0502864345908165,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.562190055847168,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.05568842589855194,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.565556049346924,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.061149828135967255,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5883288383483887,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.059933699667453766,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.5710184574127197,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.06041933596134186,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.572603225708008,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.05482695251703262,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.5667271614074707,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.04712968319654465,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.554041624069214,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.04842432215809822,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.5660672187805176,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.04273047298192978,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.554839849472046,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.05356749892234802,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.5521483421325684,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.06062676012516022,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.5551886558532715,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.05431462824344635,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5654542446136475,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.0505850687623024,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5648036003112793,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.04696856811642647,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.5559377670288086,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.042585279792547226,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.5660319328308105,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.04026484116911888,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.551206588745117,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03382421284914017,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.556361198425293,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.031177924945950508,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.5413894653320312,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.028217196464538574,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.5600497722625732,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.027062954381108284,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5447568893432617,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.03001212701201439,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.540614604949951,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.028869854286313057,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.544583797454834,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.02869918756186962,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.5471596717834473,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03359716013073921,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5416982173919678,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.04000182822346687,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.534921407699585,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.043030042201280594,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5368564128875732,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.04329848289489746,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.539764881134033,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.03740967437624931,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.5246665477752686,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.03733935207128525,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.529362916946411,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.039609868079423904,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.542544364929199,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.03937261551618576,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5366649627685547,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.03261927142739296,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.5250535011291504,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.03904294595122337,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.520721435546875,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.03913239762187004,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.5122013092041016,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.03610851988196373,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.5212767124176025,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.03944548964500427,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.537186622619629,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04762452840805054,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.5204620361328125,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04427091404795647,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5286149978637695,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04366641491651535,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.5398361682891846,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.04644155502319336,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.519331932067871,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.04372382536530495,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.531022787094116,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.04404324293136597,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.521758794784546,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.042115990072488785,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.5146074295043945,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.03796013817191124,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.5040054321289062,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.04327977076172829,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.5209312438964844,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04150385782122612,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.515871286392212,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.04465346783399582,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.5247764587402344,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.05690346285700798,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.501911163330078,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04907930642366409,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8968371152877808,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6358620524406433,
+ "eval_runtime": 42.9991,
+ "eval_samples_per_second": 56.792,
+ "eval_steps_per_second": 0.465,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.03755255416035652,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.486576795578003,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.040234558284282684,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.501087188720703,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.041416723281145096,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4942233562469482,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.039655014872550964,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.503453254699707,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.041707683354616165,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.490978717803955,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.04545537009835243,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4757118225097656,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.04328389838337898,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.504516839981079,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.043312299996614456,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.4920215606689453,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.046466976404190063,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.494760513305664,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.041762981563806534,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4875552654266357,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.03713781014084816,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.495652675628662,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.040603458881378174,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4821407794952393,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.037969645112752914,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.484981060028076,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.03652074560523033,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4736409187316895,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.03783264383673668,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.483346700668335,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.03404324874281883,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.48856782913208,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.03294725716114044,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.483706474304199,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.0347001850605011,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.474888324737549,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.0359107106924057,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.478990077972412,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.03738103806972504,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.4709420204162598,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.03722188249230385,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.465726852416992,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.034907322376966476,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.482231855392456,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.035695310682058334,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4785828590393066,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.04009179025888443,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.4795875549316406,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.038996778428554535,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.478696823120117,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04082407429814339,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.465200901031494,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04856136813759804,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4605681896209717,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.05671004578471184,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.47277569770813,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.0481606163084507,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.476151466369629,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.037689510732889175,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.466142177581787,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.05005146935582161,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.4798035621643066,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.04745829105377197,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.46163272857666,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.04638158157467842,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4728071689605713,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.04530330374836922,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.4621877670288086,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.03971488028764725,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.474224090576172,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.04521876201033592,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.46122670173645,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03923020884394646,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.4603679180145264,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03986494988203049,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.4589664936065674,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.04280020296573639,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.4624319076538086,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.03930043429136276,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.47080135345459,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.0314876064658165,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.458728790283203,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.03189660236239433,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4467148780822754,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03264275938272476,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4613699913024902,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03021484799683094,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.4535279273986816,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.030740290880203247,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.4621875286102295,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03096810355782509,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.458500385284424,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03105205111205578,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.461958646774292,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.03330068290233612,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.449777364730835,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.032831475138664246,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.4616191387176514,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.0387776717543602,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.445256471633911,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.041803423315286636,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.4545960426330566,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.04148292914032936,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.433408737182617,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.04648579657077789,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.449134349822998,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.04186409339308739,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.4449565410614014,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.038569554686546326,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.443675994873047,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03908300772309303,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4626526832580566,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.038897737860679626,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.438591957092285,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.0360480472445488,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.436077117919922,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0429830439388752,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4384970664978027,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.04465422406792641,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.458029270172119,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.037096526473760605,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.436816930770874,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.047823466360569,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.44439959526062,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.05929310992360115,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.4599215984344482,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.053512319922447205,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4387903213500977,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.042625587433576584,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4359588623046875,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.056042976677417755,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.4441447257995605,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05813189223408699,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4399144649505615,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.06313478201627731,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.453447103500366,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.05281798169016838,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.4428963661193848,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.03991910442709923,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4359219074249268,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.047981832176446915,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4506607055664062,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.03646523132920265,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8225739002227783,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6196661591529846,
+ "eval_runtime": 43.8106,
+ "eval_samples_per_second": 55.74,
+ "eval_steps_per_second": 0.457,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.03673036769032478,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.4227209091186523,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.044547371566295624,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.4073169231414795,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.04089285805821419,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4279446601867676,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.03538106009364128,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.4234392642974854,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.03152978792786598,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4214606285095215,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.03324202075600624,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.4085447788238525,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.03352442383766174,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.417776107788086,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.03582802042365074,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.411144256591797,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.0351114459335804,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.405872344970703,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.038898173719644547,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4029338359832764,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.03926656022667885,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4100089073181152,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.04285634309053421,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.410360336303711,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.04179581627249718,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4093637466430664,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.03356701880693436,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4118056297302246,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.038140635937452316,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.4154677391052246,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.04468780383467674,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416846513748169,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.062421441078186035,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4157638549804688,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06317313015460968,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.418593406677246,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.049498945474624634,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4204845428466797,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.046322405338287354,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.403489589691162,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.053081002086400986,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.4148740768432617,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05222945287823677,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.416757583618164,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04893574118614197,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4241795539855957,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04951956123113632,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.401386260986328,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04019852355122566,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.4168622493743896,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.03625848889350891,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4026846885681152,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.03464345261454582,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4021215438842773,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.03750818222761154,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4113571643829346,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.0360100194811821,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.3977622985839844,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.03464623540639877,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.413273334503174,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.03193575143814087,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.423659324645996,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.02710454724729061,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.396085262298584,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.02830432914197445,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.420255422592163,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.029717663303017616,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.3958120346069336,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.030005933716893196,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3908543586730957,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.029054716229438782,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.3873138427734375,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.031556472182273865,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3892760276794434,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03518872708082199,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4113478660583496,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.03471249341964722,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.384857654571533,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03223155066370964,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3840765953063965,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.030012644827365875,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.388247489929199,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.03162285313010216,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.3987512588500977,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.03700442612171173,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.3993430137634277,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.03710822016000748,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.387359380722046,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.028411181643605232,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3925249576568604,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031246855854988098,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.38845157623291,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.034664880484342575,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.392436981201172,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.03354701027274132,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.390890121459961,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.03527764603495598,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.3954997062683105,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03325049206614494,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3847177028656006,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.028021875768899918,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.385910987854004,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.02964954823255539,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3787484169006348,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.03068520873785019,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.400784969329834,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03416873514652252,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.391033411026001,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.03534875437617302,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.3796474933624268,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.03732474148273468,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3831934928894043,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.03648678958415985,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3785600662231445,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03801294043660164,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.387587070465088,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.04186025261878967,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.3893089294433594,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.04416101053357124,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3852195739746094,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.04137178510427475,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.391308307647705,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.05010323226451874,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.3906192779541016,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.049587685614824295,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.3846192359924316,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.04125971347093582,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.384507179260254,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.039676666259765625,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.38167142868042,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.038775283843278885,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.3789167404174805,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.03634221479296684,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.383060932159424,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03464275971055031,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.388552665710449,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.031778786331415176,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.3903298377990723,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.027904309332370758,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.377493381500244,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.028964119032025337,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.381959915161133,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.0233962032943964,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7810559272766113,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6070130467414856,
+ "eval_runtime": 50.871,
+ "eval_samples_per_second": 48.004,
+ "eval_steps_per_second": 0.393,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.030209315940737724,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.3569841384887695,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.04112662002444267,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.376680374145508,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.055858466774225235,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.3621644973754883,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.06746986508369446,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3607563972473145,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.06736797094345093,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3548707962036133,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.0562627948820591,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.364654541015625,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.051833927631378174,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.3748884201049805,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.04824110120534897,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.3502087593078613,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.05518793687224388,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3687515258789062,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.06564056873321533,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.367131233215332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.0605909489095211,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.3589868545532227,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.05744732916355133,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.370117664337158,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.06430432945489883,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.3703484535217285,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.06186915934085846,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3720293045043945,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.05977998673915863,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.3512699604034424,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.06413097679615021,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.377257823944092,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.053935080766677856,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.350717544555664,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.049110863357782364,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.3688902854919434,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.04313219338655472,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3374459743499756,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.039556123316287994,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3482556343078613,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.03796547278761864,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3547937870025635,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.03646104037761688,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.344686269760132,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.03355357423424721,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.3588008880615234,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.03004707582294941,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.352043628692627,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.030186176300048828,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3500709533691406,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.029748762026429176,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.3523058891296387,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.03238696977496147,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3635129928588867,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.032599564641714096,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.346269130706787,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.028626957908272743,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.357151746749878,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.02909751422703266,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3557677268981934,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.029951099306344986,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.34753680229187,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.029768310487270355,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3408985137939453,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.03005973994731903,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.362542152404785,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03360462561249733,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.352529287338257,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.034002047032117844,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.3334460258483887,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.03292316198348999,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3446991443634033,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.0337117575109005,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.350605010986328,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03428212180733681,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.357408046722412,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.03169601410627365,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.349088191986084,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.03389411047101021,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.361576795578003,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.036497049033641815,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.35420560836792,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03574785590171814,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3678526878356934,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.030757993459701538,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3520517349243164,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.0289121363312006,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.337864637374878,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.028012607246637344,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.353999614715576,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.027558492496609688,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.342132568359375,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.02871510572731495,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3425655364990234,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03070886805653572,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3401665687561035,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.03056657873094082,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.345277786254883,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.030907969921827316,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.346494674682617,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0318334735929966,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.3390543460845947,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.029700733721256256,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.3398728370666504,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.029687359929084778,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.3474364280700684,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.027234600856900215,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3457484245300293,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.02823043428361416,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.337700366973877,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03002646192908287,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.335954189300537,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.027246717363595963,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.330207347869873,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.029703574255108833,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.328280448913574,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.028796236962080002,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3448386192321777,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.027944380417466164,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.320234537124634,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.030626650899648666,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.334698438644409,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.029164662584662437,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3450326919555664,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03034045360982418,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.3420588970184326,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.02856968529522419,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3396072387695312,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.02892562374472618,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.340027332305908,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.02938242256641388,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.33550763130188,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.029261089861392975,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3246421813964844,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.030340535566210747,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.3465514183044434,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.033335570245981216,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.3467206954956055,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.03452218323945999,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.343702793121338,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.03326255828142166,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.337631940841675,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025433633476495743,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7402067184448242,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03175730258226395,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.62623405456543,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.04780422896146774,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.623239517211914,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.0504758283495903,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.6369194984436035,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.04481920227408409,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.57640266418457,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0440552793443203,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.630992889404297,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0503336526453495,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.627910137176514,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.053885314613580704,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.637897491455078,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.05481956899166107,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.610715866088867,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.050510745495557785,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.639884948730469,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.04579266160726547,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.590188980102539,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.04817410930991173,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.636295318603516,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.058862220495939255,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.614317417144775,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.06328638643026352,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.655928611755371,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.058752477169036865,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.652004241943359,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.06049015372991562,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.6513800621032715,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.058941151946783066,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.657688140869141,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.05542639642953873,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.637376308441162,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06024057790637016,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.63592529296875,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.06040216609835625,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.6275787353515625,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.059600673615932465,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.652995586395264,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.05379730463027954,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.597358226776123,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05355260521173477,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.6106367111206055,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.06649646162986755,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.654443740844727,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.058309994637966156,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.6869659423828125,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.05403570458292961,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.654750823974609,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04678157716989517,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.653955459594727,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04297948628664017,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.644915580749512,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.04235142841935158,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.662282943725586,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.0405130498111248,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.64569091796875,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.03351007401943207,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.653165817260742,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.03182631731033325,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.606291770935059,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.031552448868751526,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.642643928527832,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027941036969423294,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.623664379119873,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.027744563296437263,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.605886459350586,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02962736040353775,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.606762886047363,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.027251560240983963,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.620492935180664,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.027894044294953346,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.613903999328613,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.02778882533311844,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.634457588195801,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.028082096949219704,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.610980033874512,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.026210976764559746,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.612090110778809,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.026537755504250526,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.581171035766602,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.028709523379802704,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.628799915313721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.028233813121914864,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.586182117462158,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.02966046892106533,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.62320613861084,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.03198636695742607,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.6018524169921875,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.033404890447854996,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.610121726989746,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.029217196628451347,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.601312160491943,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.029335757717490196,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.620820999145508,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.031007781624794006,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.626503944396973,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.028453458100557327,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.589323997497559,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02526312880218029,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.598350524902344,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.027317512780427933,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.62388277053833,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02710605598986149,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.606523513793945,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.027491388842463493,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.586644172668457,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.027158567681908607,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.606039047241211,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.0298276599496603,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.593776702880859,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.026622479781508446,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.59510612487793,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.028238043189048767,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.5899505615234375,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.029481008648872375,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.6092071533203125,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.030138323083519936,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.600069046020508,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.02951825223863125,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.600125789642334,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.029242202639579773,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.590723037719727,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.02941613830626011,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.623342514038086,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.029764151200652122,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.631185531616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.031503792852163315,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.605094909667969,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.03184918314218521,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.592767715454102,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.03223618119955063,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.5930585861206055,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.0340263694524765,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.609347343444824,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.03245621174573898,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.598440170288086,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.03204867243766785,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.619702339172363,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.030866602435708046,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.609165191650391,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.01960863545536995,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8655829429626465,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.593852162361145,
+ "eval_runtime": 60.0971,
+ "eval_samples_per_second": 40.634,
+ "eval_steps_per_second": 0.649,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03552858531475067,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.539240837097168,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.0536109134554863,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.52609395980835,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.061411239206790924,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.536109447479248,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.06352879852056503,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.546850204467773,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.05683520436286926,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.563828468322754,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.05522637069225311,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.550929069519043,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.058990392833948135,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.554411888122559,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.057443976402282715,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.545921325683594,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.05383576825261116,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.5481414794921875,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04797440022230148,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.55556583404541,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04923664405941963,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.545513153076172,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.056191347539424896,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.536705017089844,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.06088989973068237,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.570624351501465,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.04802761599421501,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.526930809020996,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.04018320515751839,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.562432289123535,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.04280981421470642,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.559737205505371,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.04258878529071808,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.561749458312988,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.04411563277244568,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.562010765075684,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.041077401489019394,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.573984622955322,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.04125386103987694,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.550564289093018,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.041974026709795,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.552739143371582,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03689907118678093,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.540107727050781,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03586242347955704,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.554999351501465,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.036602675914764404,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.546337127685547,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.03707926720380783,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.543377876281738,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.03322125971317291,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.574325084686279,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03359243646264076,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.5613603591918945,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.0314965583384037,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.530994415283203,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.031806789338588715,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.56138801574707,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.034988485276699066,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.553149700164795,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.037627559155225754,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.555401802062988,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03723530471324921,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.52452278137207,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.036220453679561615,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.54799747467041,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03228041157126427,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.574814319610596,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.030872676521539688,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.539666652679443,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.033935241401195526,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.579252243041992,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.034338828176259995,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.538324356079102,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.03441951423883438,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.540358543395996,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03328141197562218,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.520340919494629,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03224223107099533,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.5646209716796875,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.02999890223145485,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.504915714263916,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.03014362044632435,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.516615867614746,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.029469173401594162,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.560475826263428,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.027887821197509766,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.586333751678467,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.029970254749059677,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.530396461486816,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.031648170202970505,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.554404258728027,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.0325581431388855,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.544102668762207,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.03555625304579735,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.543701171875,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.033053912222385406,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.508375644683838,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.03059076890349388,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.545750617980957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.029640182852745056,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.54715633392334,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030139951035380363,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.533708572387695,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029753828421235085,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.543603897094727,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.029481034725904465,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.534200668334961,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.028991524130105972,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.5683674812316895,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.028145765885710716,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.543053150177002,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.02999844029545784,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.5482354164123535,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03134117275476456,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.5360107421875,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.029277140274643898,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.5729827880859375,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028516866266727448,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.543548583984375,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028387080878019333,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.552616596221924,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.028977863490581512,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.5456695556640625,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.028010815382003784,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.534921646118164,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.030222605913877487,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.561526298522949,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.03140007704496384,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.500124931335449,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.029269184917211533,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.536653995513916,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.02935507707297802,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.521398544311523,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02801917865872383,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.527442455291748,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.027995768934488297,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.556139945983887,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.028441084548830986,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.555442810058594,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.029100807383656502,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.532083511352539,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020385215058922768,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.8488006591796875,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5902590751647949,
+ "eval_runtime": 60.1424,
+ "eval_samples_per_second": 40.604,
+ "eval_steps_per_second": 0.648,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03513244539499283,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.486500263214111,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.052679311484098434,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.489691734313965,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.06171489134430885,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.471555709838867,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.06164781004190445,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.486891746520996,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.06410388648509979,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.494929313659668,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06399523466825485,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.5334320068359375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.0636502280831337,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.5039381980896,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.06139931082725525,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.512765884399414,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.07460282742977142,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.477229118347168,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.075187548995018,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.484212398529053,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.07584068924188614,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.495598793029785,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.07192158699035645,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.507598400115967,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.05366869643330574,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.514532089233398,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.06397061794996262,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.512920379638672,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.06391509622335434,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.465702056884766,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.059490568935871124,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.5351881980896,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.057501111179590225,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.510270595550537,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.056947533041238785,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.509241104125977,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.050393324345350266,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.50428581237793,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.057608287781476974,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.474919319152832,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.054470717906951904,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.521528244018555,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.054405540227890015,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.510211944580078,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.051328495144844055,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.480720520019531,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.045930180698633194,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.503024101257324,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.04298274591565132,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.49046516418457,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.0406254380941391,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.544064521789551,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.03557496517896652,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.496966361999512,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.0364493802189827,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.5047502517700195,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.034521300345659256,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.485365867614746,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.03305799141526222,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.484884738922119,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.033053625375032425,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.478784561157227,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03426017984747887,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.482428073883057,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.03638783469796181,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.433219909667969,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.03597250208258629,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.522584915161133,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.03086097538471222,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.500803470611572,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.029642771929502487,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.444099426269531,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.03190716728568077,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.493034362792969,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.0312873050570488,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.487887382507324,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03201943263411522,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.479177474975586,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.033088769763708115,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.5020246505737305,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.02975946106016636,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.455426216125488,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03213953226804733,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.501615047454834,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.03262908384203911,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.485121250152588,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.031102091073989868,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.488893508911133,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.02945449948310852,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.490121841430664,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.033680472522974014,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.51833438873291,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.03277525305747986,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.480165481567383,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.03303041681647301,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.501155376434326,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.034927863627672195,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.513175010681152,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.03098093718290329,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.503798484802246,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.029513906687498093,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.49006986618042,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029630785807967186,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.522514343261719,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.030888330191373825,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.4895734786987305,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.032392196357250214,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.471953868865967,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.03215661272406578,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.48297643661499,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.03346327692270279,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.518160820007324,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.031311362981796265,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.484991550445557,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.03056488186120987,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.525470733642578,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.03043230250477791,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.510650634765625,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.02916652522981167,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.496840953826904,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.030673664063215256,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.494868278503418,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030033906921744347,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.479182720184326,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.028828797861933708,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.471983909606934,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028316056355834007,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.5049147605896,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.03158092871308327,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.449270248413086,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.03041706420481205,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.508289813995361,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.02967553399503231,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.465579032897949,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030721619725227356,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.471003532409668,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03066832385957241,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.488987445831299,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.032297998666763306,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.491018295288086,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03221398591995239,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.471237659454346,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.021120507270097733,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.8080053329467773,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5888046026229858,
+ "eval_runtime": 59.1638,
+ "eval_samples_per_second": 41.275,
+ "eval_steps_per_second": 0.659,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.03799329698085785,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.401787757873535,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.05785121023654938,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.408820629119873,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.05964238941669464,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.427263259887695,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05466582626104355,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.403415679931641,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.05613672360777855,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.413750648498535,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.05396881327033043,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.41916561126709,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.05689029023051262,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.395590782165527,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.056090328842401505,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.421253204345703,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.0473211295902729,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.385158061981201,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.04464765638113022,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.431746959686279,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.046421997249126434,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.428816795349121,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.05039866268634796,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.452417850494385,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.05491664633154869,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.447483062744141,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05866196006536484,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.434992790222168,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.05532094091176987,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.450716018676758,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.051083534955978394,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.402632713317871,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.03815016150474548,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.433640956878662,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04203395918011665,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.415888786315918,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.043729111552238464,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.423260688781738,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.039426498115062714,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.428949356079102,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.03837858885526657,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.434070587158203,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.038315922021865845,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.471246719360352,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.03963236138224602,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.4434990882873535,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04027022793889046,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.436992645263672,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.0409243181347847,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.426936149597168,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03404631093144417,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.43462610244751,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.034393735229969025,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.437409400939941,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03635180741548538,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.451286315917969,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03349733352661133,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.419209003448486,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.03238677605986595,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.456454277038574,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.034473925828933716,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.3988728523254395,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.03455805033445358,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.434884071350098,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.037486057728528976,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.466741561889648,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.03617125004529953,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.429305076599121,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.033646389842033386,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.407780170440674,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03401118144392967,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.42977237701416,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.036534134298563004,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.451094627380371,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.0384049266576767,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.414777755737305,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03460627421736717,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.420644760131836,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.033746346831321716,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.461175918579102,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.037153810262680054,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.444699287414551,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03283539041876793,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.431022644042969,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03248969465494156,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.447571277618408,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03327852487564087,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.42236328125,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.031114818528294563,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.418244361877441,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.03231207653880119,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.455004692077637,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.030702251940965652,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.483610153198242,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031027289107441902,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.4470109939575195,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.03129073604941368,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.452799320220947,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.03634301945567131,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4661664962768555,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03718913346529007,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.412006855010986,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.034018341451883316,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.436755180358887,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032502319663763046,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.452866554260254,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.032902952283620834,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.449224472045898,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.0340626984834671,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.445425510406494,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.033442508429288864,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.432218551635742,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.03489428386092186,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.4205827713012695,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.031868621706962585,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.414196968078613,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.032986633479595184,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.46116304397583,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.03348355367779732,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.424936771392822,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03174544870853424,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.434438705444336,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.03293656185269356,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.440502166748047,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.035124607384204865,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.440943717956543,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03438971936702728,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.431408882141113,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.03537093102931976,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.43867301940918,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.03267325833439827,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.462226867675781,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.030995232984423637,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.433740615844727,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.03384929150342941,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.508179664611816,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.032025814056396484,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.407378196716309,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03288578242063522,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.455531120300293,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.032061848789453506,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.452212333679199,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022320276126265526,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7567386627197266,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5878658294677734,
+ "eval_runtime": 59.5054,
+ "eval_samples_per_second": 41.038,
+ "eval_steps_per_second": 0.655,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.039114587008953094,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.347412586212158,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.05164389684796333,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.391279220581055,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055882278829813004,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.355602264404297,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.05625041201710701,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3934855461120605,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05861108750104904,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.364226341247559,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.052431970834732056,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.363334655761719,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.06147193908691406,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.384820938110352,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.06701470166444778,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.373311996459961,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.06714873015880585,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.32745361328125,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.06684907525777817,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.373660087585449,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.05738082900643349,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.362325668334961,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05306725949048996,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.351685047149658,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.05831029266119003,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.374813079833984,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05547972396016121,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.361842632293701,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.048057056963443756,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.383205413818359,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.04682513326406479,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.365846157073975,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.05001377314329147,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.362151145935059,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.05140263959765434,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.376064300537109,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.04640384018421173,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.363801002502441,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04175020009279251,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.3477983474731445,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.04134776070713997,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.337469100952148,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.042178574949502945,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.351490020751953,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.03989115729928017,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.40725040435791,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04086652770638466,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.341119766235352,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.03931201249361038,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.3786468505859375,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.03855905309319496,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.3701605796813965,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.03803952783346176,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.401841163635254,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.036908265203237534,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.410647392272949,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.0398186556994915,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.380431175231934,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03899035602807999,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.374293804168701,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.03946344554424286,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.396918773651123,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.03738604858517647,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.404016017913818,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.03579043596982956,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.396914958953857,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.037139154970645905,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.407739639282227,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.03581099584698677,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.411419868469238,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.03548453748226166,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.332690238952637,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.037561237812042236,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.381314277648926,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.035827480256557465,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.371536731719971,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03538314625620842,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.395467758178711,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.036672890186309814,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.420590400695801,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.038981080055236816,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.395296096801758,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.038234200328588486,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.415799140930176,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.038830097764730453,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.413394927978516,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03990919142961502,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.39515495300293,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.036424681544303894,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.39730978012085,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03488154709339142,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.403696060180664,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.0363851860165596,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.368661403656006,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.035461440682411194,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.396518707275391,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03647598251700401,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.440667152404785,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.03667188808321953,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.3781352043151855,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034516457468271255,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.414007186889648,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.03536032512784004,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.42198371887207,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.037454646080732346,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.409174919128418,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03632275015115738,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3547682762146,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.038922134786844254,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.39993143081665,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.03562692925333977,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.409564971923828,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.033203717321157455,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.421782493591309,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03830202296376228,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.430092811584473,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.03835884481668472,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.379505634307861,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.037440430372953415,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.403937339782715,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03732272610068321,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.440683364868164,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03325957432389259,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.3949294090271,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.035377148538827896,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.400777816772461,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.03725374862551689,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.422494411468506,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03601110726594925,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.4008941650390625,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03193027526140213,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.400361061096191,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.034382354468107224,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.408535003662109,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.03747732564806938,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.412189483642578,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03641800582408905,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.381453037261963,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.0315779373049736,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.4169769287109375,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03583509847521782,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.385746955871582,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02574268914759159,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7456769943237305,
+ "step": 1080
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 4.613803139609395e+17,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-1080/training_args.bin b/runs/l2r90-ssl/checkpoint-1080/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..103048067ef8897d0386013ed89ce6764bad5521
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1080/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f320fb44ea3a4a93198f1a6c2405447428148ce142cd94ee283a3c077a327c2d
+size 4792
diff --git a/runs/l2r90-ssl/checkpoint-1440/chat_template.jinja b/runs/l2r90-ssl/checkpoint-1440/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1440/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-ssl/checkpoint-1440/config.json b/runs/l2r90-ssl/checkpoint-1440/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1440/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-ssl/checkpoint-1440/generation_config.json b/runs/l2r90-ssl/checkpoint-1440/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1440/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-ssl/checkpoint-1440/model.safetensors b/runs/l2r90-ssl/checkpoint-1440/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..6fcf55e984de3e9ac7f260e59536277e67d9abf2
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1440/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a0f37f79f8f7496166b722ef480a483d7610460c28224b7f4296c3ffff6019d2
+size 583356232
diff --git a/runs/l2r90-ssl/checkpoint-1440/optimizer.pt b/runs/l2r90-ssl/checkpoint-1440/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..63843d4dfab4870ac9ac77cfb8f912be456e30cd
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1440/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4998c17fc0ecc6c806a67345a0eb02e3a30e47d1b744e101abc2515ae9f14c61
+size 1166827898
diff --git a/runs/l2r90-ssl/checkpoint-1440/rng_state_0.pth b/runs/l2r90-ssl/checkpoint-1440/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..e966ee38a27de32a4c7eda7c7f241e89d80c0d88
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1440/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7b6dfe0df952e9be864789885c3487b8e5cfade9f68c759414e42814ef10e5cd
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-1440/rng_state_1.pth b/runs/l2r90-ssl/checkpoint-1440/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..fabe8e923b77982f33575eef355ced13b68cc03e
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1440/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3984a09f8ba3c1f6a48b3733b965ecc6b48e3beee035ec5a8115509c64928c5c
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-1440/scheduler.pt b/runs/l2r90-ssl/checkpoint-1440/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..59b8db4a485766e1d20d11a905a37774da0cec09
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1440/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2fe49e98b5d706cd9174dc7cf5e5cd5039ca03537426b9020792a5363720022f
+size 1064
diff --git a/runs/l2r90-ssl/checkpoint-1440/tokenizer.json b/runs/l2r90-ssl/checkpoint-1440/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1440/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-ssl/checkpoint-1440/tokenizer_config.json b/runs/l2r90-ssl/checkpoint-1440/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1440/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-1440/trainer_state.json b/runs/l2r90-ssl/checkpoint-1440/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..0287891dc6185188d7269bde264e7b51401f98c3
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1440/trainer_state.json
@@ -0,0 +1,10258 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 20.0,
+ "eval_steps": 500,
+ "global_step": 1440,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.2588357925415039,
+ "learning_rate": 0.0,
+ "loss": 3.9982247352600098,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.25455471873283386,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.998683452606201,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.2555367052555084,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.9821574687957764,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.2383691370487213,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.9531917572021484,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2023179680109024,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.9148876667022705,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.17691479623317719,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.875638961791992,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.14372293651103973,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.836796283721924,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1074419692158699,
+ "learning_rate": 4.2e-05,
+ "loss": 3.8036489486694336,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.08204353600740433,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.774808406829834,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.0632256343960762,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.7503671646118164,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.047985248267650604,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.7345473766326904,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.03675895184278488,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.721525192260742,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.02908993512392044,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.713454246520996,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.02443614974617958,
+ "learning_rate": 7.8e-05,
+ "loss": 3.7078704833984375,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02109951712191105,
+ "learning_rate": 8.4e-05,
+ "loss": 3.7034401893615723,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.018616363406181335,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.7009854316711426,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.016399752348661423,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.697768211364746,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.014256286434829235,
+ "learning_rate": 0.000102,
+ "loss": 3.6974306106567383,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.013020666316151619,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.694406509399414,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.011879238300025463,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.691807270050049,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.010290366597473621,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.6937897205352783,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.009482803754508495,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.6902384757995605,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.008425266481935978,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.686815023422241,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.007676936220377684,
+ "learning_rate": 0.000138,
+ "loss": 3.685192823410034,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.007491032127290964,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.6839442253112793,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.007428477052599192,
+ "learning_rate": 0.00015,
+ "loss": 3.6783409118652344,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.006965892389416695,
+ "learning_rate": 0.000156,
+ "loss": 3.6813433170318604,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.008106694556772709,
+ "learning_rate": 0.000162,
+ "loss": 3.6741726398468018,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.0076780058443546295,
+ "learning_rate": 0.000168,
+ "loss": 3.67673921585083,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.007904314436018467,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.6724467277526855,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.008422375656664371,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.673013925552368,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00811714306473732,
+ "learning_rate": 0.000186,
+ "loss": 3.6698389053344727,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.009063836187124252,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.668990135192871,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.01878332532942295,
+ "learning_rate": 0.000198,
+ "loss": 3.6658849716186523,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.01812782697379589,
+ "learning_rate": 0.000204,
+ "loss": 3.6697864532470703,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.020587686449289322,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.662757158279419,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.01515690516680479,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.6558732986450195,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.02510508894920349,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.6495113372802734,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.020487826317548752,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.638803243637085,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.017446046695113182,
+ "learning_rate": 0.000234,
+ "loss": 3.633307456970215,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.030667584389448166,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.6277506351470947,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.023789944127202034,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.6189804077148438,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.02660694345831871,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.610522747039795,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.023814303800463676,
+ "learning_rate": 0.000258,
+ "loss": 3.615910291671753,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.021863561123609543,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.612008571624756,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03705216199159622,
+ "learning_rate": 0.00027,
+ "loss": 3.6066887378692627,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.024143919348716736,
+ "learning_rate": 0.000276,
+ "loss": 3.5987331867218018,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.03251276910305023,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.595367431640625,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06619632989168167,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.59018611907959,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.042714253067970276,
+ "learning_rate": 0.000294,
+ "loss": 3.58981990814209,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.018893254920840263,
+ "learning_rate": 0.0003,
+ "loss": 3.578472375869751,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0749685987830162,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.5828654766082764,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.12142857164144516,
+ "learning_rate": 0.000312,
+ "loss": 3.5840892791748047,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.06944665312767029,
+ "learning_rate": 0.000318,
+ "loss": 3.580961227416992,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.04545000568032265,
+ "learning_rate": 0.000324,
+ "loss": 3.5674569606781006,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.061073049902915955,
+ "learning_rate": 0.00033,
+ "loss": 3.569991111755371,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.04133498668670654,
+ "learning_rate": 0.000336,
+ "loss": 3.5657706260681152,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.055982090532779694,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.559979200363159,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.05259208381175995,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.558039426803589,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.029976001009345055,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.5491275787353516,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.04998327046632767,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.5406670570373535,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.03230690956115723,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.5391530990600586,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.06102178618311882,
+ "learning_rate": 0.000372,
+ "loss": 3.5311684608459473,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.031715717166662216,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.518899440765381,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.06841139495372772,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.514200210571289,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.10323675721883774,
+ "learning_rate": 0.00039,
+ "loss": 3.513646125793457,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.1448228508234024,
+ "learning_rate": 0.000396,
+ "loss": 3.5110087394714355,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.12129563093185425,
+ "learning_rate": 0.000402,
+ "loss": 3.501497507095337,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.030224842950701714,
+ "learning_rate": 0.000408,
+ "loss": 3.4870352745056152,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.07492101192474365,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.482234001159668,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.03942760080099106,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.4746525287628174,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.04967200756072998,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.604069232940674,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8671882152557373,
+ "eval_runtime": 42.6602,
+ "eval_samples_per_second": 57.243,
+ "eval_steps_per_second": 0.469,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.05385325849056244,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.462052822113037,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.04870929941534996,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.459104537963867,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.08701689541339874,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.4533209800720215,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.12219952791929245,
+ "learning_rate": 0.00045,
+ "loss": 3.4558796882629395,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.0823654904961586,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.443049192428589,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.12765446305274963,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.4373486042022705,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.14215047657489777,
+ "learning_rate": 0.000468,
+ "loss": 3.449429988861084,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.03649887070059776,
+ "learning_rate": 0.000474,
+ "loss": 3.42877197265625,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09551949054002762,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.4352874755859375,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.030483486130833626,
+ "learning_rate": 0.000486,
+ "loss": 3.4125874042510986,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.0745147094130516,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.4153835773468018,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.06620422005653381,
+ "learning_rate": 0.000498,
+ "loss": 3.4053854942321777,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.06098725646734238,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.404346466064453,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.047781601548194885,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.3991761207580566,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.04409855231642723,
+ "learning_rate": 0.000516,
+ "loss": 3.396144390106201,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.04355994611978531,
+ "learning_rate": 0.000522,
+ "loss": 3.3928911685943604,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.0670008510351181,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.3831443786621094,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.0812622532248497,
+ "learning_rate": 0.000534,
+ "loss": 3.382413148880005,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.0632157102227211,
+ "learning_rate": 0.00054,
+ "loss": 3.3784892559051514,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.028686825186014175,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.3673207759857178,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.034818850457668304,
+ "learning_rate": 0.000552,
+ "loss": 3.3648972511291504,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.04509204626083374,
+ "learning_rate": 0.000558,
+ "loss": 3.35561203956604,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.054211195558309555,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.3612167835235596,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.03752049803733826,
+ "learning_rate": 0.00057,
+ "loss": 3.357103109359741,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.03620241954922676,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.3469700813293457,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.02454194985330105,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.3429036140441895,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.041030462831258774,
+ "learning_rate": 0.000588,
+ "loss": 3.338247299194336,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.03266424685716629,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.332688331604004,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.04730091616511345,
+ "learning_rate": 0.0006,
+ "loss": 3.335902452468872,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.065738245844841,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.3289852142333984,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.09062394499778748,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.3278567790985107,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.1482805460691452,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.332685947418213,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.20066532492637634,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.363180637359619,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.11500677466392517,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.3308753967285156,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.11754138022661209,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.3383703231811523,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.09672239422798157,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.3267624378204346,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.07190799713134766,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.317737579345703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.0815449133515358,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.3198161125183105,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09612327069044113,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.315814971923828,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.14795289933681488,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.317605495452881,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06667589396238327,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.3127357959747314,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.058537885546684265,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.319807767868042,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.06428155303001404,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.297541618347168,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.03881775587797165,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.297539710998535,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.061801470816135406,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.290297508239746,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.07642851769924164,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.284428596496582,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08123870939016342,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.2820234298706055,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07174684852361679,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.2697296142578125,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.04030010476708412,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.264328956604004,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.05662153288722038,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.2699337005615234,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.03761111572384834,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.255704879760742,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.043027132749557495,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.2579030990600586,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.044409338384866714,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.248547077178955,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.0239730142056942,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.249307632446289,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03628705441951752,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.2451741695404053,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.029564054682850838,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.238445997238159,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.032036975026130676,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.2320892810821533,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.030085794627666473,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.231468915939331,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.025232303887605667,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.232199192047119,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.027053389698266983,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.210745334625244,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.028464114293456078,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.210076332092285,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04616783559322357,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.2103490829467773,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.039081938564777374,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.202662467956543,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.03428936004638672,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.1997475624084473,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.04351917654275894,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.19844913482666,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.0840945914387703,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.2013347148895264,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.36558300256729126,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.26310396194458,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.09432770311832428,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.2881736755371094,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.061970267444849014,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.2840280532836914,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.05112256482243538,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.2632040977478027,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06797203421592712,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.254606246948242,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.051026541739702225,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.429918050765991,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.8117228746414185,
+ "eval_runtime": 42.8572,
+ "eval_samples_per_second": 56.98,
+ "eval_steps_per_second": 0.467,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.0608622208237648,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.2348179817199707,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.05581629276275635,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.22896146774292,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.09162136912345886,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.2343664169311523,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.13750384747982025,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.2507717609405518,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.0872517079114914,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.2440056800842285,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.07161762565374374,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.229494571685791,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04082738608121872,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.2183239459991455,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.04972497373819351,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.218075752258301,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.049254074692726135,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.213379383087158,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.056671254336833954,
+ "learning_rate": 0.000599663984012462,
+ "loss": 3.209568977355957,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.05541909113526344,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.207760810852051,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.03819117695093155,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 3.199732780456543,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.0529443584382534,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.1945419311523438,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.043010979890823364,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.185936212539673,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.07311078161001205,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.189894914627075,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06110258400440216,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 3.1911773681640625,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.06880149245262146,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 3.187437057495117,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.051326874643564224,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 3.1826558113098145,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.03363822400569916,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 3.175527572631836,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.10268989205360413,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 3.172367572784424,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.09723010659217834,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 3.1873984336853027,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.04129317030310631,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 3.165799617767334,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.05292229354381561,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 3.171780824661255,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.03873143345117569,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 3.1680874824523926,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.047018859535455704,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 3.161402940750122,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.03942955285310745,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 3.1530532836914062,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.039801858365535736,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 3.1465344429016113,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.03795155510306358,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 3.1364030838012695,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.0342850424349308,
+ "learning_rate": 0.000599379982511273,
+ "loss": 3.1391525268554688,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.03243280574679375,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 3.1331253051757812,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.029679352417588234,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 3.1346120834350586,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.030230293050408363,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 3.132615804672241,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.02845083735883236,
+ "learning_rate": 0.000599309205504528,
+ "loss": 3.1242547035217285,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.03208369016647339,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 3.119110584259033,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.040268607437610626,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 3.113833427429199,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.06951889395713806,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 3.1186914443969727,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.08879823237657547,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 3.1140987873077393,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07374681532382965,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 3.1147828102111816,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.038190409541130066,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 3.1161253452301025,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.03297388181090355,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 3.1006083488464355,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.05905177444219589,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 3.106564521789551,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.037567321211099625,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 3.0884456634521484,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.02725658379495144,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 3.09000825881958,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.04272424429655075,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 3.090528964996338,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.027587950229644775,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 3.085131883621216,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.03282274678349495,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 3.0880112648010254,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.04064026474952698,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 3.079259157180786,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.023374097421765327,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 3.0747299194335938,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.025003870949149132,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 3.0721259117126465,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.030445197597146034,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 3.0733323097229004,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.025935636833310127,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 3.070709228515625,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.021337421610951424,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 3.0627551078796387,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.040481697767972946,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 3.0529394149780273,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.047608256340026855,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 3.041020631790161,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.03356955572962761,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 3.0532307624816895,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.02101682312786579,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 3.0492985248565674,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.030098581686615944,
+ "learning_rate": 0.000598804365316586,
+ "loss": 3.0527801513671875,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.040166422724723816,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 3.0505762100219727,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.051579318940639496,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 3.0422110557556152,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.06485334783792496,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 3.0435101985931396,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.055590901523828506,
+ "learning_rate": 0.000598706872384039,
+ "loss": 3.0448739528656006,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.03183475881814957,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 3.0416858196258545,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.03995629772543907,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 3.0255980491638184,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.03907827287912369,
+ "learning_rate": 0.000598631251036868,
+ "loss": 3.032339096069336,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.027681387960910797,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 3.025439739227295,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.03470184653997421,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 3.02883243560791,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.031407181173563004,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 3.0289883613586426,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.0371522456407547,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 3.013608932495117,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.03651345893740654,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 3.0120253562927246,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.06004595756530762,
+ "learning_rate": 0.000598473577921366,
+ "loss": 3.017958402633667,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.09979518502950668,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 3.023041248321533,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.10138462483882904,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.275729179382324,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7571707367897034,
+ "eval_runtime": 42.8283,
+ "eval_samples_per_second": 57.018,
+ "eval_steps_per_second": 0.467,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.04982660710811615,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 3.016209602355957,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.06532539427280426,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 3.015043020248413,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.04369623214006424,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 3.0096778869628906,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.05180002748966217,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 3.0121543407440186,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.0440233014523983,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.9937305450439453,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.05645671859383583,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 3.0023865699768066,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.03990716487169266,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 3.0003232955932617,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.03818608075380325,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.994137763977051,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.035388972610235214,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.993170738220215,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.03162284567952156,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.9922614097595215,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.029497336596250534,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.9852986335754395,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.02586926706135273,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.9798672199249268,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.027355050668120384,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.985158920288086,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.0341978520154953,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.982581615447998,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.030727429315447807,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.9701411724090576,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.04429870471358299,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.9789254665374756,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.05775132030248642,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.9758143424987793,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.05635962635278702,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.981480121612549,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.03639192506670952,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.968752861022949,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.028748342767357826,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.9584972858428955,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.035653006285429,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.967545747756958,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.03697885945439339,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.9578099250793457,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.03518654406070709,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.9680328369140625,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.034762222319841385,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.9670262336730957,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03813016042113304,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.957522392272949,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.04372904822230339,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.9540514945983887,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.06646446138620377,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.951028823852539,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.07831092923879623,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.95898699760437,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.05779588967561722,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.965884208679199,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.04414296895265579,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.9510560035705566,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.04797815904021263,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.955803394317627,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.04482145980000496,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.954059362411499,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.037104446440935135,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.94792103767395,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.03616701066493988,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.9500932693481445,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.03882797807455063,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.9558305740356445,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.04356204345822334,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.9237029552459717,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.04535257816314697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.925853729248047,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.04519587755203247,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.924984931945801,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.055922407656908035,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.932561159133911,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.06330762803554535,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.9388322830200195,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.0517115443944931,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.9272942543029785,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.057369694113731384,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.9201431274414062,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.05687880516052246,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.922886371612549,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.055525410920381546,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.9211201667785645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04503652825951576,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.9164583683013916,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.03385302051901817,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.9145395755767822,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03564384952187538,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.916003704071045,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.04188712686300278,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.915839433670044,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03657370060682297,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.903120279312134,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.03126626834273338,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.895824432373047,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.04196290299296379,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.9093422889709473,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.1706087589263916,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.9183316230773926,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.05810421332716942,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.911245107650757,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.061330314725637436,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.9159231185913086,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.06062281131744385,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.9211535453796387,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.05590800568461418,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.9140496253967285,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.04496239125728607,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.9040920734405518,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.03835335746407509,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.9026548862457275,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.05137286335229874,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.889634132385254,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.04091101884841919,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.886838674545288,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.04076292738318443,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.892918109893799,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.045437734574079514,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.893305778503418,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.041114673018455505,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.88517165184021,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.03600773215293884,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.8779795169830322,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.04311220720410347,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.87247896194458,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.04818756878376007,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.889103412628174,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.053460877388715744,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.8899683952331543,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.06763825565576553,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.8823537826538086,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.07304801046848297,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.877340316772461,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.061337269842624664,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.8820111751556396,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.04660911113023758,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.868886709213257,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.053929202258586884,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.159337043762207,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.721762478351593,
+ "eval_runtime": 42.7977,
+ "eval_samples_per_second": 57.059,
+ "eval_steps_per_second": 0.467,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.04259369149804115,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.859579086303711,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.044425517320632935,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.8681139945983887,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.04301869496703148,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.853911876678467,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.04574296995997429,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.8724427223205566,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.050247322767972946,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.8528172969818115,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.046683840453624725,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.853405714035034,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.05966169387102127,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.8605127334594727,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.08935574442148209,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.862339973449707,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.08019179105758667,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.8655967712402344,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.04753892868757248,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.8657939434051514,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06868117302656174,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.861976146697998,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.07040470093488693,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.8562135696411133,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.05683179199695587,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.8576629161834717,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.04918001592159271,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.861943244934082,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.04921915754675865,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.8495378494262695,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.037651967257261276,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.8428618907928467,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.03522057458758354,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.8372769355773926,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.03111186809837818,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.8377151489257812,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.03906773030757904,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.846196174621582,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.02546556107699871,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.8288016319274902,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.03119821846485138,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.8334617614746094,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.030583377927541733,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.8209362030029297,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.029693076387047768,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.8267486095428467,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.028978517279028893,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.819244861602783,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.027871523052453995,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.8289647102355957,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.029287064447999,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.8235747814178467,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.024990517646074295,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.819016218185425,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.030980689451098442,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.828155040740967,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03721801936626434,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.813418388366699,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03738148882985115,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.808328866958618,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.03580411896109581,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.812070846557617,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.039036594331264496,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.807159423828125,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.044666968286037445,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.799992561340332,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.055589817464351654,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.804142951965332,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.06139020994305611,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.8182730674743652,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.07769318670034409,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.8166518211364746,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.06521332263946533,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.8142249584198,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.09083619713783264,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.8166933059692383,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.09737467020750046,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.8111555576324463,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.07148188352584839,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.8168413639068604,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.055516090244054794,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.8183553218841553,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.05338391661643982,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.8127028942108154,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.040785472840070724,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.8124775886535645,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.037135496735572815,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.8065357208251953,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.032455578446388245,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.798459053039551,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.03465662896633148,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.798649549484253,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.032026033848524094,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.7906997203826904,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.026019956916570663,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.791969060897827,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.030298635363578796,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.784593105316162,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.028147678822278976,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.782454013824463,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.023933131247758865,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.787851333618164,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.026789527386426926,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.788067579269409,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.024554884061217308,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.777033567428589,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.028679484501481056,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.786342144012451,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.026410218328237534,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.7790610790252686,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.027413051575422287,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.778944969177246,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.02908242493867874,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.763850212097168,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.029632994905114174,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.7672982215881348,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.027458716183900833,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.7702178955078125,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.03263958916068077,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.7581803798675537,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04427146911621094,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.768141508102417,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.07142394036054611,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.75819730758667,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.08833584934473038,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.782716751098633,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.08168833702802658,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.7763240337371826,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07253874838352203,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.7717690467834473,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.08186902105808258,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.767303466796875,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.05854064226150513,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.761544704437256,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.04602917283773422,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.7669177055358887,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.03920970857143402,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.766085624694824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.04551195353269577,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.7623450756073,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.040861159563064575,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.7471840381622314,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.026907291263341904,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 2.0710244178771973,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6934371590614319,
+ "eval_runtime": 42.6015,
+ "eval_samples_per_second": 57.322,
+ "eval_steps_per_second": 0.469,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.03801899403333664,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.739086627960205,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.03707238286733627,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.741638422012329,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.0387551411986351,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.740777015686035,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.028010359033942223,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.7353010177612305,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.026474518701434135,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.73842453956604,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.02920856513082981,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.736295461654663,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.026074763387441635,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.7295751571655273,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.026862047612667084,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.723040819168091,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.026300370693206787,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.7199485301971436,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.026030950248241425,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.7276217937469482,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.028342435136437416,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.7193045616149902,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.032135091722011566,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.7143118381500244,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.03068603202700615,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.7191050052642822,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.035828713327646255,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.718085765838623,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.03343002125620842,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.707131862640381,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.035051796585321426,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.7091879844665527,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.03395913913846016,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6980862617492676,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.034198079258203506,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.699704647064209,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.0322839580476284,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.702791213989258,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.041531339287757874,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.7018051147460938,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.040806882083415985,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.7096598148345947,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.04156940430402756,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6956775188446045,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.05608797073364258,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.70127272605896,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0729481428861618,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.708817720413208,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.0905836969614029,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.718085289001465,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.08345075696706772,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.7235910892486572,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.08599896728992462,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.7178783416748047,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.07301691174507141,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.714029312133789,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.08231015503406525,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.7196826934814453,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.08286412060260773,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.7176082134246826,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.08590883016586304,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.7242164611816406,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.07531397044658661,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.7225024700164795,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.04711190238595009,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.695887327194214,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.05076327919960022,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.7029781341552734,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.04495669901371002,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.7057900428771973,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.04312257468700409,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.697537899017334,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.043580975383520126,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.67315673828125,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.03360886126756668,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.6883413791656494,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.030360296368598938,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.700927257537842,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.0310810673981905,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.6917803287506104,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.02703232690691948,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.6667580604553223,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.02756045199930668,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.6702821254730225,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.02611083723604679,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.669790744781494,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.02615324780344963,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.6672821044921875,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.026000119745731354,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.6671881675720215,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.02882794849574566,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.67586088180542,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03327213600277901,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.6618690490722656,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.032920945435762405,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.665444850921631,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.02812962979078293,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.6680707931518555,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.025886602699756622,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.652898073196411,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.025222789496183395,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.6560399532318115,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.023131893947720528,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.66127347946167,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.02513662725687027,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.6463732719421387,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.027243830263614655,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.6507630348205566,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.025437965989112854,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.6485800743103027,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.02770340070128441,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.657241106033325,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.03594067320227623,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.6281137466430664,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04486451297998428,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.6340909004211426,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.05102410912513733,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.6463394165039062,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04514491558074951,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.636554002761841,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04469453915953636,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.6453680992126465,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.05421523004770279,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.646010398864746,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04779462888836861,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.6315622329711914,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.06697724014520645,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.6416730880737305,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.07520416378974915,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.641632318496704,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.055692676454782486,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.6289877891540527,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.04539895057678223,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.6363251209259033,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.04829065501689911,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.641608476638794,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.041314542293548584,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.6269643306732178,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.04100198298692703,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.6243174076080322,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.04105962812900543,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.626585006713867,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.03172122687101364,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9550899267196655,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6614077091217041,
+ "eval_runtime": 42.9009,
+ "eval_samples_per_second": 56.922,
+ "eval_steps_per_second": 0.466,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.03258531913161278,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.612126588821411,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.040250636637210846,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.6025328636169434,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.034234337508678436,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.6085972785949707,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.03635786101222038,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.6126222610473633,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.039404451847076416,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.6035327911376953,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.038504838943481445,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5985798835754395,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.03744236007332802,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5873115062713623,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.034184787422418594,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.5899314880371094,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.03766429424285889,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.602734327316284,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.03741588443517685,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.592207193374634,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.03187853842973709,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5850372314453125,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.038447409868240356,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.592654228210449,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.04046263173222542,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.5881710052490234,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.03872602432966232,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5881471633911133,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.03960775211453438,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5826945304870605,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.033132150769233704,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.584038734436035,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.029504230245947838,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.584357261657715,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.03665253147482872,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5677168369293213,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.037208545953035355,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.560795783996582,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.038244519382715225,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.591458559036255,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.043307431042194366,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.5642588138580322,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.055119726806879044,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.57377290725708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.053323689848184586,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5767312049865723,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.0502864345908165,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.562190055847168,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.05568842589855194,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.565556049346924,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.061149828135967255,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5883288383483887,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.059933699667453766,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.5710184574127197,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.06041933596134186,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.572603225708008,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.05482695251703262,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.5667271614074707,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.04712968319654465,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.554041624069214,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.04842432215809822,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.5660672187805176,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.04273047298192978,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.554839849472046,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.05356749892234802,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.5521483421325684,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.06062676012516022,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.5551886558532715,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.05431462824344635,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5654542446136475,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.0505850687623024,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5648036003112793,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.04696856811642647,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.5559377670288086,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.042585279792547226,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.5660319328308105,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.04026484116911888,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.551206588745117,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03382421284914017,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.556361198425293,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.031177924945950508,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.5413894653320312,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.028217196464538574,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.5600497722625732,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.027062954381108284,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5447568893432617,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.03001212701201439,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.540614604949951,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.028869854286313057,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.544583797454834,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.02869918756186962,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.5471596717834473,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03359716013073921,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5416982173919678,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.04000182822346687,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.534921407699585,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.043030042201280594,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5368564128875732,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.04329848289489746,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.539764881134033,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.03740967437624931,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.5246665477752686,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.03733935207128525,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.529362916946411,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.039609868079423904,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.542544364929199,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.03937261551618576,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5366649627685547,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.03261927142739296,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.5250535011291504,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.03904294595122337,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.520721435546875,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.03913239762187004,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.5122013092041016,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.03610851988196373,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.5212767124176025,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.03944548964500427,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.537186622619629,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04762452840805054,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.5204620361328125,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04427091404795647,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5286149978637695,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04366641491651535,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.5398361682891846,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.04644155502319336,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.519331932067871,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.04372382536530495,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.531022787094116,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.04404324293136597,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.521758794784546,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.042115990072488785,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.5146074295043945,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.03796013817191124,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.5040054321289062,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.04327977076172829,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.5209312438964844,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04150385782122612,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.515871286392212,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.04465346783399582,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.5247764587402344,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.05690346285700798,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.501911163330078,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04907930642366409,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8968371152877808,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6358620524406433,
+ "eval_runtime": 42.9991,
+ "eval_samples_per_second": 56.792,
+ "eval_steps_per_second": 0.465,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.03755255416035652,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.486576795578003,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.040234558284282684,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.501087188720703,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.041416723281145096,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4942233562469482,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.039655014872550964,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.503453254699707,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.041707683354616165,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.490978717803955,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.04545537009835243,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4757118225097656,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.04328389838337898,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.504516839981079,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.043312299996614456,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.4920215606689453,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.046466976404190063,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.494760513305664,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.041762981563806534,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4875552654266357,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.03713781014084816,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.495652675628662,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.040603458881378174,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4821407794952393,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.037969645112752914,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.484981060028076,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.03652074560523033,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4736409187316895,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.03783264383673668,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.483346700668335,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.03404324874281883,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.48856782913208,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.03294725716114044,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.483706474304199,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.0347001850605011,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.474888324737549,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.0359107106924057,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.478990077972412,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.03738103806972504,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.4709420204162598,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.03722188249230385,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.465726852416992,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.034907322376966476,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.482231855392456,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.035695310682058334,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4785828590393066,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.04009179025888443,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.4795875549316406,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.038996778428554535,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.478696823120117,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04082407429814339,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.465200901031494,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04856136813759804,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4605681896209717,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.05671004578471184,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.47277569770813,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.0481606163084507,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.476151466369629,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.037689510732889175,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.466142177581787,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.05005146935582161,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.4798035621643066,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.04745829105377197,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.46163272857666,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.04638158157467842,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4728071689605713,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.04530330374836922,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.4621877670288086,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.03971488028764725,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.474224090576172,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.04521876201033592,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.46122670173645,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03923020884394646,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.4603679180145264,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03986494988203049,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.4589664936065674,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.04280020296573639,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.4624319076538086,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.03930043429136276,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.47080135345459,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.0314876064658165,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.458728790283203,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.03189660236239433,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4467148780822754,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03264275938272476,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4613699913024902,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03021484799683094,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.4535279273986816,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.030740290880203247,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.4621875286102295,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03096810355782509,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.458500385284424,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03105205111205578,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.461958646774292,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.03330068290233612,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.449777364730835,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.032831475138664246,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.4616191387176514,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.0387776717543602,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.445256471633911,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.041803423315286636,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.4545960426330566,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.04148292914032936,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.433408737182617,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.04648579657077789,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.449134349822998,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.04186409339308739,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.4449565410614014,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.038569554686546326,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.443675994873047,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03908300772309303,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4626526832580566,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.038897737860679626,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.438591957092285,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.0360480472445488,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.436077117919922,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0429830439388752,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4384970664978027,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.04465422406792641,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.458029270172119,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.037096526473760605,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.436816930770874,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.047823466360569,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.44439959526062,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.05929310992360115,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.4599215984344482,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.053512319922447205,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4387903213500977,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.042625587433576584,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4359588623046875,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.056042976677417755,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.4441447257995605,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05813189223408699,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4399144649505615,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.06313478201627731,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.453447103500366,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.05281798169016838,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.4428963661193848,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.03991910442709923,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4359219074249268,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.047981832176446915,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4506607055664062,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.03646523132920265,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8225739002227783,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6196661591529846,
+ "eval_runtime": 43.8106,
+ "eval_samples_per_second": 55.74,
+ "eval_steps_per_second": 0.457,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.03673036769032478,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.4227209091186523,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.044547371566295624,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.4073169231414795,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.04089285805821419,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4279446601867676,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.03538106009364128,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.4234392642974854,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.03152978792786598,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4214606285095215,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.03324202075600624,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.4085447788238525,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.03352442383766174,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.417776107788086,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.03582802042365074,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.411144256591797,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.0351114459335804,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.405872344970703,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.038898173719644547,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4029338359832764,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.03926656022667885,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4100089073181152,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.04285634309053421,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.410360336303711,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.04179581627249718,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4093637466430664,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.03356701880693436,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4118056297302246,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.038140635937452316,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.4154677391052246,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.04468780383467674,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416846513748169,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.062421441078186035,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4157638549804688,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06317313015460968,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.418593406677246,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.049498945474624634,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4204845428466797,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.046322405338287354,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.403489589691162,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.053081002086400986,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.4148740768432617,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05222945287823677,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.416757583618164,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04893574118614197,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4241795539855957,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04951956123113632,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.401386260986328,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04019852355122566,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.4168622493743896,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.03625848889350891,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4026846885681152,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.03464345261454582,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4021215438842773,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.03750818222761154,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4113571643829346,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.0360100194811821,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.3977622985839844,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.03464623540639877,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.413273334503174,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.03193575143814087,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.423659324645996,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.02710454724729061,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.396085262298584,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.02830432914197445,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.420255422592163,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.029717663303017616,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.3958120346069336,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.030005933716893196,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3908543586730957,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.029054716229438782,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.3873138427734375,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.031556472182273865,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3892760276794434,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03518872708082199,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4113478660583496,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.03471249341964722,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.384857654571533,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03223155066370964,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3840765953063965,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.030012644827365875,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.388247489929199,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.03162285313010216,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.3987512588500977,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.03700442612171173,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.3993430137634277,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.03710822016000748,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.387359380722046,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.028411181643605232,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3925249576568604,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031246855854988098,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.38845157623291,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.034664880484342575,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.392436981201172,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.03354701027274132,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.390890121459961,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.03527764603495598,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.3954997062683105,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03325049206614494,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3847177028656006,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.028021875768899918,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.385910987854004,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.02964954823255539,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3787484169006348,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.03068520873785019,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.400784969329834,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03416873514652252,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.391033411026001,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.03534875437617302,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.3796474933624268,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.03732474148273468,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3831934928894043,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.03648678958415985,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3785600662231445,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03801294043660164,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.387587070465088,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.04186025261878967,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.3893089294433594,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.04416101053357124,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3852195739746094,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.04137178510427475,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.391308307647705,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.05010323226451874,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.3906192779541016,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.049587685614824295,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.3846192359924316,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.04125971347093582,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.384507179260254,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.039676666259765625,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.38167142868042,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.038775283843278885,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.3789167404174805,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.03634221479296684,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.383060932159424,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03464275971055031,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.388552665710449,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.031778786331415176,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.3903298377990723,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.027904309332370758,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.377493381500244,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.028964119032025337,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.381959915161133,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.0233962032943964,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7810559272766113,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6070130467414856,
+ "eval_runtime": 50.871,
+ "eval_samples_per_second": 48.004,
+ "eval_steps_per_second": 0.393,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.030209315940737724,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.3569841384887695,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.04112662002444267,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.376680374145508,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.055858466774225235,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.3621644973754883,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.06746986508369446,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3607563972473145,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.06736797094345093,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3548707962036133,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.0562627948820591,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.364654541015625,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.051833927631378174,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.3748884201049805,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.04824110120534897,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.3502087593078613,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.05518793687224388,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3687515258789062,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.06564056873321533,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.367131233215332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.0605909489095211,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.3589868545532227,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.05744732916355133,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.370117664337158,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.06430432945489883,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.3703484535217285,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.06186915934085846,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3720293045043945,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.05977998673915863,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.3512699604034424,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.06413097679615021,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.377257823944092,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.053935080766677856,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.350717544555664,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.049110863357782364,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.3688902854919434,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.04313219338655472,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3374459743499756,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.039556123316287994,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3482556343078613,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.03796547278761864,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3547937870025635,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.03646104037761688,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.344686269760132,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.03355357423424721,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.3588008880615234,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.03004707582294941,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.352043628692627,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.030186176300048828,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3500709533691406,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.029748762026429176,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.3523058891296387,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.03238696977496147,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3635129928588867,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.032599564641714096,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.346269130706787,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.028626957908272743,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.357151746749878,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.02909751422703266,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3557677268981934,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.029951099306344986,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.34753680229187,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.029768310487270355,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3408985137939453,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.03005973994731903,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.362542152404785,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03360462561249733,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.352529287338257,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.034002047032117844,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.3334460258483887,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.03292316198348999,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3446991443634033,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.0337117575109005,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.350605010986328,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03428212180733681,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.357408046722412,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.03169601410627365,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.349088191986084,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.03389411047101021,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.361576795578003,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.036497049033641815,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.35420560836792,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03574785590171814,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3678526878356934,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.030757993459701538,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3520517349243164,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.0289121363312006,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.337864637374878,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.028012607246637344,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.353999614715576,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.027558492496609688,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.342132568359375,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.02871510572731495,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3425655364990234,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03070886805653572,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3401665687561035,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.03056657873094082,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.345277786254883,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.030907969921827316,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.346494674682617,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0318334735929966,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.3390543460845947,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.029700733721256256,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.3398728370666504,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.029687359929084778,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.3474364280700684,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.027234600856900215,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3457484245300293,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.02823043428361416,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.337700366973877,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03002646192908287,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.335954189300537,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.027246717363595963,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.330207347869873,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.029703574255108833,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.328280448913574,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.028796236962080002,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3448386192321777,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.027944380417466164,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.320234537124634,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.030626650899648666,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.334698438644409,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.029164662584662437,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3450326919555664,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03034045360982418,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.3420588970184326,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.02856968529522419,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3396072387695312,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.02892562374472618,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.340027332305908,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.02938242256641388,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.33550763130188,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.029261089861392975,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3246421813964844,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.030340535566210747,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.3465514183044434,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.033335570245981216,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.3467206954956055,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.03452218323945999,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.343702793121338,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.03326255828142166,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.337631940841675,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025433633476495743,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7402067184448242,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03175730258226395,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.62623405456543,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.04780422896146774,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.623239517211914,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.0504758283495903,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.6369194984436035,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.04481920227408409,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.57640266418457,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0440552793443203,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.630992889404297,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0503336526453495,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.627910137176514,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.053885314613580704,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.637897491455078,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.05481956899166107,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.610715866088867,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.050510745495557785,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.639884948730469,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.04579266160726547,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.590188980102539,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.04817410930991173,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.636295318603516,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.058862220495939255,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.614317417144775,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.06328638643026352,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.655928611755371,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.058752477169036865,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.652004241943359,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.06049015372991562,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.6513800621032715,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.058941151946783066,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.657688140869141,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.05542639642953873,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.637376308441162,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06024057790637016,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.63592529296875,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.06040216609835625,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.6275787353515625,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.059600673615932465,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.652995586395264,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.05379730463027954,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.597358226776123,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05355260521173477,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.6106367111206055,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.06649646162986755,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.654443740844727,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.058309994637966156,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.6869659423828125,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.05403570458292961,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.654750823974609,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04678157716989517,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.653955459594727,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04297948628664017,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.644915580749512,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.04235142841935158,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.662282943725586,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.0405130498111248,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.64569091796875,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.03351007401943207,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.653165817260742,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.03182631731033325,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.606291770935059,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.031552448868751526,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.642643928527832,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027941036969423294,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.623664379119873,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.027744563296437263,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.605886459350586,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02962736040353775,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.606762886047363,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.027251560240983963,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.620492935180664,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.027894044294953346,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.613903999328613,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.02778882533311844,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.634457588195801,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.028082096949219704,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.610980033874512,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.026210976764559746,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.612090110778809,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.026537755504250526,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.581171035766602,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.028709523379802704,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.628799915313721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.028233813121914864,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.586182117462158,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.02966046892106533,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.62320613861084,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.03198636695742607,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.6018524169921875,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.033404890447854996,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.610121726989746,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.029217196628451347,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.601312160491943,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.029335757717490196,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.620820999145508,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.031007781624794006,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.626503944396973,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.028453458100557327,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.589323997497559,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02526312880218029,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.598350524902344,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.027317512780427933,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.62388277053833,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02710605598986149,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.606523513793945,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.027491388842463493,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.586644172668457,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.027158567681908607,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.606039047241211,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.0298276599496603,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.593776702880859,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.026622479781508446,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.59510612487793,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.028238043189048767,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.5899505615234375,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.029481008648872375,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.6092071533203125,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.030138323083519936,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.600069046020508,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.02951825223863125,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.600125789642334,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.029242202639579773,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.590723037719727,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.02941613830626011,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.623342514038086,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.029764151200652122,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.631185531616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.031503792852163315,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.605094909667969,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.03184918314218521,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.592767715454102,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.03223618119955063,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.5930585861206055,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.0340263694524765,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.609347343444824,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.03245621174573898,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.598440170288086,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.03204867243766785,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.619702339172363,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.030866602435708046,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.609165191650391,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.01960863545536995,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8655829429626465,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.593852162361145,
+ "eval_runtime": 60.0971,
+ "eval_samples_per_second": 40.634,
+ "eval_steps_per_second": 0.649,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03552858531475067,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.539240837097168,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.0536109134554863,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.52609395980835,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.061411239206790924,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.536109447479248,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.06352879852056503,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.546850204467773,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.05683520436286926,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.563828468322754,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.05522637069225311,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.550929069519043,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.058990392833948135,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.554411888122559,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.057443976402282715,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.545921325683594,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.05383576825261116,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.5481414794921875,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04797440022230148,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.55556583404541,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04923664405941963,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.545513153076172,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.056191347539424896,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.536705017089844,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.06088989973068237,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.570624351501465,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.04802761599421501,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.526930809020996,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.04018320515751839,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.562432289123535,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.04280981421470642,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.559737205505371,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.04258878529071808,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.561749458312988,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.04411563277244568,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.562010765075684,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.041077401489019394,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.573984622955322,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.04125386103987694,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.550564289093018,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.041974026709795,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.552739143371582,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03689907118678093,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.540107727050781,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03586242347955704,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.554999351501465,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.036602675914764404,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.546337127685547,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.03707926720380783,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.543377876281738,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.03322125971317291,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.574325084686279,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03359243646264076,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.5613603591918945,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.0314965583384037,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.530994415283203,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.031806789338588715,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.56138801574707,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.034988485276699066,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.553149700164795,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.037627559155225754,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.555401802062988,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03723530471324921,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.52452278137207,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.036220453679561615,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.54799747467041,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03228041157126427,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.574814319610596,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.030872676521539688,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.539666652679443,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.033935241401195526,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.579252243041992,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.034338828176259995,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.538324356079102,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.03441951423883438,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.540358543395996,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03328141197562218,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.520340919494629,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03224223107099533,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.5646209716796875,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.02999890223145485,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.504915714263916,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.03014362044632435,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.516615867614746,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.029469173401594162,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.560475826263428,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.027887821197509766,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.586333751678467,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.029970254749059677,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.530396461486816,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.031648170202970505,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.554404258728027,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.0325581431388855,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.544102668762207,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.03555625304579735,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.543701171875,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.033053912222385406,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.508375644683838,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.03059076890349388,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.545750617980957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.029640182852745056,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.54715633392334,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030139951035380363,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.533708572387695,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029753828421235085,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.543603897094727,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.029481034725904465,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.534200668334961,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.028991524130105972,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.5683674812316895,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.028145765885710716,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.543053150177002,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.02999844029545784,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.5482354164123535,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03134117275476456,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.5360107421875,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.029277140274643898,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.5729827880859375,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028516866266727448,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.543548583984375,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028387080878019333,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.552616596221924,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.028977863490581512,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.5456695556640625,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.028010815382003784,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.534921646118164,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.030222605913877487,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.561526298522949,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.03140007704496384,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.500124931335449,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.029269184917211533,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.536653995513916,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.02935507707297802,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.521398544311523,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02801917865872383,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.527442455291748,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.027995768934488297,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.556139945983887,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.028441084548830986,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.555442810058594,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.029100807383656502,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.532083511352539,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020385215058922768,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.8488006591796875,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5902590751647949,
+ "eval_runtime": 60.1424,
+ "eval_samples_per_second": 40.604,
+ "eval_steps_per_second": 0.648,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03513244539499283,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.486500263214111,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.052679311484098434,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.489691734313965,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.06171489134430885,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.471555709838867,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.06164781004190445,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.486891746520996,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.06410388648509979,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.494929313659668,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06399523466825485,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.5334320068359375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.0636502280831337,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.5039381980896,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.06139931082725525,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.512765884399414,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.07460282742977142,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.477229118347168,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.075187548995018,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.484212398529053,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.07584068924188614,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.495598793029785,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.07192158699035645,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.507598400115967,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.05366869643330574,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.514532089233398,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.06397061794996262,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.512920379638672,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.06391509622335434,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.465702056884766,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.059490568935871124,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.5351881980896,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.057501111179590225,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.510270595550537,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.056947533041238785,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.509241104125977,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.050393324345350266,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.50428581237793,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.057608287781476974,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.474919319152832,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.054470717906951904,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.521528244018555,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.054405540227890015,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.510211944580078,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.051328495144844055,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.480720520019531,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.045930180698633194,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.503024101257324,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.04298274591565132,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.49046516418457,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.0406254380941391,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.544064521789551,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.03557496517896652,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.496966361999512,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.0364493802189827,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.5047502517700195,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.034521300345659256,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.485365867614746,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.03305799141526222,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.484884738922119,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.033053625375032425,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.478784561157227,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03426017984747887,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.482428073883057,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.03638783469796181,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.433219909667969,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.03597250208258629,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.522584915161133,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.03086097538471222,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.500803470611572,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.029642771929502487,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.444099426269531,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.03190716728568077,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.493034362792969,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.0312873050570488,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.487887382507324,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03201943263411522,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.479177474975586,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.033088769763708115,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.5020246505737305,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.02975946106016636,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.455426216125488,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03213953226804733,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.501615047454834,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.03262908384203911,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.485121250152588,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.031102091073989868,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.488893508911133,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.02945449948310852,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.490121841430664,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.033680472522974014,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.51833438873291,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.03277525305747986,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.480165481567383,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.03303041681647301,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.501155376434326,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.034927863627672195,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.513175010681152,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.03098093718290329,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.503798484802246,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.029513906687498093,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.49006986618042,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029630785807967186,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.522514343261719,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.030888330191373825,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.4895734786987305,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.032392196357250214,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.471953868865967,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.03215661272406578,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.48297643661499,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.03346327692270279,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.518160820007324,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.031311362981796265,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.484991550445557,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.03056488186120987,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.525470733642578,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.03043230250477791,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.510650634765625,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.02916652522981167,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.496840953826904,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.030673664063215256,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.494868278503418,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030033906921744347,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.479182720184326,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.028828797861933708,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.471983909606934,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028316056355834007,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.5049147605896,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.03158092871308327,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.449270248413086,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.03041706420481205,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.508289813995361,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.02967553399503231,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.465579032897949,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030721619725227356,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.471003532409668,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03066832385957241,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.488987445831299,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.032297998666763306,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.491018295288086,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03221398591995239,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.471237659454346,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.021120507270097733,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.8080053329467773,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5888046026229858,
+ "eval_runtime": 59.1638,
+ "eval_samples_per_second": 41.275,
+ "eval_steps_per_second": 0.659,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.03799329698085785,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.401787757873535,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.05785121023654938,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.408820629119873,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.05964238941669464,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.427263259887695,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05466582626104355,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.403415679931641,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.05613672360777855,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.413750648498535,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.05396881327033043,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.41916561126709,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.05689029023051262,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.395590782165527,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.056090328842401505,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.421253204345703,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.0473211295902729,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.385158061981201,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.04464765638113022,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.431746959686279,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.046421997249126434,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.428816795349121,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.05039866268634796,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.452417850494385,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.05491664633154869,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.447483062744141,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05866196006536484,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.434992790222168,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.05532094091176987,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.450716018676758,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.051083534955978394,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.402632713317871,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.03815016150474548,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.433640956878662,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04203395918011665,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.415888786315918,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.043729111552238464,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.423260688781738,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.039426498115062714,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.428949356079102,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.03837858885526657,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.434070587158203,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.038315922021865845,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.471246719360352,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.03963236138224602,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.4434990882873535,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04027022793889046,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.436992645263672,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.0409243181347847,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.426936149597168,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03404631093144417,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.43462610244751,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.034393735229969025,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.437409400939941,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03635180741548538,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.451286315917969,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03349733352661133,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.419209003448486,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.03238677605986595,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.456454277038574,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.034473925828933716,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.3988728523254395,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.03455805033445358,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.434884071350098,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.037486057728528976,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.466741561889648,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.03617125004529953,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.429305076599121,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.033646389842033386,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.407780170440674,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03401118144392967,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.42977237701416,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.036534134298563004,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.451094627380371,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.0384049266576767,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.414777755737305,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03460627421736717,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.420644760131836,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.033746346831321716,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.461175918579102,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.037153810262680054,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.444699287414551,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03283539041876793,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.431022644042969,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03248969465494156,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.447571277618408,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03327852487564087,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.42236328125,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.031114818528294563,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.418244361877441,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.03231207653880119,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.455004692077637,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.030702251940965652,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.483610153198242,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031027289107441902,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.4470109939575195,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.03129073604941368,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.452799320220947,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.03634301945567131,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4661664962768555,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03718913346529007,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.412006855010986,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.034018341451883316,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.436755180358887,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032502319663763046,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.452866554260254,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.032902952283620834,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.449224472045898,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.0340626984834671,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.445425510406494,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.033442508429288864,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.432218551635742,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.03489428386092186,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.4205827713012695,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.031868621706962585,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.414196968078613,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.032986633479595184,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.46116304397583,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.03348355367779732,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.424936771392822,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03174544870853424,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.434438705444336,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.03293656185269356,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.440502166748047,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.035124607384204865,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.440943717956543,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03438971936702728,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.431408882141113,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.03537093102931976,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.43867301940918,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.03267325833439827,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.462226867675781,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.030995232984423637,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.433740615844727,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.03384929150342941,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.508179664611816,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.032025814056396484,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.407378196716309,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03288578242063522,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.455531120300293,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.032061848789453506,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.452212333679199,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022320276126265526,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7567386627197266,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5878658294677734,
+ "eval_runtime": 59.5054,
+ "eval_samples_per_second": 41.038,
+ "eval_steps_per_second": 0.655,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.039114587008953094,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.347412586212158,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.05164389684796333,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.391279220581055,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055882278829813004,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.355602264404297,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.05625041201710701,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3934855461120605,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05861108750104904,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.364226341247559,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.052431970834732056,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.363334655761719,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.06147193908691406,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.384820938110352,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.06701470166444778,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.373311996459961,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.06714873015880585,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.32745361328125,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.06684907525777817,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.373660087585449,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.05738082900643349,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.362325668334961,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05306725949048996,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.351685047149658,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.05831029266119003,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.374813079833984,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05547972396016121,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.361842632293701,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.048057056963443756,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.383205413818359,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.04682513326406479,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.365846157073975,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.05001377314329147,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.362151145935059,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.05140263959765434,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.376064300537109,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.04640384018421173,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.363801002502441,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04175020009279251,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.3477983474731445,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.04134776070713997,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.337469100952148,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.042178574949502945,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.351490020751953,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.03989115729928017,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.40725040435791,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04086652770638466,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.341119766235352,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.03931201249361038,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.3786468505859375,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.03855905309319496,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.3701605796813965,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.03803952783346176,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.401841163635254,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.036908265203237534,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.410647392272949,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.0398186556994915,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.380431175231934,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03899035602807999,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.374293804168701,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.03946344554424286,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.396918773651123,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.03738604858517647,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.404016017913818,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.03579043596982956,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.396914958953857,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.037139154970645905,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.407739639282227,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.03581099584698677,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.411419868469238,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.03548453748226166,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.332690238952637,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.037561237812042236,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.381314277648926,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.035827480256557465,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.371536731719971,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03538314625620842,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.395467758178711,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.036672890186309814,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.420590400695801,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.038981080055236816,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.395296096801758,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.038234200328588486,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.415799140930176,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.038830097764730453,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.413394927978516,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03990919142961502,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.39515495300293,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.036424681544303894,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.39730978012085,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03488154709339142,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.403696060180664,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.0363851860165596,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.368661403656006,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.035461440682411194,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.396518707275391,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03647598251700401,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.440667152404785,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.03667188808321953,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.3781352043151855,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034516457468271255,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.414007186889648,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.03536032512784004,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.42198371887207,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.037454646080732346,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.409174919128418,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03632275015115738,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3547682762146,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.038922134786844254,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.39993143081665,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.03562692925333977,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.409564971923828,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.033203717321157455,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.421782493591309,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03830202296376228,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.430092811584473,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.03835884481668472,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.379505634307861,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.037440430372953415,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.403937339782715,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03732272610068321,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.440683364868164,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03325957432389259,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.3949294090271,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.035377148538827896,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.400777816772461,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.03725374862551689,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.422494411468506,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03601110726594925,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.4008941650390625,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03193027526140213,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.400361061096191,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.034382354468107224,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.408535003662109,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.03747732564806938,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.412189483642578,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03641800582408905,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.381453037261963,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.0315779373049736,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.4169769287109375,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03583509847521782,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.385746955871582,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02574268914759159,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7456769943237305,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.588273823261261,
+ "eval_runtime": 59.9356,
+ "eval_samples_per_second": 40.744,
+ "eval_steps_per_second": 0.651,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013961605584642,
+ "grad_norm": 0.037516895681619644,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 4.33476448059082,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027923211169284,
+ "grad_norm": 0.048450272530317307,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 4.296454429626465,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041884816753926,
+ "grad_norm": 0.05368480458855629,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 4.314815998077393,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055846422338568,
+ "grad_norm": 0.05520009621977806,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 4.330798625946045,
+ "step": 1084
+ },
+ {
+ "epoch": 15.06980802792321,
+ "grad_norm": 0.05928800627589226,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 4.317663192749023,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083769633507853,
+ "grad_norm": 0.06315027177333832,
+ "learning_rate": 0.000491285979372924,
+ "loss": 4.329221725463867,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097731239092496,
+ "grad_norm": 0.06342213600873947,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 4.32546329498291,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111692844677139,
+ "grad_norm": 0.0667002946138382,
+ "learning_rate": 0.000490873338088198,
+ "loss": 4.309828758239746,
+ "step": 1088
+ },
+ {
+ "epoch": 15.12565445026178,
+ "grad_norm": 0.05264135077595711,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 4.307073593139648,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139616055846423,
+ "grad_norm": 0.051895879209041595,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 4.333526611328125,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153577661431065,
+ "grad_norm": 0.05508396774530411,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 4.330171585083008,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167539267015707,
+ "grad_norm": 0.057272691279649734,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 4.329268455505371,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181500872600349,
+ "grad_norm": 0.058956146240234375,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 4.341983795166016,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195462478184991,
+ "grad_norm": 0.0557059645652771,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 4.348274230957031,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209424083769633,
+ "grad_norm": 0.05661974847316742,
+ "learning_rate": 0.000489424334303338,
+ "loss": 4.347631454467773,
+ "step": 1095
+ },
+ {
+ "epoch": 15.223385689354275,
+ "grad_norm": 0.05338657647371292,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 4.334639072418213,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237347294938917,
+ "grad_norm": 0.05091148987412453,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 4.355465412139893,
+ "step": 1097
+ },
+ {
+ "epoch": 15.25130890052356,
+ "grad_norm": 0.04996965080499649,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 4.346457004547119,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265270506108202,
+ "grad_norm": 0.05573750659823418,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 4.346520900726318,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279232111692846,
+ "grad_norm": 0.05776917189359665,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 4.359007835388184,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293193717277488,
+ "grad_norm": 0.050820041447877884,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 4.331710338592529,
+ "step": 1101
+ },
+ {
+ "epoch": 15.30715532286213,
+ "grad_norm": 0.04441820830106735,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 4.3126420974731445,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321116928446772,
+ "grad_norm": 0.04582618921995163,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 4.312925338745117,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335078534031414,
+ "grad_norm": 0.043592359870672226,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 4.345674514770508,
+ "step": 1104
+ },
+ {
+ "epoch": 15.349040139616056,
+ "grad_norm": 0.04069037735462189,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 4.334713935852051,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363001745200698,
+ "grad_norm": 0.04083041846752167,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 4.358171463012695,
+ "step": 1106
+ },
+ {
+ "epoch": 15.37696335078534,
+ "grad_norm": 0.04119594022631645,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 4.317891597747803,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390924956369982,
+ "grad_norm": 0.042060643434524536,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 4.355810165405273,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404886561954624,
+ "grad_norm": 0.04721749201416969,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 4.353546619415283,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418848167539267,
+ "grad_norm": 0.04657730832695961,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 4.341683387756348,
+ "step": 1110
+ },
+ {
+ "epoch": 15.432809773123909,
+ "grad_norm": 0.04096715524792671,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 4.357192039489746,
+ "step": 1111
+ },
+ {
+ "epoch": 15.44677137870855,
+ "grad_norm": 0.03904155269265175,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 4.354775428771973,
+ "step": 1112
+ },
+ {
+ "epoch": 15.460732984293193,
+ "grad_norm": 0.03828521445393562,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 4.356499671936035,
+ "step": 1113
+ },
+ {
+ "epoch": 15.474694589877837,
+ "grad_norm": 0.03688685595989227,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 4.339485168457031,
+ "step": 1114
+ },
+ {
+ "epoch": 15.488656195462479,
+ "grad_norm": 0.03945618495345116,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 4.363500118255615,
+ "step": 1115
+ },
+ {
+ "epoch": 15.502617801047121,
+ "grad_norm": 0.03821462020277977,
+ "learning_rate": 0.000485033394934934,
+ "loss": 4.353529930114746,
+ "step": 1116
+ },
+ {
+ "epoch": 15.516579406631763,
+ "grad_norm": 0.036265499889850616,
+ "learning_rate": 0.000484822676912974,
+ "loss": 4.337663173675537,
+ "step": 1117
+ },
+ {
+ "epoch": 15.530541012216405,
+ "grad_norm": 0.03803930804133415,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 4.34660530090332,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544502617801047,
+ "grad_norm": 0.038425102829933167,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 4.373246192932129,
+ "step": 1119
+ },
+ {
+ "epoch": 15.55846422338569,
+ "grad_norm": 0.036852262914180756,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 4.331226825714111,
+ "step": 1120
+ },
+ {
+ "epoch": 15.572425828970331,
+ "grad_norm": 0.03700809180736542,
+ "learning_rate": 0.000483978341612154,
+ "loss": 4.375818729400635,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586387434554974,
+ "grad_norm": 0.03628664091229439,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 4.32705020904541,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600349040139616,
+ "grad_norm": 0.036831192672252655,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 4.337127208709717,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614310645724258,
+ "grad_norm": 0.03736541047692299,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 4.36502742767334,
+ "step": 1124
+ },
+ {
+ "epoch": 15.6282722513089,
+ "grad_norm": 0.03658579662442207,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 4.359844207763672,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642233856893542,
+ "grad_norm": 0.03688633441925049,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 4.3355793952941895,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656195462478184,
+ "grad_norm": 0.03477008268237114,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 4.333405494689941,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670157068062828,
+ "grad_norm": 0.03655015677213669,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 4.338570594787598,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68411867364747,
+ "grad_norm": 0.03664765506982803,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 4.352496147155762,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698080279232112,
+ "grad_norm": 0.037019114941358566,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 4.335735321044922,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712041884816754,
+ "grad_norm": 0.03949085995554924,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 4.342585563659668,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726003490401396,
+ "grad_norm": 0.04084350913763046,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 4.329622268676758,
+ "step": 1132
+ },
+ {
+ "epoch": 15.739965095986038,
+ "grad_norm": 0.038439832627773285,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 4.360318660736084,
+ "step": 1133
+ },
+ {
+ "epoch": 15.75392670157068,
+ "grad_norm": 0.03932987526059151,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 4.343446254730225,
+ "step": 1134
+ },
+ {
+ "epoch": 15.767888307155323,
+ "grad_norm": 0.041595373302698135,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 4.351495742797852,
+ "step": 1135
+ },
+ {
+ "epoch": 15.781849912739965,
+ "grad_norm": 0.03917490318417549,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 4.380537033081055,
+ "step": 1136
+ },
+ {
+ "epoch": 15.795811518324607,
+ "grad_norm": 0.038597021251916885,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 4.3527116775512695,
+ "step": 1137
+ },
+ {
+ "epoch": 15.809773123909249,
+ "grad_norm": 0.03651711344718933,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 4.369256496429443,
+ "step": 1138
+ },
+ {
+ "epoch": 15.823734729493891,
+ "grad_norm": 0.035118963569402695,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 4.350616455078125,
+ "step": 1139
+ },
+ {
+ "epoch": 15.837696335078533,
+ "grad_norm": 0.034945692867040634,
+ "learning_rate": 0.000479936075927473,
+ "loss": 4.365446090698242,
+ "step": 1140
+ },
+ {
+ "epoch": 15.851657940663177,
+ "grad_norm": 0.03519820421934128,
+ "learning_rate": 0.000479721889242305,
+ "loss": 4.357390403747559,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86561954624782,
+ "grad_norm": 0.03407098725438118,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 4.350573539733887,
+ "step": 1142
+ },
+ {
+ "epoch": 15.879581151832461,
+ "grad_norm": 0.036511123180389404,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 4.328500270843506,
+ "step": 1143
+ },
+ {
+ "epoch": 15.893542757417103,
+ "grad_norm": 0.039194293320178986,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 4.346471786499023,
+ "step": 1144
+ },
+ {
+ "epoch": 15.907504363001745,
+ "grad_norm": 0.03777911514043808,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 4.358243942260742,
+ "step": 1145
+ },
+ {
+ "epoch": 15.921465968586388,
+ "grad_norm": 0.0354655385017395,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 4.368768692016602,
+ "step": 1146
+ },
+ {
+ "epoch": 15.93542757417103,
+ "grad_norm": 0.03710418567061424,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 4.3482184410095215,
+ "step": 1147
+ },
+ {
+ "epoch": 15.949389179755672,
+ "grad_norm": 0.03545419126749039,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 4.360692501068115,
+ "step": 1148
+ },
+ {
+ "epoch": 15.963350785340314,
+ "grad_norm": 0.035963475704193115,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 4.3592119216918945,
+ "step": 1149
+ },
+ {
+ "epoch": 15.977312390924956,
+ "grad_norm": 0.03425297513604164,
+ "learning_rate": 0.000477787822648007,
+ "loss": 4.329504013061523,
+ "step": 1150
+ },
+ {
+ "epoch": 15.991273996509598,
+ "grad_norm": 0.03453255817294121,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 4.298142433166504,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.025766054168343544,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 2.7026405334472656,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.5888971090316772,
+ "eval_runtime": 58.9721,
+ "eval_samples_per_second": 41.409,
+ "eval_steps_per_second": 0.661,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013961605584644,
+ "grad_norm": 0.04088004305958748,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 4.268601417541504,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027923211169284,
+ "grad_norm": 0.05194641277194023,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 4.267483711242676,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041884816753928,
+ "grad_norm": 0.05488700792193413,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 4.2527570724487305,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05584642233857,
+ "grad_norm": 0.059292279183864594,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 4.283864974975586,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069808027923212,
+ "grad_norm": 0.0567227303981781,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 4.291148662567139,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083769633507853,
+ "grad_norm": 0.05582835525274277,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 4.250524520874023,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097731239092496,
+ "grad_norm": 0.055709999054670334,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 4.259716033935547,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111692844677137,
+ "grad_norm": 0.053324732929468155,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 4.247231483459473,
+ "step": 1160
+ },
+ {
+ "epoch": 16.12565445026178,
+ "grad_norm": 0.05367910861968994,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 4.263905048370361,
+ "step": 1161
+ },
+ {
+ "epoch": 16.13961605584642,
+ "grad_norm": 0.057094886898994446,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 4.266510009765625,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153577661431065,
+ "grad_norm": 0.05921962857246399,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 4.278281211853027,
+ "step": 1163
+ },
+ {
+ "epoch": 16.167539267015705,
+ "grad_norm": 0.060246821492910385,
+ "learning_rate": 0.000474756648870413,
+ "loss": 4.2848052978515625,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18150087260035,
+ "grad_norm": 0.05686050280928612,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 4.286434173583984,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195462478184993,
+ "grad_norm": 0.05207265168428421,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 4.246121406555176,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209424083769633,
+ "grad_norm": 0.0477963425219059,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 4.287411689758301,
+ "step": 1167
+ },
+ {
+ "epoch": 16.223385689354277,
+ "grad_norm": 0.05586806684732437,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 4.2721333503723145,
+ "step": 1168
+ },
+ {
+ "epoch": 16.237347294938917,
+ "grad_norm": 0.05427060276269913,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 4.2751641273498535,
+ "step": 1169
+ },
+ {
+ "epoch": 16.25130890052356,
+ "grad_norm": 0.0518609918653965,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 4.287444114685059,
+ "step": 1170
+ },
+ {
+ "epoch": 16.2652705061082,
+ "grad_norm": 0.053606316447257996,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 4.287125587463379,
+ "step": 1171
+ },
+ {
+ "epoch": 16.279232111692846,
+ "grad_norm": 0.05118813365697861,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 4.291893005371094,
+ "step": 1172
+ },
+ {
+ "epoch": 16.293193717277486,
+ "grad_norm": 0.04774491861462593,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 4.325250148773193,
+ "step": 1173
+ },
+ {
+ "epoch": 16.30715532286213,
+ "grad_norm": 0.04634870961308479,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 4.311764717102051,
+ "step": 1174
+ },
+ {
+ "epoch": 16.32111692844677,
+ "grad_norm": 0.04411927983164787,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 4.277068138122559,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335078534031414,
+ "grad_norm": 0.043819356709718704,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 4.309638977050781,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349040139616054,
+ "grad_norm": 0.0418904684484005,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 4.306234836578369,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363001745200698,
+ "grad_norm": 0.04327298328280449,
+ "learning_rate": 0.000471698352726896,
+ "loss": 4.296874046325684,
+ "step": 1178
+ },
+ {
+ "epoch": 16.376963350785342,
+ "grad_norm": 0.045296184718608856,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 4.288957595825195,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390924956369982,
+ "grad_norm": 0.04447923228144646,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 4.313834190368652,
+ "step": 1180
+ },
+ {
+ "epoch": 16.404886561954626,
+ "grad_norm": 0.04548390209674835,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 4.284562587738037,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418848167539267,
+ "grad_norm": 0.04567817226052284,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 4.295994758605957,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43280977312391,
+ "grad_norm": 0.04277439042925835,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 4.262106895446777,
+ "step": 1183
+ },
+ {
+ "epoch": 16.44677137870855,
+ "grad_norm": 0.043597545474767685,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 4.294536590576172,
+ "step": 1184
+ },
+ {
+ "epoch": 16.460732984293195,
+ "grad_norm": 0.04424033313989639,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 4.297497749328613,
+ "step": 1185
+ },
+ {
+ "epoch": 16.474694589877835,
+ "grad_norm": 0.04354134574532509,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 4.321508884429932,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48865619546248,
+ "grad_norm": 0.04241178184747696,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 4.322726249694824,
+ "step": 1187
+ },
+ {
+ "epoch": 16.50261780104712,
+ "grad_norm": 0.039317622780799866,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 4.311155319213867,
+ "step": 1188
+ },
+ {
+ "epoch": 16.516579406631763,
+ "grad_norm": 0.04126536473631859,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 4.306636810302734,
+ "step": 1189
+ },
+ {
+ "epoch": 16.530541012216403,
+ "grad_norm": 0.040116794407367706,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 4.314671516418457,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544502617801047,
+ "grad_norm": 0.04045999422669411,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 4.301012992858887,
+ "step": 1191
+ },
+ {
+ "epoch": 16.55846422338569,
+ "grad_norm": 0.04013422504067421,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 4.29913330078125,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57242582897033,
+ "grad_norm": 0.03878502920269966,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 4.351600646972656,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586387434554975,
+ "grad_norm": 0.038935888558626175,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 4.27667236328125,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600349040139616,
+ "grad_norm": 0.04055892676115036,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 4.29019021987915,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61431064572426,
+ "grad_norm": 0.03959144279360771,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 4.327725887298584,
+ "step": 1196
+ },
+ {
+ "epoch": 16.6282722513089,
+ "grad_norm": 0.039878539741039276,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 4.297952651977539,
+ "step": 1197
+ },
+ {
+ "epoch": 16.642233856893544,
+ "grad_norm": 0.041209928691387177,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 4.313030242919922,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656195462478184,
+ "grad_norm": 0.04346393793821335,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 4.274901390075684,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670157068062828,
+ "grad_norm": 0.04090269282460213,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 4.330657958984375,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68411867364747,
+ "grad_norm": 0.04004506766796112,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 4.323106288909912,
+ "step": 1201
+ },
+ {
+ "epoch": 16.698080279232112,
+ "grad_norm": 0.041541460901498795,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 4.280947685241699,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712041884816752,
+ "grad_norm": 0.04033653065562248,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 4.341434955596924,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726003490401396,
+ "grad_norm": 0.04104280099272728,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 4.2931671142578125,
+ "step": 1204
+ },
+ {
+ "epoch": 16.739965095986037,
+ "grad_norm": 0.039398156106472015,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 4.316840648651123,
+ "step": 1205
+ },
+ {
+ "epoch": 16.75392670157068,
+ "grad_norm": 0.03980560600757599,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 4.325345039367676,
+ "step": 1206
+ },
+ {
+ "epoch": 16.767888307155324,
+ "grad_norm": 0.040656186640262604,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 4.32675838470459,
+ "step": 1207
+ },
+ {
+ "epoch": 16.781849912739965,
+ "grad_norm": 0.03775986284017563,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 4.314423561096191,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79581151832461,
+ "grad_norm": 0.04043284431099892,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 4.344202041625977,
+ "step": 1209
+ },
+ {
+ "epoch": 16.80977312390925,
+ "grad_norm": 0.04074503108859062,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 4.350123405456543,
+ "step": 1210
+ },
+ {
+ "epoch": 16.823734729493893,
+ "grad_norm": 0.04117420315742493,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 4.350917816162109,
+ "step": 1211
+ },
+ {
+ "epoch": 16.837696335078533,
+ "grad_norm": 0.03799610957503319,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 4.289987564086914,
+ "step": 1212
+ },
+ {
+ "epoch": 16.851657940663177,
+ "grad_norm": 0.04071265831589699,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 4.306744575500488,
+ "step": 1213
+ },
+ {
+ "epoch": 16.865619546247817,
+ "grad_norm": 0.04109715297818184,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 4.318753242492676,
+ "step": 1214
+ },
+ {
+ "epoch": 16.87958115183246,
+ "grad_norm": 0.03910145163536072,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 4.281325340270996,
+ "step": 1215
+ },
+ {
+ "epoch": 16.8935427574171,
+ "grad_norm": 0.03752733767032623,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 4.3231048583984375,
+ "step": 1216
+ },
+ {
+ "epoch": 16.907504363001745,
+ "grad_norm": 0.03752103075385094,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 4.351336479187012,
+ "step": 1217
+ },
+ {
+ "epoch": 16.921465968586386,
+ "grad_norm": 0.03675422817468643,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 4.282256126403809,
+ "step": 1218
+ },
+ {
+ "epoch": 16.93542757417103,
+ "grad_norm": 0.03850258141756058,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 4.322597503662109,
+ "step": 1219
+ },
+ {
+ "epoch": 16.949389179755673,
+ "grad_norm": 0.03692392632365227,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 4.320333957672119,
+ "step": 1220
+ },
+ {
+ "epoch": 16.963350785340314,
+ "grad_norm": 0.035366240888834,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 4.31341552734375,
+ "step": 1221
+ },
+ {
+ "epoch": 16.977312390924958,
+ "grad_norm": 0.037120968103408813,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 4.321141242980957,
+ "step": 1222
+ },
+ {
+ "epoch": 16.991273996509598,
+ "grad_norm": 0.03721754252910614,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 4.305757522583008,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.026698043569922447,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 2.716372489929199,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.5894472599029541,
+ "eval_runtime": 58.927,
+ "eval_samples_per_second": 41.441,
+ "eval_steps_per_second": 0.662,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013961605584644,
+ "grad_norm": 0.044532276690006256,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 4.243244171142578,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027923211169284,
+ "grad_norm": 0.050529688596725464,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 4.221382141113281,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041884816753928,
+ "grad_norm": 0.05690870061516762,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 4.240285873413086,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05584642233857,
+ "grad_norm": 0.054803814738988876,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 4.213047027587891,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069808027923212,
+ "grad_norm": 0.051825862377882004,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 4.226880073547363,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083769633507853,
+ "grad_norm": 0.049802035093307495,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 4.25303840637207,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097731239092496,
+ "grad_norm": 0.05290328338742256,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 4.199422836303711,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111692844677137,
+ "grad_norm": 0.0531814880669117,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 4.201648712158203,
+ "step": 1232
+ },
+ {
+ "epoch": 17.12565445026178,
+ "grad_norm": 0.048972588032484055,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 4.207578182220459,
+ "step": 1233
+ },
+ {
+ "epoch": 17.13961605584642,
+ "grad_norm": 0.051345713436603546,
+ "learning_rate": 0.000459203665939781,
+ "loss": 4.206205368041992,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153577661431065,
+ "grad_norm": 0.059012677520513535,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 4.228360176086426,
+ "step": 1235
+ },
+ {
+ "epoch": 17.167539267015705,
+ "grad_norm": 0.05984492972493172,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 4.234362602233887,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18150087260035,
+ "grad_norm": 0.060886185616254807,
+ "learning_rate": 0.000458522886595554,
+ "loss": 4.253525257110596,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195462478184993,
+ "grad_norm": 0.06476911902427673,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 4.249889373779297,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209424083769633,
+ "grad_norm": 0.059773124754428864,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 4.225249290466309,
+ "step": 1239
+ },
+ {
+ "epoch": 17.223385689354277,
+ "grad_norm": 0.06063077598810196,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 4.25465202331543,
+ "step": 1240
+ },
+ {
+ "epoch": 17.237347294938917,
+ "grad_norm": 0.059922896325588226,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 4.252101898193359,
+ "step": 1241
+ },
+ {
+ "epoch": 17.25130890052356,
+ "grad_norm": 0.05849695950746536,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 4.268218040466309,
+ "step": 1242
+ },
+ {
+ "epoch": 17.2652705061082,
+ "grad_norm": 0.05769657716155052,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 4.2395782470703125,
+ "step": 1243
+ },
+ {
+ "epoch": 17.279232111692846,
+ "grad_norm": 0.058189719915390015,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 4.257606506347656,
+ "step": 1244
+ },
+ {
+ "epoch": 17.293193717277486,
+ "grad_norm": 0.059679049998521805,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 4.278588771820068,
+ "step": 1245
+ },
+ {
+ "epoch": 17.30715532286213,
+ "grad_norm": 0.0579286590218544,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 4.25511360168457,
+ "step": 1246
+ },
+ {
+ "epoch": 17.32111692844677,
+ "grad_norm": 0.05269965901970863,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 4.241326332092285,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335078534031414,
+ "grad_norm": 0.04828358441591263,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 4.243182182312012,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349040139616054,
+ "grad_norm": 0.050172045826911926,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 4.295144081115723,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363001745200698,
+ "grad_norm": 0.047297943383455276,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 4.250682830810547,
+ "step": 1250
+ },
+ {
+ "epoch": 17.376963350785342,
+ "grad_norm": 0.04574039950966835,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 4.269444465637207,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390924956369982,
+ "grad_norm": 0.042310427874326706,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 4.255361557006836,
+ "step": 1252
+ },
+ {
+ "epoch": 17.404886561954626,
+ "grad_norm": 0.04268861189484596,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 4.252292156219482,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418848167539267,
+ "grad_norm": 0.0424286350607872,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 4.24439811706543,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43280977312391,
+ "grad_norm": 0.042049869894981384,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 4.236266136169434,
+ "step": 1255
+ },
+ {
+ "epoch": 17.44677137870855,
+ "grad_norm": 0.04185241088271141,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 4.268826484680176,
+ "step": 1256
+ },
+ {
+ "epoch": 17.460732984293195,
+ "grad_norm": 0.04423609375953674,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 4.240240573883057,
+ "step": 1257
+ },
+ {
+ "epoch": 17.474694589877835,
+ "grad_norm": 0.041972566395998,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 4.269453048706055,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48865619546248,
+ "grad_norm": 0.04493343085050583,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 4.270544052124023,
+ "step": 1259
+ },
+ {
+ "epoch": 17.50261780104712,
+ "grad_norm": 0.04502921551465988,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 4.271880149841309,
+ "step": 1260
+ },
+ {
+ "epoch": 17.516579406631763,
+ "grad_norm": 0.044518254697322845,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 4.249786376953125,
+ "step": 1261
+ },
+ {
+ "epoch": 17.530541012216403,
+ "grad_norm": 0.044482357800006866,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 4.300779819488525,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544502617801047,
+ "grad_norm": 0.0424671471118927,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 4.248597621917725,
+ "step": 1263
+ },
+ {
+ "epoch": 17.55846422338569,
+ "grad_norm": 0.04136480763554573,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 4.276763916015625,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57242582897033,
+ "grad_norm": 0.04094843938946724,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 4.263628005981445,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586387434554975,
+ "grad_norm": 0.04415388032793999,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 4.257946014404297,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600349040139616,
+ "grad_norm": 0.04177384823560715,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 4.289669036865234,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61431064572426,
+ "grad_norm": 0.040863484144210815,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 4.320903778076172,
+ "step": 1268
+ },
+ {
+ "epoch": 17.6282722513089,
+ "grad_norm": 0.04025879129767418,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 4.268717288970947,
+ "step": 1269
+ },
+ {
+ "epoch": 17.642233856893544,
+ "grad_norm": 0.04161325842142105,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 4.275268077850342,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656195462478184,
+ "grad_norm": 0.04260355979204178,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 4.263361930847168,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670157068062828,
+ "grad_norm": 0.043366871774196625,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 4.254581928253174,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68411867364747,
+ "grad_norm": 0.04299113526940346,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 4.268922328948975,
+ "step": 1273
+ },
+ {
+ "epoch": 17.698080279232112,
+ "grad_norm": 0.041865330189466476,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 4.290097236633301,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712041884816752,
+ "grad_norm": 0.04515853524208069,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 4.261376857757568,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726003490401396,
+ "grad_norm": 0.046030085533857346,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 4.290409564971924,
+ "step": 1276
+ },
+ {
+ "epoch": 17.739965095986037,
+ "grad_norm": 0.045034267008304596,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 4.28201961517334,
+ "step": 1277
+ },
+ {
+ "epoch": 17.75392670157068,
+ "grad_norm": 0.04293077066540718,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 4.282614707946777,
+ "step": 1278
+ },
+ {
+ "epoch": 17.767888307155324,
+ "grad_norm": 0.043866172432899475,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 4.2777557373046875,
+ "step": 1279
+ },
+ {
+ "epoch": 17.781849912739965,
+ "grad_norm": 0.04693951830267906,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 4.272597312927246,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79581151832461,
+ "grad_norm": 0.04072774201631546,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 4.284282684326172,
+ "step": 1281
+ },
+ {
+ "epoch": 17.80977312390925,
+ "grad_norm": 0.04241872951388359,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 4.254406452178955,
+ "step": 1282
+ },
+ {
+ "epoch": 17.823734729493893,
+ "grad_norm": 0.04614942893385887,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 4.297187805175781,
+ "step": 1283
+ },
+ {
+ "epoch": 17.837696335078533,
+ "grad_norm": 0.04326155036687851,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 4.292263984680176,
+ "step": 1284
+ },
+ {
+ "epoch": 17.851657940663177,
+ "grad_norm": 0.04053817689418793,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 4.2925920486450195,
+ "step": 1285
+ },
+ {
+ "epoch": 17.865619546247817,
+ "grad_norm": 0.040867388248443604,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 4.276965141296387,
+ "step": 1286
+ },
+ {
+ "epoch": 17.87958115183246,
+ "grad_norm": 0.041000183671712875,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 4.257638454437256,
+ "step": 1287
+ },
+ {
+ "epoch": 17.8935427574171,
+ "grad_norm": 0.04014866054058075,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 4.2734479904174805,
+ "step": 1288
+ },
+ {
+ "epoch": 17.907504363001745,
+ "grad_norm": 0.041237372905015945,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 4.263461112976074,
+ "step": 1289
+ },
+ {
+ "epoch": 17.921465968586386,
+ "grad_norm": 0.039109110832214355,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 4.295286655426025,
+ "step": 1290
+ },
+ {
+ "epoch": 17.93542757417103,
+ "grad_norm": 0.03911379352211952,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 4.254448413848877,
+ "step": 1291
+ },
+ {
+ "epoch": 17.949389179755673,
+ "grad_norm": 0.03997721150517464,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 4.284714698791504,
+ "step": 1292
+ },
+ {
+ "epoch": 17.963350785340314,
+ "grad_norm": 0.04011402651667595,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 4.278304100036621,
+ "step": 1293
+ },
+ {
+ "epoch": 17.977312390924958,
+ "grad_norm": 0.04040595516562462,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 4.288183212280273,
+ "step": 1294
+ },
+ {
+ "epoch": 17.991273996509598,
+ "grad_norm": 0.038857635110616684,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 4.274664402008057,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.028110472485423088,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 2.6625723838806152,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.5912551283836365,
+ "eval_runtime": 60.273,
+ "eval_samples_per_second": 40.516,
+ "eval_steps_per_second": 0.647,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013961605584644,
+ "grad_norm": 0.041634928435087204,
+ "learning_rate": 0.000444677103403854,
+ "loss": 4.181324005126953,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027923211169284,
+ "grad_norm": 0.051596369594335556,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 4.204722881317139,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041884816753928,
+ "grad_norm": 0.058580417186021805,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 4.207996845245361,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05584642233857,
+ "grad_norm": 0.05986522138118744,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 4.184011459350586,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069808027923212,
+ "grad_norm": 0.051045969128608704,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 4.148118019104004,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083769633507853,
+ "grad_norm": 0.045477766543626785,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 4.207469940185547,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097731239092496,
+ "grad_norm": 0.0488731674849987,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 4.201406478881836,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111692844677137,
+ "grad_norm": 0.05013517662882805,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 4.192239761352539,
+ "step": 1304
+ },
+ {
+ "epoch": 18.12565445026178,
+ "grad_norm": 0.05091651529073715,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 4.188252925872803,
+ "step": 1305
+ },
+ {
+ "epoch": 18.13961605584642,
+ "grad_norm": 0.051825884729623795,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 4.157929420471191,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153577661431065,
+ "grad_norm": 0.049625713378190994,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 4.208776473999023,
+ "step": 1307
+ },
+ {
+ "epoch": 18.167539267015705,
+ "grad_norm": 0.04816461354494095,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 4.180496692657471,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18150087260035,
+ "grad_norm": 0.04720531404018402,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 4.178536415100098,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195462478184993,
+ "grad_norm": 0.0475568026304245,
+ "learning_rate": 0.000441621628895411,
+ "loss": 4.203497886657715,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209424083769633,
+ "grad_norm": 0.04613947868347168,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 4.226914405822754,
+ "step": 1311
+ },
+ {
+ "epoch": 18.223385689354277,
+ "grad_norm": 0.04585929960012436,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 4.175119400024414,
+ "step": 1312
+ },
+ {
+ "epoch": 18.237347294938917,
+ "grad_norm": 0.045777060091495514,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 4.181760311126709,
+ "step": 1313
+ },
+ {
+ "epoch": 18.25130890052356,
+ "grad_norm": 0.04583602771162987,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 4.179868698120117,
+ "step": 1314
+ },
+ {
+ "epoch": 18.2652705061082,
+ "grad_norm": 0.04706503078341484,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 4.203325271606445,
+ "step": 1315
+ },
+ {
+ "epoch": 18.279232111692846,
+ "grad_norm": 0.045122336596250534,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 4.187152862548828,
+ "step": 1316
+ },
+ {
+ "epoch": 18.293193717277486,
+ "grad_norm": 0.04495825991034508,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 4.2080535888671875,
+ "step": 1317
+ },
+ {
+ "epoch": 18.30715532286213,
+ "grad_norm": 0.04767543077468872,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 4.207225322723389,
+ "step": 1318
+ },
+ {
+ "epoch": 18.32111692844677,
+ "grad_norm": 0.053396135568618774,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 4.227478981018066,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335078534031414,
+ "grad_norm": 0.0512089841067791,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 4.1904754638671875,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349040139616054,
+ "grad_norm": 0.04967977851629257,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 4.2079925537109375,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363001745200698,
+ "grad_norm": 0.04800970479846001,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 4.206037998199463,
+ "step": 1322
+ },
+ {
+ "epoch": 18.376963350785342,
+ "grad_norm": 0.04993556812405586,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 4.210054397583008,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390924956369982,
+ "grad_norm": 0.0480492040514946,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 4.229897499084473,
+ "step": 1324
+ },
+ {
+ "epoch": 18.404886561954626,
+ "grad_norm": 0.048699282109737396,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 4.17500114440918,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418848167539267,
+ "grad_norm": 0.04681490734219551,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 4.196526527404785,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43280977312391,
+ "grad_norm": 0.04674449935555458,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 4.246790409088135,
+ "step": 1327
+ },
+ {
+ "epoch": 18.44677137870855,
+ "grad_norm": 0.04650607705116272,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 4.225878715515137,
+ "step": 1328
+ },
+ {
+ "epoch": 18.460732984293195,
+ "grad_norm": 0.047200269997119904,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 4.20685338973999,
+ "step": 1329
+ },
+ {
+ "epoch": 18.474694589877835,
+ "grad_norm": 0.04322655498981476,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 4.212920665740967,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48865619546248,
+ "grad_norm": 0.04489128664135933,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 4.266653060913086,
+ "step": 1331
+ },
+ {
+ "epoch": 18.50261780104712,
+ "grad_norm": 0.048838935792446136,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 4.2364501953125,
+ "step": 1332
+ },
+ {
+ "epoch": 18.516579406631763,
+ "grad_norm": 0.04893286153674126,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 4.202394485473633,
+ "step": 1333
+ },
+ {
+ "epoch": 18.530541012216403,
+ "grad_norm": 0.046900101006031036,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 4.209832191467285,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544502617801047,
+ "grad_norm": 0.044056762009859085,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 4.229903221130371,
+ "step": 1335
+ },
+ {
+ "epoch": 18.55846422338569,
+ "grad_norm": 0.04681328684091568,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 4.22783899307251,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57242582897033,
+ "grad_norm": 0.046087440103292465,
+ "learning_rate": 0.000435215814386134,
+ "loss": 4.2078986167907715,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586387434554975,
+ "grad_norm": 0.04363090172410011,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 4.218686103820801,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600349040139616,
+ "grad_norm": 0.046255532652139664,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 4.2475128173828125,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61431064572426,
+ "grad_norm": 0.045545510947704315,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 4.233272552490234,
+ "step": 1340
+ },
+ {
+ "epoch": 18.6282722513089,
+ "grad_norm": 0.045259326696395874,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 4.252043724060059,
+ "step": 1341
+ },
+ {
+ "epoch": 18.642233856893544,
+ "grad_norm": 0.042144861072301865,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 4.247762203216553,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656195462478184,
+ "grad_norm": 0.04298558086156845,
+ "learning_rate": 0.000433781621332752,
+ "loss": 4.222329616546631,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670157068062828,
+ "grad_norm": 0.044070884585380554,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 4.222488880157471,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68411867364747,
+ "grad_norm": 0.044758375734090805,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 4.238160133361816,
+ "step": 1345
+ },
+ {
+ "epoch": 18.698080279232112,
+ "grad_norm": 0.04324857518076897,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 4.220003128051758,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712041884816752,
+ "grad_norm": 0.04349793121218681,
+ "learning_rate": 0.000432823382695327,
+ "loss": 4.251608848571777,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726003490401396,
+ "grad_norm": 0.04483242705464363,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 4.243319511413574,
+ "step": 1348
+ },
+ {
+ "epoch": 18.739965095986037,
+ "grad_norm": 0.04391994699835777,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 4.237479209899902,
+ "step": 1349
+ },
+ {
+ "epoch": 18.75392670157068,
+ "grad_norm": 0.042584922164678574,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 4.2309980392456055,
+ "step": 1350
+ },
+ {
+ "epoch": 18.767888307155324,
+ "grad_norm": 0.04363991320133209,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 4.216980934143066,
+ "step": 1351
+ },
+ {
+ "epoch": 18.781849912739965,
+ "grad_norm": 0.04662707448005676,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 4.257450103759766,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79581151832461,
+ "grad_norm": 0.0462074875831604,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 4.222329139709473,
+ "step": 1353
+ },
+ {
+ "epoch": 18.80977312390925,
+ "grad_norm": 0.043543554842472076,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 4.231960773468018,
+ "step": 1354
+ },
+ {
+ "epoch": 18.823734729493893,
+ "grad_norm": 0.04444567486643791,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 4.225714683532715,
+ "step": 1355
+ },
+ {
+ "epoch": 18.837696335078533,
+ "grad_norm": 0.04507977142930031,
+ "learning_rate": 0.000430661245733797,
+ "loss": 4.234722137451172,
+ "step": 1356
+ },
+ {
+ "epoch": 18.851657940663177,
+ "grad_norm": 0.043786536902189255,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 4.274998664855957,
+ "step": 1357
+ },
+ {
+ "epoch": 18.865619546247817,
+ "grad_norm": 0.04188325256109238,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 4.260570526123047,
+ "step": 1358
+ },
+ {
+ "epoch": 18.87958115183246,
+ "grad_norm": 0.0439927875995636,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 4.250492572784424,
+ "step": 1359
+ },
+ {
+ "epoch": 18.8935427574171,
+ "grad_norm": 0.04520147293806076,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 4.250911712646484,
+ "step": 1360
+ },
+ {
+ "epoch": 18.907504363001745,
+ "grad_norm": 0.04442249983549118,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 4.255832672119141,
+ "step": 1361
+ },
+ {
+ "epoch": 18.921465968586386,
+ "grad_norm": 0.044756144285202026,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 4.226641654968262,
+ "step": 1362
+ },
+ {
+ "epoch": 18.93542757417103,
+ "grad_norm": 0.04647127911448479,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 4.2110915184021,
+ "step": 1363
+ },
+ {
+ "epoch": 18.949389179755673,
+ "grad_norm": 0.045986004173755646,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 4.243263244628906,
+ "step": 1364
+ },
+ {
+ "epoch": 18.963350785340314,
+ "grad_norm": 0.04533308371901512,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 4.253750801086426,
+ "step": 1365
+ },
+ {
+ "epoch": 18.977312390924958,
+ "grad_norm": 0.04307471215724945,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 4.233024597167969,
+ "step": 1366
+ },
+ {
+ "epoch": 18.991273996509598,
+ "grad_norm": 0.0416078194975853,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 4.264965534210205,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.030926328152418137,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 2.651331901550293,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.592528223991394,
+ "eval_runtime": 59.67,
+ "eval_samples_per_second": 40.925,
+ "eval_steps_per_second": 0.654,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013961605584644,
+ "grad_norm": 0.04524356126785278,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 4.153796672821045,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027923211169284,
+ "grad_norm": 0.05105067789554596,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 4.154603004455566,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041884816753928,
+ "grad_norm": 0.052823506295681,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 4.1599273681640625,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05584642233857,
+ "grad_norm": 0.05688423290848732,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 4.124622344970703,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069808027923212,
+ "grad_norm": 0.053178343921899796,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 4.138644218444824,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083769633507853,
+ "grad_norm": 0.05473732203245163,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 4.138207912445068,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097731239092496,
+ "grad_norm": 0.055207233875989914,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 4.155735015869141,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111692844677137,
+ "grad_norm": 0.05484927073121071,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 4.140205383300781,
+ "step": 1376
+ },
+ {
+ "epoch": 19.12565445026178,
+ "grad_norm": 0.05589596927165985,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 4.160840034484863,
+ "step": 1377
+ },
+ {
+ "epoch": 19.13961605584642,
+ "grad_norm": 0.05222902446985245,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 4.154781341552734,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153577661431065,
+ "grad_norm": 0.051824700087308884,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 4.134243488311768,
+ "step": 1379
+ },
+ {
+ "epoch": 19.167539267015705,
+ "grad_norm": 0.0517590269446373,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 4.125945568084717,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18150087260035,
+ "grad_norm": 0.05399726703763008,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 4.146993160247803,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195462478184993,
+ "grad_norm": 0.054430052638053894,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 4.1310906410217285,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209424083769633,
+ "grad_norm": 0.0551968477666378,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 4.170640468597412,
+ "step": 1383
+ },
+ {
+ "epoch": 19.223385689354277,
+ "grad_norm": 0.06050962209701538,
+ "learning_rate": 0.000423881957237287,
+ "loss": 4.166876316070557,
+ "step": 1384
+ },
+ {
+ "epoch": 19.237347294938917,
+ "grad_norm": 0.055587392300367355,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 4.150339603424072,
+ "step": 1385
+ },
+ {
+ "epoch": 19.25130890052356,
+ "grad_norm": 0.05193328112363815,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 4.171754360198975,
+ "step": 1386
+ },
+ {
+ "epoch": 19.2652705061082,
+ "grad_norm": 0.05272402986884117,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 4.165149688720703,
+ "step": 1387
+ },
+ {
+ "epoch": 19.279232111692846,
+ "grad_norm": 0.056283943355083466,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 4.193273544311523,
+ "step": 1388
+ },
+ {
+ "epoch": 19.293193717277486,
+ "grad_norm": 0.05825977027416229,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 4.192652225494385,
+ "step": 1389
+ },
+ {
+ "epoch": 19.30715532286213,
+ "grad_norm": 0.05539379641413689,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 4.187044143676758,
+ "step": 1390
+ },
+ {
+ "epoch": 19.32111692844677,
+ "grad_norm": 0.05300498753786087,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 4.168220043182373,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335078534031414,
+ "grad_norm": 0.05425756797194481,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 4.185193061828613,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349040139616054,
+ "grad_norm": 0.05527182295918465,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 4.1618332862854,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363001745200698,
+ "grad_norm": 0.054422613233327866,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 4.195295810699463,
+ "step": 1394
+ },
+ {
+ "epoch": 19.376963350785342,
+ "grad_norm": 0.052717432379722595,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 4.143290996551514,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390924956369982,
+ "grad_norm": 0.05154350399971008,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 4.115989685058594,
+ "step": 1396
+ },
+ {
+ "epoch": 19.404886561954626,
+ "grad_norm": 0.0489451140165329,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 4.169506072998047,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418848167539267,
+ "grad_norm": 0.049093786627054214,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 4.164122104644775,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43280977312391,
+ "grad_norm": 0.04690805822610855,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 4.186679840087891,
+ "step": 1399
+ },
+ {
+ "epoch": 19.44677137870855,
+ "grad_norm": 0.046708665788173676,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 4.184325695037842,
+ "step": 1400
+ },
+ {
+ "epoch": 19.460732984293195,
+ "grad_norm": 0.04617927595973015,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 4.201678276062012,
+ "step": 1401
+ },
+ {
+ "epoch": 19.474694589877835,
+ "grad_norm": 0.05017530545592308,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 4.18928337097168,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48865619546248,
+ "grad_norm": 0.05197977274656296,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 4.188801288604736,
+ "step": 1403
+ },
+ {
+ "epoch": 19.50261780104712,
+ "grad_norm": 0.049691297113895416,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 4.209467887878418,
+ "step": 1404
+ },
+ {
+ "epoch": 19.516579406631763,
+ "grad_norm": 0.049709953367710114,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 4.19733190536499,
+ "step": 1405
+ },
+ {
+ "epoch": 19.530541012216403,
+ "grad_norm": 0.05274137482047081,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 4.221798896789551,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544502617801047,
+ "grad_norm": 0.05015392601490021,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 4.210565567016602,
+ "step": 1407
+ },
+ {
+ "epoch": 19.55846422338569,
+ "grad_norm": 0.05023058131337166,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 4.184835910797119,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57242582897033,
+ "grad_norm": 0.0479239858686924,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 4.199846267700195,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586387434554975,
+ "grad_norm": 0.0481228232383728,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 4.185157775878906,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600349040139616,
+ "grad_norm": 0.04909738525748253,
+ "learning_rate": 0.000417272427439646,
+ "loss": 4.207550525665283,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61431064572426,
+ "grad_norm": 0.0475567989051342,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 4.213225841522217,
+ "step": 1412
+ },
+ {
+ "epoch": 19.6282722513089,
+ "grad_norm": 0.048073310405015945,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 4.212669372558594,
+ "step": 1413
+ },
+ {
+ "epoch": 19.642233856893544,
+ "grad_norm": 0.04523796588182449,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 4.1950860023498535,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656195462478184,
+ "grad_norm": 0.04359745234251022,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 4.2136640548706055,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670157068062828,
+ "grad_norm": 0.047282423824071884,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 4.209680557250977,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68411867364747,
+ "grad_norm": 0.0474061481654644,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 4.188494682312012,
+ "step": 1417
+ },
+ {
+ "epoch": 19.698080279232112,
+ "grad_norm": 0.04402102902531624,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 4.207236289978027,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712041884816752,
+ "grad_norm": 0.04398655146360397,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 4.210899353027344,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726003490401396,
+ "grad_norm": 0.045767925679683685,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 4.171738624572754,
+ "step": 1420
+ },
+ {
+ "epoch": 19.739965095986037,
+ "grad_norm": 0.04498986527323723,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 4.222414970397949,
+ "step": 1421
+ },
+ {
+ "epoch": 19.75392670157068,
+ "grad_norm": 0.04282551258802414,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 4.202138900756836,
+ "step": 1422
+ },
+ {
+ "epoch": 19.767888307155324,
+ "grad_norm": 0.04392262175679207,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 4.184439659118652,
+ "step": 1423
+ },
+ {
+ "epoch": 19.781849912739965,
+ "grad_norm": 0.04343154653906822,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 4.204674243927002,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79581151832461,
+ "grad_norm": 0.04496419429779053,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 4.23316764831543,
+ "step": 1425
+ },
+ {
+ "epoch": 19.80977312390925,
+ "grad_norm": 0.04488370940089226,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 4.201659202575684,
+ "step": 1426
+ },
+ {
+ "epoch": 19.823734729493893,
+ "grad_norm": 0.045598775148391724,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 4.187412261962891,
+ "step": 1427
+ },
+ {
+ "epoch": 19.837696335078533,
+ "grad_norm": 0.043846048414707184,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 4.215444564819336,
+ "step": 1428
+ },
+ {
+ "epoch": 19.851657940663177,
+ "grad_norm": 0.043062254786491394,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 4.2074055671691895,
+ "step": 1429
+ },
+ {
+ "epoch": 19.865619546247817,
+ "grad_norm": 0.04544302448630333,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 4.184919357299805,
+ "step": 1430
+ },
+ {
+ "epoch": 19.87958115183246,
+ "grad_norm": 0.046144865453243256,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 4.2175750732421875,
+ "step": 1431
+ },
+ {
+ "epoch": 19.8935427574171,
+ "grad_norm": 0.04465445503592491,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 4.1851115226745605,
+ "step": 1432
+ },
+ {
+ "epoch": 19.907504363001745,
+ "grad_norm": 0.045913565903902054,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 4.2226152420043945,
+ "step": 1433
+ },
+ {
+ "epoch": 19.921465968586386,
+ "grad_norm": 0.04401300847530365,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 4.216772556304932,
+ "step": 1434
+ },
+ {
+ "epoch": 19.93542757417103,
+ "grad_norm": 0.042624711990356445,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 4.2133684158325195,
+ "step": 1435
+ },
+ {
+ "epoch": 19.949389179755673,
+ "grad_norm": 0.04533256217837334,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 4.218449592590332,
+ "step": 1436
+ },
+ {
+ "epoch": 19.963350785340314,
+ "grad_norm": 0.04764309898018837,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 4.191402912139893,
+ "step": 1437
+ },
+ {
+ "epoch": 19.977312390924958,
+ "grad_norm": 0.04689318686723709,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 4.2113037109375,
+ "step": 1438
+ },
+ {
+ "epoch": 19.991273996509598,
+ "grad_norm": 0.04479913413524628,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 4.194989204406738,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.03151386231184006,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 2.6285438537597656,
+ "step": 1440
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 6.149950262791373e+17,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-1440/training_args.bin b/runs/l2r90-ssl/checkpoint-1440/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..103048067ef8897d0386013ed89ce6764bad5521
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1440/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f320fb44ea3a4a93198f1a6c2405447428148ce142cd94ee283a3c077a327c2d
+size 4792
diff --git a/runs/l2r90-ssl/checkpoint-1800/chat_template.jinja b/runs/l2r90-ssl/checkpoint-1800/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1800/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-ssl/checkpoint-1800/config.json b/runs/l2r90-ssl/checkpoint-1800/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1800/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-ssl/checkpoint-1800/generation_config.json b/runs/l2r90-ssl/checkpoint-1800/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1800/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-ssl/checkpoint-1800/model.safetensors b/runs/l2r90-ssl/checkpoint-1800/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..f571d1a160e2037074528f17cb170a50a210be4c
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1800/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9d8d57d927c32f53c708a12353ae8f46375adc287beb0f15f6219dad866b5a0f
+size 583356232
diff --git a/runs/l2r90-ssl/checkpoint-1800/optimizer.pt b/runs/l2r90-ssl/checkpoint-1800/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..1d0e067e9fc07662ed8f00582c06e998cb032404
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1800/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:32eff8db573b44e702061c3e04f899dba31f83910de192610505e1793ef85b3f
+size 1166827898
diff --git a/runs/l2r90-ssl/checkpoint-1800/rng_state_0.pth b/runs/l2r90-ssl/checkpoint-1800/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..7d6b18c29f439e3010a8b54373d317696d66573f
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1800/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bedcb09a1c8950a65096d4085b70ba38f132583fa00a66de92fcc356e19a09d9
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-1800/rng_state_1.pth b/runs/l2r90-ssl/checkpoint-1800/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..3989a9d6a9b159dc1be67742f3434f6c50d1fb5f
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1800/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:68d1b7677dc8ab3e76010e23c7c2711560a2616f56d2a2907e1a8544b829f4e7
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-1800/scheduler.pt b/runs/l2r90-ssl/checkpoint-1800/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..6e0ba2afe167e5831737cd4b431b166033f17fab
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1800/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ee9cab552a4cabfd465ec6f079c65fbe7304d4a3df75e2e2b3e2eb614e447c70
+size 1064
diff --git a/runs/l2r90-ssl/checkpoint-1800/tokenizer.json b/runs/l2r90-ssl/checkpoint-1800/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1800/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-ssl/checkpoint-1800/tokenizer_config.json b/runs/l2r90-ssl/checkpoint-1800/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1800/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-1800/trainer_state.json b/runs/l2r90-ssl/checkpoint-1800/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..2b1ac567226576a14285894018dd93622bc8bdcd
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1800/trainer_state.json
@@ -0,0 +1,12818 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 25.0,
+ "eval_steps": 500,
+ "global_step": 1800,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.2588357925415039,
+ "learning_rate": 0.0,
+ "loss": 3.9982247352600098,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.25455471873283386,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.998683452606201,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.2555367052555084,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.9821574687957764,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.2383691370487213,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.9531917572021484,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2023179680109024,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.9148876667022705,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.17691479623317719,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.875638961791992,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.14372293651103973,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.836796283721924,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1074419692158699,
+ "learning_rate": 4.2e-05,
+ "loss": 3.8036489486694336,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.08204353600740433,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.774808406829834,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.0632256343960762,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.7503671646118164,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.047985248267650604,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.7345473766326904,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.03675895184278488,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.721525192260742,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.02908993512392044,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.713454246520996,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.02443614974617958,
+ "learning_rate": 7.8e-05,
+ "loss": 3.7078704833984375,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02109951712191105,
+ "learning_rate": 8.4e-05,
+ "loss": 3.7034401893615723,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.018616363406181335,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.7009854316711426,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.016399752348661423,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.697768211364746,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.014256286434829235,
+ "learning_rate": 0.000102,
+ "loss": 3.6974306106567383,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.013020666316151619,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.694406509399414,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.011879238300025463,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.691807270050049,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.010290366597473621,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.6937897205352783,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.009482803754508495,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.6902384757995605,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.008425266481935978,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.686815023422241,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.007676936220377684,
+ "learning_rate": 0.000138,
+ "loss": 3.685192823410034,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.007491032127290964,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.6839442253112793,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.007428477052599192,
+ "learning_rate": 0.00015,
+ "loss": 3.6783409118652344,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.006965892389416695,
+ "learning_rate": 0.000156,
+ "loss": 3.6813433170318604,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.008106694556772709,
+ "learning_rate": 0.000162,
+ "loss": 3.6741726398468018,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.0076780058443546295,
+ "learning_rate": 0.000168,
+ "loss": 3.67673921585083,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.007904314436018467,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.6724467277526855,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.008422375656664371,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.673013925552368,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00811714306473732,
+ "learning_rate": 0.000186,
+ "loss": 3.6698389053344727,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.009063836187124252,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.668990135192871,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.01878332532942295,
+ "learning_rate": 0.000198,
+ "loss": 3.6658849716186523,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.01812782697379589,
+ "learning_rate": 0.000204,
+ "loss": 3.6697864532470703,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.020587686449289322,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.662757158279419,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.01515690516680479,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.6558732986450195,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.02510508894920349,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.6495113372802734,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.020487826317548752,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.638803243637085,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.017446046695113182,
+ "learning_rate": 0.000234,
+ "loss": 3.633307456970215,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.030667584389448166,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.6277506351470947,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.023789944127202034,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.6189804077148438,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.02660694345831871,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.610522747039795,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.023814303800463676,
+ "learning_rate": 0.000258,
+ "loss": 3.615910291671753,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.021863561123609543,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.612008571624756,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03705216199159622,
+ "learning_rate": 0.00027,
+ "loss": 3.6066887378692627,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.024143919348716736,
+ "learning_rate": 0.000276,
+ "loss": 3.5987331867218018,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.03251276910305023,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.595367431640625,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06619632989168167,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.59018611907959,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.042714253067970276,
+ "learning_rate": 0.000294,
+ "loss": 3.58981990814209,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.018893254920840263,
+ "learning_rate": 0.0003,
+ "loss": 3.578472375869751,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0749685987830162,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.5828654766082764,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.12142857164144516,
+ "learning_rate": 0.000312,
+ "loss": 3.5840892791748047,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.06944665312767029,
+ "learning_rate": 0.000318,
+ "loss": 3.580961227416992,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.04545000568032265,
+ "learning_rate": 0.000324,
+ "loss": 3.5674569606781006,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.061073049902915955,
+ "learning_rate": 0.00033,
+ "loss": 3.569991111755371,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.04133498668670654,
+ "learning_rate": 0.000336,
+ "loss": 3.5657706260681152,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.055982090532779694,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.559979200363159,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.05259208381175995,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.558039426803589,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.029976001009345055,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.5491275787353516,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.04998327046632767,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.5406670570373535,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.03230690956115723,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.5391530990600586,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.06102178618311882,
+ "learning_rate": 0.000372,
+ "loss": 3.5311684608459473,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.031715717166662216,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.518899440765381,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.06841139495372772,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.514200210571289,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.10323675721883774,
+ "learning_rate": 0.00039,
+ "loss": 3.513646125793457,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.1448228508234024,
+ "learning_rate": 0.000396,
+ "loss": 3.5110087394714355,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.12129563093185425,
+ "learning_rate": 0.000402,
+ "loss": 3.501497507095337,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.030224842950701714,
+ "learning_rate": 0.000408,
+ "loss": 3.4870352745056152,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.07492101192474365,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.482234001159668,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.03942760080099106,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.4746525287628174,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.04967200756072998,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.604069232940674,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8671882152557373,
+ "eval_runtime": 42.6602,
+ "eval_samples_per_second": 57.243,
+ "eval_steps_per_second": 0.469,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.05385325849056244,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.462052822113037,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.04870929941534996,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.459104537963867,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.08701689541339874,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.4533209800720215,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.12219952791929245,
+ "learning_rate": 0.00045,
+ "loss": 3.4558796882629395,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.0823654904961586,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.443049192428589,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.12765446305274963,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.4373486042022705,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.14215047657489777,
+ "learning_rate": 0.000468,
+ "loss": 3.449429988861084,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.03649887070059776,
+ "learning_rate": 0.000474,
+ "loss": 3.42877197265625,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09551949054002762,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.4352874755859375,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.030483486130833626,
+ "learning_rate": 0.000486,
+ "loss": 3.4125874042510986,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.0745147094130516,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.4153835773468018,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.06620422005653381,
+ "learning_rate": 0.000498,
+ "loss": 3.4053854942321777,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.06098725646734238,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.404346466064453,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.047781601548194885,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.3991761207580566,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.04409855231642723,
+ "learning_rate": 0.000516,
+ "loss": 3.396144390106201,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.04355994611978531,
+ "learning_rate": 0.000522,
+ "loss": 3.3928911685943604,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.0670008510351181,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.3831443786621094,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.0812622532248497,
+ "learning_rate": 0.000534,
+ "loss": 3.382413148880005,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.0632157102227211,
+ "learning_rate": 0.00054,
+ "loss": 3.3784892559051514,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.028686825186014175,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.3673207759857178,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.034818850457668304,
+ "learning_rate": 0.000552,
+ "loss": 3.3648972511291504,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.04509204626083374,
+ "learning_rate": 0.000558,
+ "loss": 3.35561203956604,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.054211195558309555,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.3612167835235596,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.03752049803733826,
+ "learning_rate": 0.00057,
+ "loss": 3.357103109359741,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.03620241954922676,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.3469700813293457,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.02454194985330105,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.3429036140441895,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.041030462831258774,
+ "learning_rate": 0.000588,
+ "loss": 3.338247299194336,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.03266424685716629,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.332688331604004,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.04730091616511345,
+ "learning_rate": 0.0006,
+ "loss": 3.335902452468872,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.065738245844841,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.3289852142333984,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.09062394499778748,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.3278567790985107,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.1482805460691452,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.332685947418213,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.20066532492637634,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.363180637359619,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.11500677466392517,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.3308753967285156,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.11754138022661209,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.3383703231811523,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.09672239422798157,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.3267624378204346,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.07190799713134766,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.317737579345703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.0815449133515358,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.3198161125183105,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09612327069044113,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.315814971923828,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.14795289933681488,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.317605495452881,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06667589396238327,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.3127357959747314,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.058537885546684265,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.319807767868042,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.06428155303001404,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.297541618347168,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.03881775587797165,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.297539710998535,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.061801470816135406,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.290297508239746,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.07642851769924164,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.284428596496582,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08123870939016342,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.2820234298706055,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07174684852361679,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.2697296142578125,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.04030010476708412,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.264328956604004,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.05662153288722038,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.2699337005615234,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.03761111572384834,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.255704879760742,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.043027132749557495,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.2579030990600586,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.044409338384866714,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.248547077178955,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.0239730142056942,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.249307632446289,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03628705441951752,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.2451741695404053,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.029564054682850838,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.238445997238159,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.032036975026130676,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.2320892810821533,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.030085794627666473,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.231468915939331,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.025232303887605667,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.232199192047119,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.027053389698266983,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.210745334625244,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.028464114293456078,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.210076332092285,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04616783559322357,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.2103490829467773,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.039081938564777374,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.202662467956543,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.03428936004638672,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.1997475624084473,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.04351917654275894,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.19844913482666,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.0840945914387703,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.2013347148895264,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.36558300256729126,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.26310396194458,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.09432770311832428,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.2881736755371094,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.061970267444849014,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.2840280532836914,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.05112256482243538,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.2632040977478027,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06797203421592712,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.254606246948242,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.051026541739702225,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.429918050765991,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.8117228746414185,
+ "eval_runtime": 42.8572,
+ "eval_samples_per_second": 56.98,
+ "eval_steps_per_second": 0.467,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.0608622208237648,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.2348179817199707,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.05581629276275635,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.22896146774292,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.09162136912345886,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.2343664169311523,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.13750384747982025,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.2507717609405518,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.0872517079114914,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.2440056800842285,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.07161762565374374,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.229494571685791,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04082738608121872,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.2183239459991455,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.04972497373819351,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.218075752258301,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.049254074692726135,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.213379383087158,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.056671254336833954,
+ "learning_rate": 0.000599663984012462,
+ "loss": 3.209568977355957,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.05541909113526344,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.207760810852051,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.03819117695093155,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 3.199732780456543,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.0529443584382534,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.1945419311523438,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.043010979890823364,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.185936212539673,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.07311078161001205,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.189894914627075,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06110258400440216,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 3.1911773681640625,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.06880149245262146,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 3.187437057495117,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.051326874643564224,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 3.1826558113098145,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.03363822400569916,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 3.175527572631836,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.10268989205360413,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 3.172367572784424,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.09723010659217834,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 3.1873984336853027,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.04129317030310631,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 3.165799617767334,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.05292229354381561,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 3.171780824661255,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.03873143345117569,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 3.1680874824523926,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.047018859535455704,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 3.161402940750122,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.03942955285310745,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 3.1530532836914062,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.039801858365535736,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 3.1465344429016113,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.03795155510306358,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 3.1364030838012695,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.0342850424349308,
+ "learning_rate": 0.000599379982511273,
+ "loss": 3.1391525268554688,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.03243280574679375,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 3.1331253051757812,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.029679352417588234,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 3.1346120834350586,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.030230293050408363,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 3.132615804672241,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.02845083735883236,
+ "learning_rate": 0.000599309205504528,
+ "loss": 3.1242547035217285,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.03208369016647339,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 3.119110584259033,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.040268607437610626,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 3.113833427429199,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.06951889395713806,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 3.1186914443969727,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.08879823237657547,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 3.1140987873077393,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07374681532382965,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 3.1147828102111816,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.038190409541130066,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 3.1161253452301025,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.03297388181090355,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 3.1006083488464355,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.05905177444219589,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 3.106564521789551,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.037567321211099625,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 3.0884456634521484,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.02725658379495144,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 3.09000825881958,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.04272424429655075,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 3.090528964996338,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.027587950229644775,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 3.085131883621216,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.03282274678349495,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 3.0880112648010254,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.04064026474952698,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 3.079259157180786,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.023374097421765327,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 3.0747299194335938,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.025003870949149132,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 3.0721259117126465,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.030445197597146034,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 3.0733323097229004,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.025935636833310127,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 3.070709228515625,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.021337421610951424,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 3.0627551078796387,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.040481697767972946,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 3.0529394149780273,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.047608256340026855,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 3.041020631790161,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.03356955572962761,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 3.0532307624816895,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.02101682312786579,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 3.0492985248565674,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.030098581686615944,
+ "learning_rate": 0.000598804365316586,
+ "loss": 3.0527801513671875,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.040166422724723816,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 3.0505762100219727,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.051579318940639496,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 3.0422110557556152,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.06485334783792496,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 3.0435101985931396,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.055590901523828506,
+ "learning_rate": 0.000598706872384039,
+ "loss": 3.0448739528656006,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.03183475881814957,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 3.0416858196258545,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.03995629772543907,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 3.0255980491638184,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.03907827287912369,
+ "learning_rate": 0.000598631251036868,
+ "loss": 3.032339096069336,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.027681387960910797,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 3.025439739227295,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.03470184653997421,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 3.02883243560791,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.031407181173563004,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 3.0289883613586426,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.0371522456407547,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 3.013608932495117,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.03651345893740654,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 3.0120253562927246,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.06004595756530762,
+ "learning_rate": 0.000598473577921366,
+ "loss": 3.017958402633667,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.09979518502950668,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 3.023041248321533,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.10138462483882904,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.275729179382324,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7571707367897034,
+ "eval_runtime": 42.8283,
+ "eval_samples_per_second": 57.018,
+ "eval_steps_per_second": 0.467,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.04982660710811615,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 3.016209602355957,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.06532539427280426,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 3.015043020248413,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.04369623214006424,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 3.0096778869628906,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.05180002748966217,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 3.0121543407440186,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.0440233014523983,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.9937305450439453,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.05645671859383583,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 3.0023865699768066,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.03990716487169266,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 3.0003232955932617,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.03818608075380325,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.994137763977051,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.035388972610235214,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.993170738220215,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.03162284567952156,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.9922614097595215,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.029497336596250534,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.9852986335754395,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.02586926706135273,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.9798672199249268,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.027355050668120384,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.985158920288086,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.0341978520154953,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.982581615447998,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.030727429315447807,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.9701411724090576,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.04429870471358299,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.9789254665374756,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.05775132030248642,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.9758143424987793,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.05635962635278702,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.981480121612549,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.03639192506670952,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.968752861022949,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.028748342767357826,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.9584972858428955,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.035653006285429,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.967545747756958,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.03697885945439339,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.9578099250793457,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.03518654406070709,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.9680328369140625,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.034762222319841385,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.9670262336730957,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03813016042113304,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.957522392272949,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.04372904822230339,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.9540514945983887,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.06646446138620377,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.951028823852539,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.07831092923879623,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.95898699760437,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.05779588967561722,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.965884208679199,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.04414296895265579,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.9510560035705566,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.04797815904021263,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.955803394317627,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.04482145980000496,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.954059362411499,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.037104446440935135,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.94792103767395,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.03616701066493988,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.9500932693481445,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.03882797807455063,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.9558305740356445,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.04356204345822334,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.9237029552459717,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.04535257816314697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.925853729248047,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.04519587755203247,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.924984931945801,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.055922407656908035,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.932561159133911,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.06330762803554535,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.9388322830200195,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.0517115443944931,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.9272942543029785,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.057369694113731384,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.9201431274414062,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.05687880516052246,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.922886371612549,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.055525410920381546,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.9211201667785645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04503652825951576,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.9164583683013916,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.03385302051901817,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.9145395755767822,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03564384952187538,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.916003704071045,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.04188712686300278,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.915839433670044,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03657370060682297,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.903120279312134,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.03126626834273338,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.895824432373047,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.04196290299296379,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.9093422889709473,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.1706087589263916,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.9183316230773926,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.05810421332716942,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.911245107650757,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.061330314725637436,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.9159231185913086,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.06062281131744385,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.9211535453796387,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.05590800568461418,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.9140496253967285,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.04496239125728607,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.9040920734405518,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.03835335746407509,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.9026548862457275,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.05137286335229874,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.889634132385254,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.04091101884841919,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.886838674545288,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.04076292738318443,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.892918109893799,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.045437734574079514,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.893305778503418,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.041114673018455505,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.88517165184021,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.03600773215293884,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.8779795169830322,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.04311220720410347,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.87247896194458,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.04818756878376007,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.889103412628174,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.053460877388715744,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.8899683952331543,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.06763825565576553,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.8823537826538086,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.07304801046848297,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.877340316772461,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.061337269842624664,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.8820111751556396,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.04660911113023758,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.868886709213257,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.053929202258586884,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.159337043762207,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.721762478351593,
+ "eval_runtime": 42.7977,
+ "eval_samples_per_second": 57.059,
+ "eval_steps_per_second": 0.467,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.04259369149804115,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.859579086303711,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.044425517320632935,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.8681139945983887,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.04301869496703148,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.853911876678467,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.04574296995997429,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.8724427223205566,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.050247322767972946,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.8528172969818115,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.046683840453624725,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.853405714035034,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.05966169387102127,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.8605127334594727,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.08935574442148209,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.862339973449707,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.08019179105758667,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.8655967712402344,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.04753892868757248,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.8657939434051514,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06868117302656174,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.861976146697998,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.07040470093488693,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.8562135696411133,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.05683179199695587,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.8576629161834717,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.04918001592159271,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.861943244934082,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.04921915754675865,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.8495378494262695,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.037651967257261276,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.8428618907928467,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.03522057458758354,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.8372769355773926,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.03111186809837818,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.8377151489257812,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.03906773030757904,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.846196174621582,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.02546556107699871,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.8288016319274902,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.03119821846485138,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.8334617614746094,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.030583377927541733,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.8209362030029297,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.029693076387047768,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.8267486095428467,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.028978517279028893,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.819244861602783,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.027871523052453995,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.8289647102355957,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.029287064447999,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.8235747814178467,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.024990517646074295,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.819016218185425,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.030980689451098442,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.828155040740967,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03721801936626434,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.813418388366699,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03738148882985115,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.808328866958618,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.03580411896109581,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.812070846557617,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.039036594331264496,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.807159423828125,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.044666968286037445,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.799992561340332,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.055589817464351654,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.804142951965332,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.06139020994305611,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.8182730674743652,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.07769318670034409,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.8166518211364746,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.06521332263946533,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.8142249584198,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.09083619713783264,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.8166933059692383,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.09737467020750046,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.8111555576324463,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.07148188352584839,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.8168413639068604,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.055516090244054794,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.8183553218841553,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.05338391661643982,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.8127028942108154,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.040785472840070724,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.8124775886535645,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.037135496735572815,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.8065357208251953,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.032455578446388245,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.798459053039551,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.03465662896633148,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.798649549484253,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.032026033848524094,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.7906997203826904,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.026019956916570663,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.791969060897827,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.030298635363578796,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.784593105316162,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.028147678822278976,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.782454013824463,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.023933131247758865,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.787851333618164,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.026789527386426926,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.788067579269409,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.024554884061217308,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.777033567428589,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.028679484501481056,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.786342144012451,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.026410218328237534,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.7790610790252686,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.027413051575422287,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.778944969177246,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.02908242493867874,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.763850212097168,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.029632994905114174,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.7672982215881348,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.027458716183900833,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.7702178955078125,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.03263958916068077,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.7581803798675537,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04427146911621094,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.768141508102417,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.07142394036054611,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.75819730758667,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.08833584934473038,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.782716751098633,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.08168833702802658,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.7763240337371826,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07253874838352203,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.7717690467834473,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.08186902105808258,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.767303466796875,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.05854064226150513,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.761544704437256,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.04602917283773422,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.7669177055358887,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.03920970857143402,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.766085624694824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.04551195353269577,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.7623450756073,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.040861159563064575,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.7471840381622314,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.026907291263341904,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 2.0710244178771973,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6934371590614319,
+ "eval_runtime": 42.6015,
+ "eval_samples_per_second": 57.322,
+ "eval_steps_per_second": 0.469,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.03801899403333664,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.739086627960205,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.03707238286733627,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.741638422012329,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.0387551411986351,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.740777015686035,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.028010359033942223,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.7353010177612305,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.026474518701434135,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.73842453956604,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.02920856513082981,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.736295461654663,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.026074763387441635,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.7295751571655273,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.026862047612667084,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.723040819168091,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.026300370693206787,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.7199485301971436,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.026030950248241425,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.7276217937469482,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.028342435136437416,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.7193045616149902,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.032135091722011566,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.7143118381500244,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.03068603202700615,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.7191050052642822,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.035828713327646255,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.718085765838623,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.03343002125620842,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.707131862640381,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.035051796585321426,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.7091879844665527,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.03395913913846016,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6980862617492676,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.034198079258203506,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.699704647064209,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.0322839580476284,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.702791213989258,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.041531339287757874,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.7018051147460938,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.040806882083415985,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.7096598148345947,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.04156940430402756,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6956775188446045,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.05608797073364258,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.70127272605896,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0729481428861618,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.708817720413208,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.0905836969614029,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.718085289001465,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.08345075696706772,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.7235910892486572,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.08599896728992462,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.7178783416748047,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.07301691174507141,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.714029312133789,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.08231015503406525,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.7196826934814453,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.08286412060260773,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.7176082134246826,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.08590883016586304,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.7242164611816406,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.07531397044658661,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.7225024700164795,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.04711190238595009,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.695887327194214,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.05076327919960022,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.7029781341552734,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.04495669901371002,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.7057900428771973,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.04312257468700409,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.697537899017334,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.043580975383520126,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.67315673828125,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.03360886126756668,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.6883413791656494,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.030360296368598938,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.700927257537842,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.0310810673981905,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.6917803287506104,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.02703232690691948,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.6667580604553223,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.02756045199930668,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.6702821254730225,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.02611083723604679,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.669790744781494,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.02615324780344963,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.6672821044921875,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.026000119745731354,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.6671881675720215,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.02882794849574566,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.67586088180542,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03327213600277901,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.6618690490722656,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.032920945435762405,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.665444850921631,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.02812962979078293,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.6680707931518555,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.025886602699756622,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.652898073196411,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.025222789496183395,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.6560399532318115,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.023131893947720528,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.66127347946167,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.02513662725687027,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.6463732719421387,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.027243830263614655,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.6507630348205566,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.025437965989112854,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.6485800743103027,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.02770340070128441,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.657241106033325,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.03594067320227623,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.6281137466430664,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04486451297998428,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.6340909004211426,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.05102410912513733,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.6463394165039062,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04514491558074951,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.636554002761841,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04469453915953636,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.6453680992126465,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.05421523004770279,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.646010398864746,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04779462888836861,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.6315622329711914,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.06697724014520645,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.6416730880737305,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.07520416378974915,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.641632318496704,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.055692676454782486,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.6289877891540527,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.04539895057678223,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.6363251209259033,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.04829065501689911,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.641608476638794,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.041314542293548584,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.6269643306732178,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.04100198298692703,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.6243174076080322,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.04105962812900543,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.626585006713867,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.03172122687101364,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9550899267196655,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6614077091217041,
+ "eval_runtime": 42.9009,
+ "eval_samples_per_second": 56.922,
+ "eval_steps_per_second": 0.466,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.03258531913161278,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.612126588821411,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.040250636637210846,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.6025328636169434,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.034234337508678436,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.6085972785949707,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.03635786101222038,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.6126222610473633,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.039404451847076416,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.6035327911376953,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.038504838943481445,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5985798835754395,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.03744236007332802,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5873115062713623,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.034184787422418594,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.5899314880371094,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.03766429424285889,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.602734327316284,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.03741588443517685,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.592207193374634,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.03187853842973709,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5850372314453125,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.038447409868240356,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.592654228210449,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.04046263173222542,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.5881710052490234,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.03872602432966232,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5881471633911133,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.03960775211453438,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5826945304870605,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.033132150769233704,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.584038734436035,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.029504230245947838,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.584357261657715,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.03665253147482872,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5677168369293213,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.037208545953035355,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.560795783996582,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.038244519382715225,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.591458559036255,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.043307431042194366,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.5642588138580322,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.055119726806879044,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.57377290725708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.053323689848184586,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5767312049865723,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.0502864345908165,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.562190055847168,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.05568842589855194,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.565556049346924,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.061149828135967255,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5883288383483887,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.059933699667453766,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.5710184574127197,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.06041933596134186,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.572603225708008,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.05482695251703262,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.5667271614074707,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.04712968319654465,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.554041624069214,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.04842432215809822,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.5660672187805176,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.04273047298192978,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.554839849472046,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.05356749892234802,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.5521483421325684,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.06062676012516022,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.5551886558532715,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.05431462824344635,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5654542446136475,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.0505850687623024,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5648036003112793,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.04696856811642647,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.5559377670288086,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.042585279792547226,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.5660319328308105,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.04026484116911888,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.551206588745117,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03382421284914017,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.556361198425293,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.031177924945950508,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.5413894653320312,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.028217196464538574,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.5600497722625732,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.027062954381108284,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5447568893432617,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.03001212701201439,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.540614604949951,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.028869854286313057,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.544583797454834,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.02869918756186962,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.5471596717834473,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03359716013073921,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5416982173919678,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.04000182822346687,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.534921407699585,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.043030042201280594,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5368564128875732,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.04329848289489746,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.539764881134033,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.03740967437624931,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.5246665477752686,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.03733935207128525,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.529362916946411,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.039609868079423904,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.542544364929199,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.03937261551618576,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5366649627685547,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.03261927142739296,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.5250535011291504,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.03904294595122337,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.520721435546875,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.03913239762187004,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.5122013092041016,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.03610851988196373,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.5212767124176025,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.03944548964500427,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.537186622619629,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04762452840805054,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.5204620361328125,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04427091404795647,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5286149978637695,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04366641491651535,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.5398361682891846,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.04644155502319336,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.519331932067871,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.04372382536530495,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.531022787094116,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.04404324293136597,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.521758794784546,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.042115990072488785,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.5146074295043945,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.03796013817191124,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.5040054321289062,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.04327977076172829,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.5209312438964844,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04150385782122612,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.515871286392212,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.04465346783399582,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.5247764587402344,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.05690346285700798,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.501911163330078,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04907930642366409,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8968371152877808,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6358620524406433,
+ "eval_runtime": 42.9991,
+ "eval_samples_per_second": 56.792,
+ "eval_steps_per_second": 0.465,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.03755255416035652,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.486576795578003,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.040234558284282684,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.501087188720703,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.041416723281145096,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4942233562469482,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.039655014872550964,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.503453254699707,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.041707683354616165,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.490978717803955,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.04545537009835243,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4757118225097656,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.04328389838337898,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.504516839981079,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.043312299996614456,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.4920215606689453,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.046466976404190063,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.494760513305664,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.041762981563806534,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4875552654266357,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.03713781014084816,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.495652675628662,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.040603458881378174,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4821407794952393,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.037969645112752914,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.484981060028076,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.03652074560523033,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4736409187316895,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.03783264383673668,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.483346700668335,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.03404324874281883,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.48856782913208,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.03294725716114044,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.483706474304199,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.0347001850605011,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.474888324737549,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.0359107106924057,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.478990077972412,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.03738103806972504,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.4709420204162598,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.03722188249230385,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.465726852416992,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.034907322376966476,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.482231855392456,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.035695310682058334,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4785828590393066,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.04009179025888443,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.4795875549316406,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.038996778428554535,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.478696823120117,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04082407429814339,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.465200901031494,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04856136813759804,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4605681896209717,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.05671004578471184,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.47277569770813,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.0481606163084507,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.476151466369629,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.037689510732889175,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.466142177581787,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.05005146935582161,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.4798035621643066,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.04745829105377197,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.46163272857666,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.04638158157467842,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4728071689605713,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.04530330374836922,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.4621877670288086,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.03971488028764725,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.474224090576172,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.04521876201033592,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.46122670173645,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03923020884394646,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.4603679180145264,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03986494988203049,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.4589664936065674,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.04280020296573639,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.4624319076538086,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.03930043429136276,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.47080135345459,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.0314876064658165,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.458728790283203,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.03189660236239433,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4467148780822754,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03264275938272476,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4613699913024902,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03021484799683094,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.4535279273986816,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.030740290880203247,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.4621875286102295,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03096810355782509,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.458500385284424,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03105205111205578,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.461958646774292,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.03330068290233612,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.449777364730835,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.032831475138664246,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.4616191387176514,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.0387776717543602,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.445256471633911,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.041803423315286636,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.4545960426330566,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.04148292914032936,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.433408737182617,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.04648579657077789,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.449134349822998,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.04186409339308739,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.4449565410614014,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.038569554686546326,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.443675994873047,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03908300772309303,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4626526832580566,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.038897737860679626,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.438591957092285,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.0360480472445488,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.436077117919922,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0429830439388752,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4384970664978027,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.04465422406792641,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.458029270172119,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.037096526473760605,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.436816930770874,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.047823466360569,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.44439959526062,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.05929310992360115,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.4599215984344482,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.053512319922447205,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4387903213500977,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.042625587433576584,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4359588623046875,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.056042976677417755,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.4441447257995605,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05813189223408699,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4399144649505615,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.06313478201627731,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.453447103500366,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.05281798169016838,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.4428963661193848,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.03991910442709923,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4359219074249268,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.047981832176446915,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4506607055664062,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.03646523132920265,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8225739002227783,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6196661591529846,
+ "eval_runtime": 43.8106,
+ "eval_samples_per_second": 55.74,
+ "eval_steps_per_second": 0.457,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.03673036769032478,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.4227209091186523,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.044547371566295624,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.4073169231414795,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.04089285805821419,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4279446601867676,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.03538106009364128,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.4234392642974854,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.03152978792786598,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4214606285095215,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.03324202075600624,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.4085447788238525,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.03352442383766174,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.417776107788086,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.03582802042365074,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.411144256591797,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.0351114459335804,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.405872344970703,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.038898173719644547,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4029338359832764,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.03926656022667885,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4100089073181152,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.04285634309053421,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.410360336303711,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.04179581627249718,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4093637466430664,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.03356701880693436,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4118056297302246,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.038140635937452316,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.4154677391052246,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.04468780383467674,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416846513748169,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.062421441078186035,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4157638549804688,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06317313015460968,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.418593406677246,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.049498945474624634,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4204845428466797,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.046322405338287354,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.403489589691162,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.053081002086400986,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.4148740768432617,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05222945287823677,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.416757583618164,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04893574118614197,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4241795539855957,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04951956123113632,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.401386260986328,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04019852355122566,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.4168622493743896,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.03625848889350891,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4026846885681152,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.03464345261454582,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4021215438842773,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.03750818222761154,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4113571643829346,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.0360100194811821,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.3977622985839844,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.03464623540639877,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.413273334503174,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.03193575143814087,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.423659324645996,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.02710454724729061,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.396085262298584,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.02830432914197445,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.420255422592163,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.029717663303017616,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.3958120346069336,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.030005933716893196,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3908543586730957,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.029054716229438782,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.3873138427734375,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.031556472182273865,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3892760276794434,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03518872708082199,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4113478660583496,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.03471249341964722,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.384857654571533,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03223155066370964,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3840765953063965,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.030012644827365875,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.388247489929199,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.03162285313010216,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.3987512588500977,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.03700442612171173,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.3993430137634277,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.03710822016000748,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.387359380722046,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.028411181643605232,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3925249576568604,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031246855854988098,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.38845157623291,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.034664880484342575,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.392436981201172,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.03354701027274132,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.390890121459961,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.03527764603495598,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.3954997062683105,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03325049206614494,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3847177028656006,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.028021875768899918,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.385910987854004,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.02964954823255539,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3787484169006348,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.03068520873785019,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.400784969329834,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03416873514652252,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.391033411026001,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.03534875437617302,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.3796474933624268,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.03732474148273468,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3831934928894043,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.03648678958415985,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3785600662231445,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03801294043660164,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.387587070465088,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.04186025261878967,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.3893089294433594,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.04416101053357124,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3852195739746094,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.04137178510427475,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.391308307647705,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.05010323226451874,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.3906192779541016,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.049587685614824295,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.3846192359924316,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.04125971347093582,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.384507179260254,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.039676666259765625,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.38167142868042,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.038775283843278885,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.3789167404174805,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.03634221479296684,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.383060932159424,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03464275971055031,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.388552665710449,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.031778786331415176,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.3903298377990723,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.027904309332370758,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.377493381500244,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.028964119032025337,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.381959915161133,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.0233962032943964,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7810559272766113,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6070130467414856,
+ "eval_runtime": 50.871,
+ "eval_samples_per_second": 48.004,
+ "eval_steps_per_second": 0.393,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.030209315940737724,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.3569841384887695,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.04112662002444267,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.376680374145508,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.055858466774225235,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.3621644973754883,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.06746986508369446,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3607563972473145,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.06736797094345093,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3548707962036133,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.0562627948820591,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.364654541015625,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.051833927631378174,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.3748884201049805,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.04824110120534897,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.3502087593078613,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.05518793687224388,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3687515258789062,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.06564056873321533,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.367131233215332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.0605909489095211,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.3589868545532227,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.05744732916355133,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.370117664337158,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.06430432945489883,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.3703484535217285,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.06186915934085846,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3720293045043945,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.05977998673915863,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.3512699604034424,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.06413097679615021,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.377257823944092,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.053935080766677856,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.350717544555664,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.049110863357782364,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.3688902854919434,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.04313219338655472,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3374459743499756,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.039556123316287994,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3482556343078613,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.03796547278761864,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3547937870025635,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.03646104037761688,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.344686269760132,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.03355357423424721,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.3588008880615234,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.03004707582294941,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.352043628692627,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.030186176300048828,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3500709533691406,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.029748762026429176,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.3523058891296387,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.03238696977496147,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3635129928588867,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.032599564641714096,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.346269130706787,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.028626957908272743,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.357151746749878,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.02909751422703266,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3557677268981934,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.029951099306344986,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.34753680229187,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.029768310487270355,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3408985137939453,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.03005973994731903,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.362542152404785,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03360462561249733,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.352529287338257,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.034002047032117844,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.3334460258483887,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.03292316198348999,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3446991443634033,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.0337117575109005,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.350605010986328,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03428212180733681,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.357408046722412,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.03169601410627365,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.349088191986084,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.03389411047101021,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.361576795578003,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.036497049033641815,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.35420560836792,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03574785590171814,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3678526878356934,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.030757993459701538,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3520517349243164,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.0289121363312006,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.337864637374878,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.028012607246637344,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.353999614715576,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.027558492496609688,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.342132568359375,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.02871510572731495,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3425655364990234,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03070886805653572,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3401665687561035,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.03056657873094082,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.345277786254883,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.030907969921827316,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.346494674682617,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0318334735929966,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.3390543460845947,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.029700733721256256,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.3398728370666504,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.029687359929084778,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.3474364280700684,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.027234600856900215,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3457484245300293,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.02823043428361416,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.337700366973877,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03002646192908287,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.335954189300537,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.027246717363595963,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.330207347869873,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.029703574255108833,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.328280448913574,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.028796236962080002,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3448386192321777,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.027944380417466164,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.320234537124634,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.030626650899648666,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.334698438644409,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.029164662584662437,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3450326919555664,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03034045360982418,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.3420588970184326,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.02856968529522419,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3396072387695312,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.02892562374472618,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.340027332305908,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.02938242256641388,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.33550763130188,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.029261089861392975,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3246421813964844,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.030340535566210747,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.3465514183044434,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.033335570245981216,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.3467206954956055,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.03452218323945999,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.343702793121338,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.03326255828142166,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.337631940841675,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025433633476495743,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7402067184448242,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03175730258226395,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.62623405456543,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.04780422896146774,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.623239517211914,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.0504758283495903,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.6369194984436035,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.04481920227408409,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.57640266418457,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0440552793443203,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.630992889404297,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0503336526453495,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.627910137176514,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.053885314613580704,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.637897491455078,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.05481956899166107,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.610715866088867,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.050510745495557785,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.639884948730469,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.04579266160726547,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.590188980102539,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.04817410930991173,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.636295318603516,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.058862220495939255,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.614317417144775,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.06328638643026352,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.655928611755371,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.058752477169036865,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.652004241943359,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.06049015372991562,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.6513800621032715,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.058941151946783066,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.657688140869141,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.05542639642953873,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.637376308441162,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06024057790637016,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.63592529296875,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.06040216609835625,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.6275787353515625,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.059600673615932465,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.652995586395264,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.05379730463027954,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.597358226776123,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05355260521173477,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.6106367111206055,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.06649646162986755,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.654443740844727,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.058309994637966156,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.6869659423828125,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.05403570458292961,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.654750823974609,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04678157716989517,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.653955459594727,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04297948628664017,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.644915580749512,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.04235142841935158,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.662282943725586,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.0405130498111248,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.64569091796875,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.03351007401943207,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.653165817260742,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.03182631731033325,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.606291770935059,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.031552448868751526,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.642643928527832,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027941036969423294,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.623664379119873,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.027744563296437263,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.605886459350586,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02962736040353775,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.606762886047363,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.027251560240983963,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.620492935180664,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.027894044294953346,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.613903999328613,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.02778882533311844,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.634457588195801,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.028082096949219704,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.610980033874512,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.026210976764559746,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.612090110778809,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.026537755504250526,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.581171035766602,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.028709523379802704,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.628799915313721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.028233813121914864,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.586182117462158,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.02966046892106533,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.62320613861084,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.03198636695742607,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.6018524169921875,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.033404890447854996,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.610121726989746,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.029217196628451347,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.601312160491943,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.029335757717490196,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.620820999145508,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.031007781624794006,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.626503944396973,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.028453458100557327,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.589323997497559,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02526312880218029,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.598350524902344,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.027317512780427933,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.62388277053833,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02710605598986149,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.606523513793945,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.027491388842463493,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.586644172668457,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.027158567681908607,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.606039047241211,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.0298276599496603,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.593776702880859,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.026622479781508446,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.59510612487793,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.028238043189048767,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.5899505615234375,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.029481008648872375,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.6092071533203125,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.030138323083519936,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.600069046020508,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.02951825223863125,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.600125789642334,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.029242202639579773,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.590723037719727,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.02941613830626011,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.623342514038086,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.029764151200652122,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.631185531616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.031503792852163315,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.605094909667969,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.03184918314218521,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.592767715454102,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.03223618119955063,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.5930585861206055,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.0340263694524765,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.609347343444824,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.03245621174573898,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.598440170288086,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.03204867243766785,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.619702339172363,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.030866602435708046,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.609165191650391,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.01960863545536995,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8655829429626465,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.593852162361145,
+ "eval_runtime": 60.0971,
+ "eval_samples_per_second": 40.634,
+ "eval_steps_per_second": 0.649,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03552858531475067,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.539240837097168,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.0536109134554863,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.52609395980835,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.061411239206790924,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.536109447479248,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.06352879852056503,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.546850204467773,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.05683520436286926,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.563828468322754,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.05522637069225311,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.550929069519043,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.058990392833948135,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.554411888122559,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.057443976402282715,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.545921325683594,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.05383576825261116,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.5481414794921875,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04797440022230148,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.55556583404541,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04923664405941963,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.545513153076172,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.056191347539424896,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.536705017089844,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.06088989973068237,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.570624351501465,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.04802761599421501,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.526930809020996,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.04018320515751839,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.562432289123535,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.04280981421470642,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.559737205505371,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.04258878529071808,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.561749458312988,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.04411563277244568,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.562010765075684,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.041077401489019394,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.573984622955322,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.04125386103987694,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.550564289093018,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.041974026709795,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.552739143371582,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03689907118678093,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.540107727050781,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03586242347955704,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.554999351501465,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.036602675914764404,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.546337127685547,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.03707926720380783,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.543377876281738,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.03322125971317291,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.574325084686279,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03359243646264076,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.5613603591918945,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.0314965583384037,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.530994415283203,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.031806789338588715,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.56138801574707,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.034988485276699066,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.553149700164795,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.037627559155225754,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.555401802062988,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03723530471324921,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.52452278137207,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.036220453679561615,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.54799747467041,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03228041157126427,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.574814319610596,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.030872676521539688,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.539666652679443,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.033935241401195526,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.579252243041992,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.034338828176259995,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.538324356079102,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.03441951423883438,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.540358543395996,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03328141197562218,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.520340919494629,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03224223107099533,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.5646209716796875,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.02999890223145485,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.504915714263916,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.03014362044632435,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.516615867614746,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.029469173401594162,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.560475826263428,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.027887821197509766,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.586333751678467,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.029970254749059677,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.530396461486816,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.031648170202970505,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.554404258728027,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.0325581431388855,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.544102668762207,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.03555625304579735,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.543701171875,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.033053912222385406,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.508375644683838,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.03059076890349388,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.545750617980957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.029640182852745056,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.54715633392334,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030139951035380363,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.533708572387695,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029753828421235085,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.543603897094727,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.029481034725904465,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.534200668334961,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.028991524130105972,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.5683674812316895,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.028145765885710716,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.543053150177002,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.02999844029545784,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.5482354164123535,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03134117275476456,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.5360107421875,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.029277140274643898,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.5729827880859375,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028516866266727448,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.543548583984375,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028387080878019333,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.552616596221924,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.028977863490581512,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.5456695556640625,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.028010815382003784,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.534921646118164,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.030222605913877487,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.561526298522949,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.03140007704496384,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.500124931335449,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.029269184917211533,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.536653995513916,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.02935507707297802,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.521398544311523,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02801917865872383,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.527442455291748,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.027995768934488297,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.556139945983887,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.028441084548830986,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.555442810058594,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.029100807383656502,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.532083511352539,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020385215058922768,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.8488006591796875,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5902590751647949,
+ "eval_runtime": 60.1424,
+ "eval_samples_per_second": 40.604,
+ "eval_steps_per_second": 0.648,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03513244539499283,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.486500263214111,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.052679311484098434,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.489691734313965,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.06171489134430885,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.471555709838867,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.06164781004190445,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.486891746520996,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.06410388648509979,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.494929313659668,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06399523466825485,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.5334320068359375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.0636502280831337,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.5039381980896,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.06139931082725525,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.512765884399414,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.07460282742977142,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.477229118347168,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.075187548995018,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.484212398529053,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.07584068924188614,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.495598793029785,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.07192158699035645,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.507598400115967,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.05366869643330574,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.514532089233398,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.06397061794996262,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.512920379638672,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.06391509622335434,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.465702056884766,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.059490568935871124,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.5351881980896,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.057501111179590225,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.510270595550537,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.056947533041238785,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.509241104125977,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.050393324345350266,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.50428581237793,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.057608287781476974,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.474919319152832,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.054470717906951904,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.521528244018555,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.054405540227890015,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.510211944580078,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.051328495144844055,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.480720520019531,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.045930180698633194,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.503024101257324,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.04298274591565132,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.49046516418457,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.0406254380941391,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.544064521789551,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.03557496517896652,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.496966361999512,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.0364493802189827,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.5047502517700195,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.034521300345659256,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.485365867614746,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.03305799141526222,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.484884738922119,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.033053625375032425,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.478784561157227,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03426017984747887,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.482428073883057,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.03638783469796181,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.433219909667969,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.03597250208258629,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.522584915161133,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.03086097538471222,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.500803470611572,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.029642771929502487,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.444099426269531,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.03190716728568077,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.493034362792969,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.0312873050570488,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.487887382507324,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03201943263411522,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.479177474975586,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.033088769763708115,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.5020246505737305,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.02975946106016636,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.455426216125488,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03213953226804733,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.501615047454834,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.03262908384203911,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.485121250152588,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.031102091073989868,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.488893508911133,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.02945449948310852,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.490121841430664,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.033680472522974014,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.51833438873291,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.03277525305747986,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.480165481567383,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.03303041681647301,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.501155376434326,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.034927863627672195,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.513175010681152,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.03098093718290329,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.503798484802246,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.029513906687498093,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.49006986618042,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029630785807967186,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.522514343261719,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.030888330191373825,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.4895734786987305,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.032392196357250214,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.471953868865967,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.03215661272406578,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.48297643661499,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.03346327692270279,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.518160820007324,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.031311362981796265,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.484991550445557,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.03056488186120987,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.525470733642578,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.03043230250477791,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.510650634765625,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.02916652522981167,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.496840953826904,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.030673664063215256,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.494868278503418,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030033906921744347,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.479182720184326,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.028828797861933708,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.471983909606934,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028316056355834007,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.5049147605896,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.03158092871308327,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.449270248413086,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.03041706420481205,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.508289813995361,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.02967553399503231,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.465579032897949,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030721619725227356,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.471003532409668,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03066832385957241,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.488987445831299,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.032297998666763306,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.491018295288086,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03221398591995239,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.471237659454346,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.021120507270097733,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.8080053329467773,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5888046026229858,
+ "eval_runtime": 59.1638,
+ "eval_samples_per_second": 41.275,
+ "eval_steps_per_second": 0.659,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.03799329698085785,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.401787757873535,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.05785121023654938,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.408820629119873,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.05964238941669464,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.427263259887695,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05466582626104355,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.403415679931641,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.05613672360777855,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.413750648498535,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.05396881327033043,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.41916561126709,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.05689029023051262,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.395590782165527,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.056090328842401505,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.421253204345703,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.0473211295902729,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.385158061981201,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.04464765638113022,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.431746959686279,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.046421997249126434,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.428816795349121,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.05039866268634796,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.452417850494385,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.05491664633154869,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.447483062744141,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05866196006536484,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.434992790222168,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.05532094091176987,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.450716018676758,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.051083534955978394,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.402632713317871,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.03815016150474548,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.433640956878662,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04203395918011665,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.415888786315918,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.043729111552238464,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.423260688781738,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.039426498115062714,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.428949356079102,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.03837858885526657,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.434070587158203,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.038315922021865845,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.471246719360352,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.03963236138224602,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.4434990882873535,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04027022793889046,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.436992645263672,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.0409243181347847,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.426936149597168,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03404631093144417,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.43462610244751,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.034393735229969025,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.437409400939941,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03635180741548538,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.451286315917969,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03349733352661133,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.419209003448486,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.03238677605986595,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.456454277038574,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.034473925828933716,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.3988728523254395,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.03455805033445358,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.434884071350098,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.037486057728528976,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.466741561889648,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.03617125004529953,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.429305076599121,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.033646389842033386,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.407780170440674,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03401118144392967,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.42977237701416,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.036534134298563004,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.451094627380371,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.0384049266576767,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.414777755737305,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03460627421736717,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.420644760131836,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.033746346831321716,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.461175918579102,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.037153810262680054,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.444699287414551,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03283539041876793,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.431022644042969,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03248969465494156,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.447571277618408,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03327852487564087,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.42236328125,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.031114818528294563,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.418244361877441,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.03231207653880119,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.455004692077637,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.030702251940965652,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.483610153198242,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031027289107441902,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.4470109939575195,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.03129073604941368,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.452799320220947,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.03634301945567131,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4661664962768555,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03718913346529007,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.412006855010986,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.034018341451883316,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.436755180358887,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032502319663763046,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.452866554260254,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.032902952283620834,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.449224472045898,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.0340626984834671,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.445425510406494,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.033442508429288864,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.432218551635742,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.03489428386092186,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.4205827713012695,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.031868621706962585,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.414196968078613,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.032986633479595184,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.46116304397583,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.03348355367779732,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.424936771392822,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03174544870853424,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.434438705444336,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.03293656185269356,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.440502166748047,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.035124607384204865,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.440943717956543,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03438971936702728,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.431408882141113,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.03537093102931976,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.43867301940918,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.03267325833439827,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.462226867675781,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.030995232984423637,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.433740615844727,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.03384929150342941,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.508179664611816,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.032025814056396484,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.407378196716309,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03288578242063522,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.455531120300293,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.032061848789453506,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.452212333679199,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022320276126265526,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7567386627197266,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5878658294677734,
+ "eval_runtime": 59.5054,
+ "eval_samples_per_second": 41.038,
+ "eval_steps_per_second": 0.655,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.039114587008953094,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.347412586212158,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.05164389684796333,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.391279220581055,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055882278829813004,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.355602264404297,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.05625041201710701,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3934855461120605,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05861108750104904,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.364226341247559,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.052431970834732056,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.363334655761719,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.06147193908691406,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.384820938110352,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.06701470166444778,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.373311996459961,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.06714873015880585,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.32745361328125,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.06684907525777817,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.373660087585449,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.05738082900643349,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.362325668334961,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05306725949048996,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.351685047149658,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.05831029266119003,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.374813079833984,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05547972396016121,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.361842632293701,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.048057056963443756,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.383205413818359,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.04682513326406479,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.365846157073975,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.05001377314329147,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.362151145935059,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.05140263959765434,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.376064300537109,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.04640384018421173,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.363801002502441,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04175020009279251,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.3477983474731445,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.04134776070713997,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.337469100952148,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.042178574949502945,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.351490020751953,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.03989115729928017,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.40725040435791,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04086652770638466,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.341119766235352,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.03931201249361038,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.3786468505859375,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.03855905309319496,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.3701605796813965,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.03803952783346176,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.401841163635254,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.036908265203237534,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.410647392272949,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.0398186556994915,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.380431175231934,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03899035602807999,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.374293804168701,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.03946344554424286,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.396918773651123,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.03738604858517647,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.404016017913818,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.03579043596982956,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.396914958953857,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.037139154970645905,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.407739639282227,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.03581099584698677,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.411419868469238,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.03548453748226166,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.332690238952637,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.037561237812042236,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.381314277648926,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.035827480256557465,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.371536731719971,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03538314625620842,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.395467758178711,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.036672890186309814,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.420590400695801,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.038981080055236816,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.395296096801758,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.038234200328588486,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.415799140930176,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.038830097764730453,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.413394927978516,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03990919142961502,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.39515495300293,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.036424681544303894,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.39730978012085,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03488154709339142,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.403696060180664,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.0363851860165596,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.368661403656006,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.035461440682411194,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.396518707275391,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03647598251700401,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.440667152404785,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.03667188808321953,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.3781352043151855,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034516457468271255,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.414007186889648,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.03536032512784004,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.42198371887207,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.037454646080732346,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.409174919128418,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03632275015115738,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3547682762146,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.038922134786844254,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.39993143081665,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.03562692925333977,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.409564971923828,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.033203717321157455,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.421782493591309,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03830202296376228,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.430092811584473,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.03835884481668472,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.379505634307861,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.037440430372953415,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.403937339782715,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03732272610068321,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.440683364868164,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03325957432389259,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.3949294090271,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.035377148538827896,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.400777816772461,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.03725374862551689,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.422494411468506,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03601110726594925,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.4008941650390625,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03193027526140213,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.400361061096191,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.034382354468107224,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.408535003662109,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.03747732564806938,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.412189483642578,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03641800582408905,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.381453037261963,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.0315779373049736,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.4169769287109375,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03583509847521782,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.385746955871582,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02574268914759159,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7456769943237305,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.588273823261261,
+ "eval_runtime": 59.9356,
+ "eval_samples_per_second": 40.744,
+ "eval_steps_per_second": 0.651,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013961605584642,
+ "grad_norm": 0.037516895681619644,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 4.33476448059082,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027923211169284,
+ "grad_norm": 0.048450272530317307,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 4.296454429626465,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041884816753926,
+ "grad_norm": 0.05368480458855629,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 4.314815998077393,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055846422338568,
+ "grad_norm": 0.05520009621977806,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 4.330798625946045,
+ "step": 1084
+ },
+ {
+ "epoch": 15.06980802792321,
+ "grad_norm": 0.05928800627589226,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 4.317663192749023,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083769633507853,
+ "grad_norm": 0.06315027177333832,
+ "learning_rate": 0.000491285979372924,
+ "loss": 4.329221725463867,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097731239092496,
+ "grad_norm": 0.06342213600873947,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 4.32546329498291,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111692844677139,
+ "grad_norm": 0.0667002946138382,
+ "learning_rate": 0.000490873338088198,
+ "loss": 4.309828758239746,
+ "step": 1088
+ },
+ {
+ "epoch": 15.12565445026178,
+ "grad_norm": 0.05264135077595711,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 4.307073593139648,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139616055846423,
+ "grad_norm": 0.051895879209041595,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 4.333526611328125,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153577661431065,
+ "grad_norm": 0.05508396774530411,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 4.330171585083008,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167539267015707,
+ "grad_norm": 0.057272691279649734,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 4.329268455505371,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181500872600349,
+ "grad_norm": 0.058956146240234375,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 4.341983795166016,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195462478184991,
+ "grad_norm": 0.0557059645652771,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 4.348274230957031,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209424083769633,
+ "grad_norm": 0.05661974847316742,
+ "learning_rate": 0.000489424334303338,
+ "loss": 4.347631454467773,
+ "step": 1095
+ },
+ {
+ "epoch": 15.223385689354275,
+ "grad_norm": 0.05338657647371292,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 4.334639072418213,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237347294938917,
+ "grad_norm": 0.05091148987412453,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 4.355465412139893,
+ "step": 1097
+ },
+ {
+ "epoch": 15.25130890052356,
+ "grad_norm": 0.04996965080499649,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 4.346457004547119,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265270506108202,
+ "grad_norm": 0.05573750659823418,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 4.346520900726318,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279232111692846,
+ "grad_norm": 0.05776917189359665,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 4.359007835388184,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293193717277488,
+ "grad_norm": 0.050820041447877884,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 4.331710338592529,
+ "step": 1101
+ },
+ {
+ "epoch": 15.30715532286213,
+ "grad_norm": 0.04441820830106735,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 4.3126420974731445,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321116928446772,
+ "grad_norm": 0.04582618921995163,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 4.312925338745117,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335078534031414,
+ "grad_norm": 0.043592359870672226,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 4.345674514770508,
+ "step": 1104
+ },
+ {
+ "epoch": 15.349040139616056,
+ "grad_norm": 0.04069037735462189,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 4.334713935852051,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363001745200698,
+ "grad_norm": 0.04083041846752167,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 4.358171463012695,
+ "step": 1106
+ },
+ {
+ "epoch": 15.37696335078534,
+ "grad_norm": 0.04119594022631645,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 4.317891597747803,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390924956369982,
+ "grad_norm": 0.042060643434524536,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 4.355810165405273,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404886561954624,
+ "grad_norm": 0.04721749201416969,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 4.353546619415283,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418848167539267,
+ "grad_norm": 0.04657730832695961,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 4.341683387756348,
+ "step": 1110
+ },
+ {
+ "epoch": 15.432809773123909,
+ "grad_norm": 0.04096715524792671,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 4.357192039489746,
+ "step": 1111
+ },
+ {
+ "epoch": 15.44677137870855,
+ "grad_norm": 0.03904155269265175,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 4.354775428771973,
+ "step": 1112
+ },
+ {
+ "epoch": 15.460732984293193,
+ "grad_norm": 0.03828521445393562,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 4.356499671936035,
+ "step": 1113
+ },
+ {
+ "epoch": 15.474694589877837,
+ "grad_norm": 0.03688685595989227,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 4.339485168457031,
+ "step": 1114
+ },
+ {
+ "epoch": 15.488656195462479,
+ "grad_norm": 0.03945618495345116,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 4.363500118255615,
+ "step": 1115
+ },
+ {
+ "epoch": 15.502617801047121,
+ "grad_norm": 0.03821462020277977,
+ "learning_rate": 0.000485033394934934,
+ "loss": 4.353529930114746,
+ "step": 1116
+ },
+ {
+ "epoch": 15.516579406631763,
+ "grad_norm": 0.036265499889850616,
+ "learning_rate": 0.000484822676912974,
+ "loss": 4.337663173675537,
+ "step": 1117
+ },
+ {
+ "epoch": 15.530541012216405,
+ "grad_norm": 0.03803930804133415,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 4.34660530090332,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544502617801047,
+ "grad_norm": 0.038425102829933167,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 4.373246192932129,
+ "step": 1119
+ },
+ {
+ "epoch": 15.55846422338569,
+ "grad_norm": 0.036852262914180756,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 4.331226825714111,
+ "step": 1120
+ },
+ {
+ "epoch": 15.572425828970331,
+ "grad_norm": 0.03700809180736542,
+ "learning_rate": 0.000483978341612154,
+ "loss": 4.375818729400635,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586387434554974,
+ "grad_norm": 0.03628664091229439,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 4.32705020904541,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600349040139616,
+ "grad_norm": 0.036831192672252655,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 4.337127208709717,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614310645724258,
+ "grad_norm": 0.03736541047692299,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 4.36502742767334,
+ "step": 1124
+ },
+ {
+ "epoch": 15.6282722513089,
+ "grad_norm": 0.03658579662442207,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 4.359844207763672,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642233856893542,
+ "grad_norm": 0.03688633441925049,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 4.3355793952941895,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656195462478184,
+ "grad_norm": 0.03477008268237114,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 4.333405494689941,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670157068062828,
+ "grad_norm": 0.03655015677213669,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 4.338570594787598,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68411867364747,
+ "grad_norm": 0.03664765506982803,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 4.352496147155762,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698080279232112,
+ "grad_norm": 0.037019114941358566,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 4.335735321044922,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712041884816754,
+ "grad_norm": 0.03949085995554924,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 4.342585563659668,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726003490401396,
+ "grad_norm": 0.04084350913763046,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 4.329622268676758,
+ "step": 1132
+ },
+ {
+ "epoch": 15.739965095986038,
+ "grad_norm": 0.038439832627773285,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 4.360318660736084,
+ "step": 1133
+ },
+ {
+ "epoch": 15.75392670157068,
+ "grad_norm": 0.03932987526059151,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 4.343446254730225,
+ "step": 1134
+ },
+ {
+ "epoch": 15.767888307155323,
+ "grad_norm": 0.041595373302698135,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 4.351495742797852,
+ "step": 1135
+ },
+ {
+ "epoch": 15.781849912739965,
+ "grad_norm": 0.03917490318417549,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 4.380537033081055,
+ "step": 1136
+ },
+ {
+ "epoch": 15.795811518324607,
+ "grad_norm": 0.038597021251916885,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 4.3527116775512695,
+ "step": 1137
+ },
+ {
+ "epoch": 15.809773123909249,
+ "grad_norm": 0.03651711344718933,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 4.369256496429443,
+ "step": 1138
+ },
+ {
+ "epoch": 15.823734729493891,
+ "grad_norm": 0.035118963569402695,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 4.350616455078125,
+ "step": 1139
+ },
+ {
+ "epoch": 15.837696335078533,
+ "grad_norm": 0.034945692867040634,
+ "learning_rate": 0.000479936075927473,
+ "loss": 4.365446090698242,
+ "step": 1140
+ },
+ {
+ "epoch": 15.851657940663177,
+ "grad_norm": 0.03519820421934128,
+ "learning_rate": 0.000479721889242305,
+ "loss": 4.357390403747559,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86561954624782,
+ "grad_norm": 0.03407098725438118,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 4.350573539733887,
+ "step": 1142
+ },
+ {
+ "epoch": 15.879581151832461,
+ "grad_norm": 0.036511123180389404,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 4.328500270843506,
+ "step": 1143
+ },
+ {
+ "epoch": 15.893542757417103,
+ "grad_norm": 0.039194293320178986,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 4.346471786499023,
+ "step": 1144
+ },
+ {
+ "epoch": 15.907504363001745,
+ "grad_norm": 0.03777911514043808,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 4.358243942260742,
+ "step": 1145
+ },
+ {
+ "epoch": 15.921465968586388,
+ "grad_norm": 0.0354655385017395,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 4.368768692016602,
+ "step": 1146
+ },
+ {
+ "epoch": 15.93542757417103,
+ "grad_norm": 0.03710418567061424,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 4.3482184410095215,
+ "step": 1147
+ },
+ {
+ "epoch": 15.949389179755672,
+ "grad_norm": 0.03545419126749039,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 4.360692501068115,
+ "step": 1148
+ },
+ {
+ "epoch": 15.963350785340314,
+ "grad_norm": 0.035963475704193115,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 4.3592119216918945,
+ "step": 1149
+ },
+ {
+ "epoch": 15.977312390924956,
+ "grad_norm": 0.03425297513604164,
+ "learning_rate": 0.000477787822648007,
+ "loss": 4.329504013061523,
+ "step": 1150
+ },
+ {
+ "epoch": 15.991273996509598,
+ "grad_norm": 0.03453255817294121,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 4.298142433166504,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.025766054168343544,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 2.7026405334472656,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.5888971090316772,
+ "eval_runtime": 58.9721,
+ "eval_samples_per_second": 41.409,
+ "eval_steps_per_second": 0.661,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013961605584644,
+ "grad_norm": 0.04088004305958748,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 4.268601417541504,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027923211169284,
+ "grad_norm": 0.05194641277194023,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 4.267483711242676,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041884816753928,
+ "grad_norm": 0.05488700792193413,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 4.2527570724487305,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05584642233857,
+ "grad_norm": 0.059292279183864594,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 4.283864974975586,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069808027923212,
+ "grad_norm": 0.0567227303981781,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 4.291148662567139,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083769633507853,
+ "grad_norm": 0.05582835525274277,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 4.250524520874023,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097731239092496,
+ "grad_norm": 0.055709999054670334,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 4.259716033935547,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111692844677137,
+ "grad_norm": 0.053324732929468155,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 4.247231483459473,
+ "step": 1160
+ },
+ {
+ "epoch": 16.12565445026178,
+ "grad_norm": 0.05367910861968994,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 4.263905048370361,
+ "step": 1161
+ },
+ {
+ "epoch": 16.13961605584642,
+ "grad_norm": 0.057094886898994446,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 4.266510009765625,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153577661431065,
+ "grad_norm": 0.05921962857246399,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 4.278281211853027,
+ "step": 1163
+ },
+ {
+ "epoch": 16.167539267015705,
+ "grad_norm": 0.060246821492910385,
+ "learning_rate": 0.000474756648870413,
+ "loss": 4.2848052978515625,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18150087260035,
+ "grad_norm": 0.05686050280928612,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 4.286434173583984,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195462478184993,
+ "grad_norm": 0.05207265168428421,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 4.246121406555176,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209424083769633,
+ "grad_norm": 0.0477963425219059,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 4.287411689758301,
+ "step": 1167
+ },
+ {
+ "epoch": 16.223385689354277,
+ "grad_norm": 0.05586806684732437,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 4.2721333503723145,
+ "step": 1168
+ },
+ {
+ "epoch": 16.237347294938917,
+ "grad_norm": 0.05427060276269913,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 4.2751641273498535,
+ "step": 1169
+ },
+ {
+ "epoch": 16.25130890052356,
+ "grad_norm": 0.0518609918653965,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 4.287444114685059,
+ "step": 1170
+ },
+ {
+ "epoch": 16.2652705061082,
+ "grad_norm": 0.053606316447257996,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 4.287125587463379,
+ "step": 1171
+ },
+ {
+ "epoch": 16.279232111692846,
+ "grad_norm": 0.05118813365697861,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 4.291893005371094,
+ "step": 1172
+ },
+ {
+ "epoch": 16.293193717277486,
+ "grad_norm": 0.04774491861462593,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 4.325250148773193,
+ "step": 1173
+ },
+ {
+ "epoch": 16.30715532286213,
+ "grad_norm": 0.04634870961308479,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 4.311764717102051,
+ "step": 1174
+ },
+ {
+ "epoch": 16.32111692844677,
+ "grad_norm": 0.04411927983164787,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 4.277068138122559,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335078534031414,
+ "grad_norm": 0.043819356709718704,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 4.309638977050781,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349040139616054,
+ "grad_norm": 0.0418904684484005,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 4.306234836578369,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363001745200698,
+ "grad_norm": 0.04327298328280449,
+ "learning_rate": 0.000471698352726896,
+ "loss": 4.296874046325684,
+ "step": 1178
+ },
+ {
+ "epoch": 16.376963350785342,
+ "grad_norm": 0.045296184718608856,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 4.288957595825195,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390924956369982,
+ "grad_norm": 0.04447923228144646,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 4.313834190368652,
+ "step": 1180
+ },
+ {
+ "epoch": 16.404886561954626,
+ "grad_norm": 0.04548390209674835,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 4.284562587738037,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418848167539267,
+ "grad_norm": 0.04567817226052284,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 4.295994758605957,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43280977312391,
+ "grad_norm": 0.04277439042925835,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 4.262106895446777,
+ "step": 1183
+ },
+ {
+ "epoch": 16.44677137870855,
+ "grad_norm": 0.043597545474767685,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 4.294536590576172,
+ "step": 1184
+ },
+ {
+ "epoch": 16.460732984293195,
+ "grad_norm": 0.04424033313989639,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 4.297497749328613,
+ "step": 1185
+ },
+ {
+ "epoch": 16.474694589877835,
+ "grad_norm": 0.04354134574532509,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 4.321508884429932,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48865619546248,
+ "grad_norm": 0.04241178184747696,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 4.322726249694824,
+ "step": 1187
+ },
+ {
+ "epoch": 16.50261780104712,
+ "grad_norm": 0.039317622780799866,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 4.311155319213867,
+ "step": 1188
+ },
+ {
+ "epoch": 16.516579406631763,
+ "grad_norm": 0.04126536473631859,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 4.306636810302734,
+ "step": 1189
+ },
+ {
+ "epoch": 16.530541012216403,
+ "grad_norm": 0.040116794407367706,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 4.314671516418457,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544502617801047,
+ "grad_norm": 0.04045999422669411,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 4.301012992858887,
+ "step": 1191
+ },
+ {
+ "epoch": 16.55846422338569,
+ "grad_norm": 0.04013422504067421,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 4.29913330078125,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57242582897033,
+ "grad_norm": 0.03878502920269966,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 4.351600646972656,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586387434554975,
+ "grad_norm": 0.038935888558626175,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 4.27667236328125,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600349040139616,
+ "grad_norm": 0.04055892676115036,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 4.29019021987915,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61431064572426,
+ "grad_norm": 0.03959144279360771,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 4.327725887298584,
+ "step": 1196
+ },
+ {
+ "epoch": 16.6282722513089,
+ "grad_norm": 0.039878539741039276,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 4.297952651977539,
+ "step": 1197
+ },
+ {
+ "epoch": 16.642233856893544,
+ "grad_norm": 0.041209928691387177,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 4.313030242919922,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656195462478184,
+ "grad_norm": 0.04346393793821335,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 4.274901390075684,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670157068062828,
+ "grad_norm": 0.04090269282460213,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 4.330657958984375,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68411867364747,
+ "grad_norm": 0.04004506766796112,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 4.323106288909912,
+ "step": 1201
+ },
+ {
+ "epoch": 16.698080279232112,
+ "grad_norm": 0.041541460901498795,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 4.280947685241699,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712041884816752,
+ "grad_norm": 0.04033653065562248,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 4.341434955596924,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726003490401396,
+ "grad_norm": 0.04104280099272728,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 4.2931671142578125,
+ "step": 1204
+ },
+ {
+ "epoch": 16.739965095986037,
+ "grad_norm": 0.039398156106472015,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 4.316840648651123,
+ "step": 1205
+ },
+ {
+ "epoch": 16.75392670157068,
+ "grad_norm": 0.03980560600757599,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 4.325345039367676,
+ "step": 1206
+ },
+ {
+ "epoch": 16.767888307155324,
+ "grad_norm": 0.040656186640262604,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 4.32675838470459,
+ "step": 1207
+ },
+ {
+ "epoch": 16.781849912739965,
+ "grad_norm": 0.03775986284017563,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 4.314423561096191,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79581151832461,
+ "grad_norm": 0.04043284431099892,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 4.344202041625977,
+ "step": 1209
+ },
+ {
+ "epoch": 16.80977312390925,
+ "grad_norm": 0.04074503108859062,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 4.350123405456543,
+ "step": 1210
+ },
+ {
+ "epoch": 16.823734729493893,
+ "grad_norm": 0.04117420315742493,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 4.350917816162109,
+ "step": 1211
+ },
+ {
+ "epoch": 16.837696335078533,
+ "grad_norm": 0.03799610957503319,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 4.289987564086914,
+ "step": 1212
+ },
+ {
+ "epoch": 16.851657940663177,
+ "grad_norm": 0.04071265831589699,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 4.306744575500488,
+ "step": 1213
+ },
+ {
+ "epoch": 16.865619546247817,
+ "grad_norm": 0.04109715297818184,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 4.318753242492676,
+ "step": 1214
+ },
+ {
+ "epoch": 16.87958115183246,
+ "grad_norm": 0.03910145163536072,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 4.281325340270996,
+ "step": 1215
+ },
+ {
+ "epoch": 16.8935427574171,
+ "grad_norm": 0.03752733767032623,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 4.3231048583984375,
+ "step": 1216
+ },
+ {
+ "epoch": 16.907504363001745,
+ "grad_norm": 0.03752103075385094,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 4.351336479187012,
+ "step": 1217
+ },
+ {
+ "epoch": 16.921465968586386,
+ "grad_norm": 0.03675422817468643,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 4.282256126403809,
+ "step": 1218
+ },
+ {
+ "epoch": 16.93542757417103,
+ "grad_norm": 0.03850258141756058,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 4.322597503662109,
+ "step": 1219
+ },
+ {
+ "epoch": 16.949389179755673,
+ "grad_norm": 0.03692392632365227,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 4.320333957672119,
+ "step": 1220
+ },
+ {
+ "epoch": 16.963350785340314,
+ "grad_norm": 0.035366240888834,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 4.31341552734375,
+ "step": 1221
+ },
+ {
+ "epoch": 16.977312390924958,
+ "grad_norm": 0.037120968103408813,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 4.321141242980957,
+ "step": 1222
+ },
+ {
+ "epoch": 16.991273996509598,
+ "grad_norm": 0.03721754252910614,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 4.305757522583008,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.026698043569922447,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 2.716372489929199,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.5894472599029541,
+ "eval_runtime": 58.927,
+ "eval_samples_per_second": 41.441,
+ "eval_steps_per_second": 0.662,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013961605584644,
+ "grad_norm": 0.044532276690006256,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 4.243244171142578,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027923211169284,
+ "grad_norm": 0.050529688596725464,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 4.221382141113281,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041884816753928,
+ "grad_norm": 0.05690870061516762,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 4.240285873413086,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05584642233857,
+ "grad_norm": 0.054803814738988876,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 4.213047027587891,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069808027923212,
+ "grad_norm": 0.051825862377882004,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 4.226880073547363,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083769633507853,
+ "grad_norm": 0.049802035093307495,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 4.25303840637207,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097731239092496,
+ "grad_norm": 0.05290328338742256,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 4.199422836303711,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111692844677137,
+ "grad_norm": 0.0531814880669117,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 4.201648712158203,
+ "step": 1232
+ },
+ {
+ "epoch": 17.12565445026178,
+ "grad_norm": 0.048972588032484055,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 4.207578182220459,
+ "step": 1233
+ },
+ {
+ "epoch": 17.13961605584642,
+ "grad_norm": 0.051345713436603546,
+ "learning_rate": 0.000459203665939781,
+ "loss": 4.206205368041992,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153577661431065,
+ "grad_norm": 0.059012677520513535,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 4.228360176086426,
+ "step": 1235
+ },
+ {
+ "epoch": 17.167539267015705,
+ "grad_norm": 0.05984492972493172,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 4.234362602233887,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18150087260035,
+ "grad_norm": 0.060886185616254807,
+ "learning_rate": 0.000458522886595554,
+ "loss": 4.253525257110596,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195462478184993,
+ "grad_norm": 0.06476911902427673,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 4.249889373779297,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209424083769633,
+ "grad_norm": 0.059773124754428864,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 4.225249290466309,
+ "step": 1239
+ },
+ {
+ "epoch": 17.223385689354277,
+ "grad_norm": 0.06063077598810196,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 4.25465202331543,
+ "step": 1240
+ },
+ {
+ "epoch": 17.237347294938917,
+ "grad_norm": 0.059922896325588226,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 4.252101898193359,
+ "step": 1241
+ },
+ {
+ "epoch": 17.25130890052356,
+ "grad_norm": 0.05849695950746536,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 4.268218040466309,
+ "step": 1242
+ },
+ {
+ "epoch": 17.2652705061082,
+ "grad_norm": 0.05769657716155052,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 4.2395782470703125,
+ "step": 1243
+ },
+ {
+ "epoch": 17.279232111692846,
+ "grad_norm": 0.058189719915390015,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 4.257606506347656,
+ "step": 1244
+ },
+ {
+ "epoch": 17.293193717277486,
+ "grad_norm": 0.059679049998521805,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 4.278588771820068,
+ "step": 1245
+ },
+ {
+ "epoch": 17.30715532286213,
+ "grad_norm": 0.0579286590218544,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 4.25511360168457,
+ "step": 1246
+ },
+ {
+ "epoch": 17.32111692844677,
+ "grad_norm": 0.05269965901970863,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 4.241326332092285,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335078534031414,
+ "grad_norm": 0.04828358441591263,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 4.243182182312012,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349040139616054,
+ "grad_norm": 0.050172045826911926,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 4.295144081115723,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363001745200698,
+ "grad_norm": 0.047297943383455276,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 4.250682830810547,
+ "step": 1250
+ },
+ {
+ "epoch": 17.376963350785342,
+ "grad_norm": 0.04574039950966835,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 4.269444465637207,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390924956369982,
+ "grad_norm": 0.042310427874326706,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 4.255361557006836,
+ "step": 1252
+ },
+ {
+ "epoch": 17.404886561954626,
+ "grad_norm": 0.04268861189484596,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 4.252292156219482,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418848167539267,
+ "grad_norm": 0.0424286350607872,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 4.24439811706543,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43280977312391,
+ "grad_norm": 0.042049869894981384,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 4.236266136169434,
+ "step": 1255
+ },
+ {
+ "epoch": 17.44677137870855,
+ "grad_norm": 0.04185241088271141,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 4.268826484680176,
+ "step": 1256
+ },
+ {
+ "epoch": 17.460732984293195,
+ "grad_norm": 0.04423609375953674,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 4.240240573883057,
+ "step": 1257
+ },
+ {
+ "epoch": 17.474694589877835,
+ "grad_norm": 0.041972566395998,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 4.269453048706055,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48865619546248,
+ "grad_norm": 0.04493343085050583,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 4.270544052124023,
+ "step": 1259
+ },
+ {
+ "epoch": 17.50261780104712,
+ "grad_norm": 0.04502921551465988,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 4.271880149841309,
+ "step": 1260
+ },
+ {
+ "epoch": 17.516579406631763,
+ "grad_norm": 0.044518254697322845,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 4.249786376953125,
+ "step": 1261
+ },
+ {
+ "epoch": 17.530541012216403,
+ "grad_norm": 0.044482357800006866,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 4.300779819488525,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544502617801047,
+ "grad_norm": 0.0424671471118927,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 4.248597621917725,
+ "step": 1263
+ },
+ {
+ "epoch": 17.55846422338569,
+ "grad_norm": 0.04136480763554573,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 4.276763916015625,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57242582897033,
+ "grad_norm": 0.04094843938946724,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 4.263628005981445,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586387434554975,
+ "grad_norm": 0.04415388032793999,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 4.257946014404297,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600349040139616,
+ "grad_norm": 0.04177384823560715,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 4.289669036865234,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61431064572426,
+ "grad_norm": 0.040863484144210815,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 4.320903778076172,
+ "step": 1268
+ },
+ {
+ "epoch": 17.6282722513089,
+ "grad_norm": 0.04025879129767418,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 4.268717288970947,
+ "step": 1269
+ },
+ {
+ "epoch": 17.642233856893544,
+ "grad_norm": 0.04161325842142105,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 4.275268077850342,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656195462478184,
+ "grad_norm": 0.04260355979204178,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 4.263361930847168,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670157068062828,
+ "grad_norm": 0.043366871774196625,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 4.254581928253174,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68411867364747,
+ "grad_norm": 0.04299113526940346,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 4.268922328948975,
+ "step": 1273
+ },
+ {
+ "epoch": 17.698080279232112,
+ "grad_norm": 0.041865330189466476,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 4.290097236633301,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712041884816752,
+ "grad_norm": 0.04515853524208069,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 4.261376857757568,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726003490401396,
+ "grad_norm": 0.046030085533857346,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 4.290409564971924,
+ "step": 1276
+ },
+ {
+ "epoch": 17.739965095986037,
+ "grad_norm": 0.045034267008304596,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 4.28201961517334,
+ "step": 1277
+ },
+ {
+ "epoch": 17.75392670157068,
+ "grad_norm": 0.04293077066540718,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 4.282614707946777,
+ "step": 1278
+ },
+ {
+ "epoch": 17.767888307155324,
+ "grad_norm": 0.043866172432899475,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 4.2777557373046875,
+ "step": 1279
+ },
+ {
+ "epoch": 17.781849912739965,
+ "grad_norm": 0.04693951830267906,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 4.272597312927246,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79581151832461,
+ "grad_norm": 0.04072774201631546,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 4.284282684326172,
+ "step": 1281
+ },
+ {
+ "epoch": 17.80977312390925,
+ "grad_norm": 0.04241872951388359,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 4.254406452178955,
+ "step": 1282
+ },
+ {
+ "epoch": 17.823734729493893,
+ "grad_norm": 0.04614942893385887,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 4.297187805175781,
+ "step": 1283
+ },
+ {
+ "epoch": 17.837696335078533,
+ "grad_norm": 0.04326155036687851,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 4.292263984680176,
+ "step": 1284
+ },
+ {
+ "epoch": 17.851657940663177,
+ "grad_norm": 0.04053817689418793,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 4.2925920486450195,
+ "step": 1285
+ },
+ {
+ "epoch": 17.865619546247817,
+ "grad_norm": 0.040867388248443604,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 4.276965141296387,
+ "step": 1286
+ },
+ {
+ "epoch": 17.87958115183246,
+ "grad_norm": 0.041000183671712875,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 4.257638454437256,
+ "step": 1287
+ },
+ {
+ "epoch": 17.8935427574171,
+ "grad_norm": 0.04014866054058075,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 4.2734479904174805,
+ "step": 1288
+ },
+ {
+ "epoch": 17.907504363001745,
+ "grad_norm": 0.041237372905015945,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 4.263461112976074,
+ "step": 1289
+ },
+ {
+ "epoch": 17.921465968586386,
+ "grad_norm": 0.039109110832214355,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 4.295286655426025,
+ "step": 1290
+ },
+ {
+ "epoch": 17.93542757417103,
+ "grad_norm": 0.03911379352211952,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 4.254448413848877,
+ "step": 1291
+ },
+ {
+ "epoch": 17.949389179755673,
+ "grad_norm": 0.03997721150517464,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 4.284714698791504,
+ "step": 1292
+ },
+ {
+ "epoch": 17.963350785340314,
+ "grad_norm": 0.04011402651667595,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 4.278304100036621,
+ "step": 1293
+ },
+ {
+ "epoch": 17.977312390924958,
+ "grad_norm": 0.04040595516562462,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 4.288183212280273,
+ "step": 1294
+ },
+ {
+ "epoch": 17.991273996509598,
+ "grad_norm": 0.038857635110616684,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 4.274664402008057,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.028110472485423088,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 2.6625723838806152,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.5912551283836365,
+ "eval_runtime": 60.273,
+ "eval_samples_per_second": 40.516,
+ "eval_steps_per_second": 0.647,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013961605584644,
+ "grad_norm": 0.041634928435087204,
+ "learning_rate": 0.000444677103403854,
+ "loss": 4.181324005126953,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027923211169284,
+ "grad_norm": 0.051596369594335556,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 4.204722881317139,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041884816753928,
+ "grad_norm": 0.058580417186021805,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 4.207996845245361,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05584642233857,
+ "grad_norm": 0.05986522138118744,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 4.184011459350586,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069808027923212,
+ "grad_norm": 0.051045969128608704,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 4.148118019104004,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083769633507853,
+ "grad_norm": 0.045477766543626785,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 4.207469940185547,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097731239092496,
+ "grad_norm": 0.0488731674849987,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 4.201406478881836,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111692844677137,
+ "grad_norm": 0.05013517662882805,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 4.192239761352539,
+ "step": 1304
+ },
+ {
+ "epoch": 18.12565445026178,
+ "grad_norm": 0.05091651529073715,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 4.188252925872803,
+ "step": 1305
+ },
+ {
+ "epoch": 18.13961605584642,
+ "grad_norm": 0.051825884729623795,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 4.157929420471191,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153577661431065,
+ "grad_norm": 0.049625713378190994,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 4.208776473999023,
+ "step": 1307
+ },
+ {
+ "epoch": 18.167539267015705,
+ "grad_norm": 0.04816461354494095,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 4.180496692657471,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18150087260035,
+ "grad_norm": 0.04720531404018402,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 4.178536415100098,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195462478184993,
+ "grad_norm": 0.0475568026304245,
+ "learning_rate": 0.000441621628895411,
+ "loss": 4.203497886657715,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209424083769633,
+ "grad_norm": 0.04613947868347168,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 4.226914405822754,
+ "step": 1311
+ },
+ {
+ "epoch": 18.223385689354277,
+ "grad_norm": 0.04585929960012436,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 4.175119400024414,
+ "step": 1312
+ },
+ {
+ "epoch": 18.237347294938917,
+ "grad_norm": 0.045777060091495514,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 4.181760311126709,
+ "step": 1313
+ },
+ {
+ "epoch": 18.25130890052356,
+ "grad_norm": 0.04583602771162987,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 4.179868698120117,
+ "step": 1314
+ },
+ {
+ "epoch": 18.2652705061082,
+ "grad_norm": 0.04706503078341484,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 4.203325271606445,
+ "step": 1315
+ },
+ {
+ "epoch": 18.279232111692846,
+ "grad_norm": 0.045122336596250534,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 4.187152862548828,
+ "step": 1316
+ },
+ {
+ "epoch": 18.293193717277486,
+ "grad_norm": 0.04495825991034508,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 4.2080535888671875,
+ "step": 1317
+ },
+ {
+ "epoch": 18.30715532286213,
+ "grad_norm": 0.04767543077468872,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 4.207225322723389,
+ "step": 1318
+ },
+ {
+ "epoch": 18.32111692844677,
+ "grad_norm": 0.053396135568618774,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 4.227478981018066,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335078534031414,
+ "grad_norm": 0.0512089841067791,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 4.1904754638671875,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349040139616054,
+ "grad_norm": 0.04967977851629257,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 4.2079925537109375,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363001745200698,
+ "grad_norm": 0.04800970479846001,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 4.206037998199463,
+ "step": 1322
+ },
+ {
+ "epoch": 18.376963350785342,
+ "grad_norm": 0.04993556812405586,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 4.210054397583008,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390924956369982,
+ "grad_norm": 0.0480492040514946,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 4.229897499084473,
+ "step": 1324
+ },
+ {
+ "epoch": 18.404886561954626,
+ "grad_norm": 0.048699282109737396,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 4.17500114440918,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418848167539267,
+ "grad_norm": 0.04681490734219551,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 4.196526527404785,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43280977312391,
+ "grad_norm": 0.04674449935555458,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 4.246790409088135,
+ "step": 1327
+ },
+ {
+ "epoch": 18.44677137870855,
+ "grad_norm": 0.04650607705116272,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 4.225878715515137,
+ "step": 1328
+ },
+ {
+ "epoch": 18.460732984293195,
+ "grad_norm": 0.047200269997119904,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 4.20685338973999,
+ "step": 1329
+ },
+ {
+ "epoch": 18.474694589877835,
+ "grad_norm": 0.04322655498981476,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 4.212920665740967,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48865619546248,
+ "grad_norm": 0.04489128664135933,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 4.266653060913086,
+ "step": 1331
+ },
+ {
+ "epoch": 18.50261780104712,
+ "grad_norm": 0.048838935792446136,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 4.2364501953125,
+ "step": 1332
+ },
+ {
+ "epoch": 18.516579406631763,
+ "grad_norm": 0.04893286153674126,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 4.202394485473633,
+ "step": 1333
+ },
+ {
+ "epoch": 18.530541012216403,
+ "grad_norm": 0.046900101006031036,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 4.209832191467285,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544502617801047,
+ "grad_norm": 0.044056762009859085,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 4.229903221130371,
+ "step": 1335
+ },
+ {
+ "epoch": 18.55846422338569,
+ "grad_norm": 0.04681328684091568,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 4.22783899307251,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57242582897033,
+ "grad_norm": 0.046087440103292465,
+ "learning_rate": 0.000435215814386134,
+ "loss": 4.2078986167907715,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586387434554975,
+ "grad_norm": 0.04363090172410011,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 4.218686103820801,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600349040139616,
+ "grad_norm": 0.046255532652139664,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 4.2475128173828125,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61431064572426,
+ "grad_norm": 0.045545510947704315,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 4.233272552490234,
+ "step": 1340
+ },
+ {
+ "epoch": 18.6282722513089,
+ "grad_norm": 0.045259326696395874,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 4.252043724060059,
+ "step": 1341
+ },
+ {
+ "epoch": 18.642233856893544,
+ "grad_norm": 0.042144861072301865,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 4.247762203216553,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656195462478184,
+ "grad_norm": 0.04298558086156845,
+ "learning_rate": 0.000433781621332752,
+ "loss": 4.222329616546631,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670157068062828,
+ "grad_norm": 0.044070884585380554,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 4.222488880157471,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68411867364747,
+ "grad_norm": 0.044758375734090805,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 4.238160133361816,
+ "step": 1345
+ },
+ {
+ "epoch": 18.698080279232112,
+ "grad_norm": 0.04324857518076897,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 4.220003128051758,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712041884816752,
+ "grad_norm": 0.04349793121218681,
+ "learning_rate": 0.000432823382695327,
+ "loss": 4.251608848571777,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726003490401396,
+ "grad_norm": 0.04483242705464363,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 4.243319511413574,
+ "step": 1348
+ },
+ {
+ "epoch": 18.739965095986037,
+ "grad_norm": 0.04391994699835777,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 4.237479209899902,
+ "step": 1349
+ },
+ {
+ "epoch": 18.75392670157068,
+ "grad_norm": 0.042584922164678574,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 4.2309980392456055,
+ "step": 1350
+ },
+ {
+ "epoch": 18.767888307155324,
+ "grad_norm": 0.04363991320133209,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 4.216980934143066,
+ "step": 1351
+ },
+ {
+ "epoch": 18.781849912739965,
+ "grad_norm": 0.04662707448005676,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 4.257450103759766,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79581151832461,
+ "grad_norm": 0.0462074875831604,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 4.222329139709473,
+ "step": 1353
+ },
+ {
+ "epoch": 18.80977312390925,
+ "grad_norm": 0.043543554842472076,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 4.231960773468018,
+ "step": 1354
+ },
+ {
+ "epoch": 18.823734729493893,
+ "grad_norm": 0.04444567486643791,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 4.225714683532715,
+ "step": 1355
+ },
+ {
+ "epoch": 18.837696335078533,
+ "grad_norm": 0.04507977142930031,
+ "learning_rate": 0.000430661245733797,
+ "loss": 4.234722137451172,
+ "step": 1356
+ },
+ {
+ "epoch": 18.851657940663177,
+ "grad_norm": 0.043786536902189255,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 4.274998664855957,
+ "step": 1357
+ },
+ {
+ "epoch": 18.865619546247817,
+ "grad_norm": 0.04188325256109238,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 4.260570526123047,
+ "step": 1358
+ },
+ {
+ "epoch": 18.87958115183246,
+ "grad_norm": 0.0439927875995636,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 4.250492572784424,
+ "step": 1359
+ },
+ {
+ "epoch": 18.8935427574171,
+ "grad_norm": 0.04520147293806076,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 4.250911712646484,
+ "step": 1360
+ },
+ {
+ "epoch": 18.907504363001745,
+ "grad_norm": 0.04442249983549118,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 4.255832672119141,
+ "step": 1361
+ },
+ {
+ "epoch": 18.921465968586386,
+ "grad_norm": 0.044756144285202026,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 4.226641654968262,
+ "step": 1362
+ },
+ {
+ "epoch": 18.93542757417103,
+ "grad_norm": 0.04647127911448479,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 4.2110915184021,
+ "step": 1363
+ },
+ {
+ "epoch": 18.949389179755673,
+ "grad_norm": 0.045986004173755646,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 4.243263244628906,
+ "step": 1364
+ },
+ {
+ "epoch": 18.963350785340314,
+ "grad_norm": 0.04533308371901512,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 4.253750801086426,
+ "step": 1365
+ },
+ {
+ "epoch": 18.977312390924958,
+ "grad_norm": 0.04307471215724945,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 4.233024597167969,
+ "step": 1366
+ },
+ {
+ "epoch": 18.991273996509598,
+ "grad_norm": 0.0416078194975853,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 4.264965534210205,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.030926328152418137,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 2.651331901550293,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.592528223991394,
+ "eval_runtime": 59.67,
+ "eval_samples_per_second": 40.925,
+ "eval_steps_per_second": 0.654,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013961605584644,
+ "grad_norm": 0.04524356126785278,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 4.153796672821045,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027923211169284,
+ "grad_norm": 0.05105067789554596,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 4.154603004455566,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041884816753928,
+ "grad_norm": 0.052823506295681,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 4.1599273681640625,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05584642233857,
+ "grad_norm": 0.05688423290848732,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 4.124622344970703,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069808027923212,
+ "grad_norm": 0.053178343921899796,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 4.138644218444824,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083769633507853,
+ "grad_norm": 0.05473732203245163,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 4.138207912445068,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097731239092496,
+ "grad_norm": 0.055207233875989914,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 4.155735015869141,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111692844677137,
+ "grad_norm": 0.05484927073121071,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 4.140205383300781,
+ "step": 1376
+ },
+ {
+ "epoch": 19.12565445026178,
+ "grad_norm": 0.05589596927165985,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 4.160840034484863,
+ "step": 1377
+ },
+ {
+ "epoch": 19.13961605584642,
+ "grad_norm": 0.05222902446985245,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 4.154781341552734,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153577661431065,
+ "grad_norm": 0.051824700087308884,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 4.134243488311768,
+ "step": 1379
+ },
+ {
+ "epoch": 19.167539267015705,
+ "grad_norm": 0.0517590269446373,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 4.125945568084717,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18150087260035,
+ "grad_norm": 0.05399726703763008,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 4.146993160247803,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195462478184993,
+ "grad_norm": 0.054430052638053894,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 4.1310906410217285,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209424083769633,
+ "grad_norm": 0.0551968477666378,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 4.170640468597412,
+ "step": 1383
+ },
+ {
+ "epoch": 19.223385689354277,
+ "grad_norm": 0.06050962209701538,
+ "learning_rate": 0.000423881957237287,
+ "loss": 4.166876316070557,
+ "step": 1384
+ },
+ {
+ "epoch": 19.237347294938917,
+ "grad_norm": 0.055587392300367355,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 4.150339603424072,
+ "step": 1385
+ },
+ {
+ "epoch": 19.25130890052356,
+ "grad_norm": 0.05193328112363815,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 4.171754360198975,
+ "step": 1386
+ },
+ {
+ "epoch": 19.2652705061082,
+ "grad_norm": 0.05272402986884117,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 4.165149688720703,
+ "step": 1387
+ },
+ {
+ "epoch": 19.279232111692846,
+ "grad_norm": 0.056283943355083466,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 4.193273544311523,
+ "step": 1388
+ },
+ {
+ "epoch": 19.293193717277486,
+ "grad_norm": 0.05825977027416229,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 4.192652225494385,
+ "step": 1389
+ },
+ {
+ "epoch": 19.30715532286213,
+ "grad_norm": 0.05539379641413689,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 4.187044143676758,
+ "step": 1390
+ },
+ {
+ "epoch": 19.32111692844677,
+ "grad_norm": 0.05300498753786087,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 4.168220043182373,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335078534031414,
+ "grad_norm": 0.05425756797194481,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 4.185193061828613,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349040139616054,
+ "grad_norm": 0.05527182295918465,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 4.1618332862854,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363001745200698,
+ "grad_norm": 0.054422613233327866,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 4.195295810699463,
+ "step": 1394
+ },
+ {
+ "epoch": 19.376963350785342,
+ "grad_norm": 0.052717432379722595,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 4.143290996551514,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390924956369982,
+ "grad_norm": 0.05154350399971008,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 4.115989685058594,
+ "step": 1396
+ },
+ {
+ "epoch": 19.404886561954626,
+ "grad_norm": 0.0489451140165329,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 4.169506072998047,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418848167539267,
+ "grad_norm": 0.049093786627054214,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 4.164122104644775,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43280977312391,
+ "grad_norm": 0.04690805822610855,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 4.186679840087891,
+ "step": 1399
+ },
+ {
+ "epoch": 19.44677137870855,
+ "grad_norm": 0.046708665788173676,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 4.184325695037842,
+ "step": 1400
+ },
+ {
+ "epoch": 19.460732984293195,
+ "grad_norm": 0.04617927595973015,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 4.201678276062012,
+ "step": 1401
+ },
+ {
+ "epoch": 19.474694589877835,
+ "grad_norm": 0.05017530545592308,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 4.18928337097168,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48865619546248,
+ "grad_norm": 0.05197977274656296,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 4.188801288604736,
+ "step": 1403
+ },
+ {
+ "epoch": 19.50261780104712,
+ "grad_norm": 0.049691297113895416,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 4.209467887878418,
+ "step": 1404
+ },
+ {
+ "epoch": 19.516579406631763,
+ "grad_norm": 0.049709953367710114,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 4.19733190536499,
+ "step": 1405
+ },
+ {
+ "epoch": 19.530541012216403,
+ "grad_norm": 0.05274137482047081,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 4.221798896789551,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544502617801047,
+ "grad_norm": 0.05015392601490021,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 4.210565567016602,
+ "step": 1407
+ },
+ {
+ "epoch": 19.55846422338569,
+ "grad_norm": 0.05023058131337166,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 4.184835910797119,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57242582897033,
+ "grad_norm": 0.0479239858686924,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 4.199846267700195,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586387434554975,
+ "grad_norm": 0.0481228232383728,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 4.185157775878906,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600349040139616,
+ "grad_norm": 0.04909738525748253,
+ "learning_rate": 0.000417272427439646,
+ "loss": 4.207550525665283,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61431064572426,
+ "grad_norm": 0.0475567989051342,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 4.213225841522217,
+ "step": 1412
+ },
+ {
+ "epoch": 19.6282722513089,
+ "grad_norm": 0.048073310405015945,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 4.212669372558594,
+ "step": 1413
+ },
+ {
+ "epoch": 19.642233856893544,
+ "grad_norm": 0.04523796588182449,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 4.1950860023498535,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656195462478184,
+ "grad_norm": 0.04359745234251022,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 4.2136640548706055,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670157068062828,
+ "grad_norm": 0.047282423824071884,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 4.209680557250977,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68411867364747,
+ "grad_norm": 0.0474061481654644,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 4.188494682312012,
+ "step": 1417
+ },
+ {
+ "epoch": 19.698080279232112,
+ "grad_norm": 0.04402102902531624,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 4.207236289978027,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712041884816752,
+ "grad_norm": 0.04398655146360397,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 4.210899353027344,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726003490401396,
+ "grad_norm": 0.045767925679683685,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 4.171738624572754,
+ "step": 1420
+ },
+ {
+ "epoch": 19.739965095986037,
+ "grad_norm": 0.04498986527323723,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 4.222414970397949,
+ "step": 1421
+ },
+ {
+ "epoch": 19.75392670157068,
+ "grad_norm": 0.04282551258802414,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 4.202138900756836,
+ "step": 1422
+ },
+ {
+ "epoch": 19.767888307155324,
+ "grad_norm": 0.04392262175679207,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 4.184439659118652,
+ "step": 1423
+ },
+ {
+ "epoch": 19.781849912739965,
+ "grad_norm": 0.04343154653906822,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 4.204674243927002,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79581151832461,
+ "grad_norm": 0.04496419429779053,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 4.23316764831543,
+ "step": 1425
+ },
+ {
+ "epoch": 19.80977312390925,
+ "grad_norm": 0.04488370940089226,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 4.201659202575684,
+ "step": 1426
+ },
+ {
+ "epoch": 19.823734729493893,
+ "grad_norm": 0.045598775148391724,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 4.187412261962891,
+ "step": 1427
+ },
+ {
+ "epoch": 19.837696335078533,
+ "grad_norm": 0.043846048414707184,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 4.215444564819336,
+ "step": 1428
+ },
+ {
+ "epoch": 19.851657940663177,
+ "grad_norm": 0.043062254786491394,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 4.2074055671691895,
+ "step": 1429
+ },
+ {
+ "epoch": 19.865619546247817,
+ "grad_norm": 0.04544302448630333,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 4.184919357299805,
+ "step": 1430
+ },
+ {
+ "epoch": 19.87958115183246,
+ "grad_norm": 0.046144865453243256,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 4.2175750732421875,
+ "step": 1431
+ },
+ {
+ "epoch": 19.8935427574171,
+ "grad_norm": 0.04465445503592491,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 4.1851115226745605,
+ "step": 1432
+ },
+ {
+ "epoch": 19.907504363001745,
+ "grad_norm": 0.045913565903902054,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 4.2226152420043945,
+ "step": 1433
+ },
+ {
+ "epoch": 19.921465968586386,
+ "grad_norm": 0.04401300847530365,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 4.216772556304932,
+ "step": 1434
+ },
+ {
+ "epoch": 19.93542757417103,
+ "grad_norm": 0.042624711990356445,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 4.2133684158325195,
+ "step": 1435
+ },
+ {
+ "epoch": 19.949389179755673,
+ "grad_norm": 0.04533256217837334,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 4.218449592590332,
+ "step": 1436
+ },
+ {
+ "epoch": 19.963350785340314,
+ "grad_norm": 0.04764309898018837,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 4.191402912139893,
+ "step": 1437
+ },
+ {
+ "epoch": 19.977312390924958,
+ "grad_norm": 0.04689318686723709,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 4.2113037109375,
+ "step": 1438
+ },
+ {
+ "epoch": 19.991273996509598,
+ "grad_norm": 0.04479913413524628,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 4.194989204406738,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.03151386231184006,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 2.6285438537597656,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 0.5938133597373962,
+ "eval_runtime": 58.9656,
+ "eval_samples_per_second": 41.414,
+ "eval_steps_per_second": 0.661,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013961605584644,
+ "grad_norm": 0.04884875938296318,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 4.1257734298706055,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027923211169284,
+ "grad_norm": 0.05968257039785385,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 4.091641426086426,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041884816753928,
+ "grad_norm": 0.06437838077545166,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 4.122465133666992,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05584642233857,
+ "grad_norm": 0.057831212878227234,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 4.129654884338379,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069808027923212,
+ "grad_norm": 0.049407217651605606,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 4.116935729980469,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083769633507853,
+ "grad_norm": 0.051957421004772186,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 4.125039577484131,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097731239092496,
+ "grad_norm": 0.05425718426704407,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 4.113485336303711,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111692844677137,
+ "grad_norm": 0.055318959057331085,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 4.119422912597656,
+ "step": 1448
+ },
+ {
+ "epoch": 20.12565445026178,
+ "grad_norm": 0.0548059344291687,
+ "learning_rate": 0.000407857329622967,
+ "loss": 4.097405910491943,
+ "step": 1449
+ },
+ {
+ "epoch": 20.13961605584642,
+ "grad_norm": 0.051046356558799744,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 4.115051746368408,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153577661431065,
+ "grad_norm": 0.05079492926597595,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 4.126070499420166,
+ "step": 1451
+ },
+ {
+ "epoch": 20.167539267015705,
+ "grad_norm": 0.04777265712618828,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 4.118952751159668,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18150087260035,
+ "grad_norm": 0.04895703122019768,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 4.095477104187012,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195462478184993,
+ "grad_norm": 0.051196496933698654,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 4.116866111755371,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209424083769633,
+ "grad_norm": 0.05625293031334877,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 4.141980171203613,
+ "step": 1455
+ },
+ {
+ "epoch": 20.223385689354277,
+ "grad_norm": 0.05852660909295082,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 4.126821517944336,
+ "step": 1456
+ },
+ {
+ "epoch": 20.237347294938917,
+ "grad_norm": 0.05419151484966278,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 4.10935640335083,
+ "step": 1457
+ },
+ {
+ "epoch": 20.25130890052356,
+ "grad_norm": 0.05251913517713547,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 4.151289463043213,
+ "step": 1458
+ },
+ {
+ "epoch": 20.2652705061082,
+ "grad_norm": 0.05223092809319496,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 4.149935722351074,
+ "step": 1459
+ },
+ {
+ "epoch": 20.279232111692846,
+ "grad_norm": 0.05314765125513077,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 4.106821060180664,
+ "step": 1460
+ },
+ {
+ "epoch": 20.293193717277486,
+ "grad_norm": 0.05472502112388611,
+ "learning_rate": 0.000404858275823277,
+ "loss": 4.120498180389404,
+ "step": 1461
+ },
+ {
+ "epoch": 20.30715532286213,
+ "grad_norm": 0.05233139917254448,
+ "learning_rate": 0.000404607816444578,
+ "loss": 4.1237335205078125,
+ "step": 1462
+ },
+ {
+ "epoch": 20.32111692844677,
+ "grad_norm": 0.05210232734680176,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 4.1271209716796875,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335078534031414,
+ "grad_norm": 0.049779925495386124,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 4.121427536010742,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349040139616054,
+ "grad_norm": 0.05232279747724533,
+ "learning_rate": 0.000403855947934478,
+ "loss": 4.136429786682129,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363001745200698,
+ "grad_norm": 0.05235981196165085,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 4.1808600425720215,
+ "step": 1466
+ },
+ {
+ "epoch": 20.376963350785342,
+ "grad_norm": 0.04902610555291176,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 4.138566017150879,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390924956369982,
+ "grad_norm": 0.04923798143863678,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 4.135126113891602,
+ "step": 1468
+ },
+ {
+ "epoch": 20.404886561954626,
+ "grad_norm": 0.05059202015399933,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 4.1548285484313965,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418848167539267,
+ "grad_norm": 0.05058307573199272,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 4.149041652679443,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43280977312391,
+ "grad_norm": 0.050172705203294754,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 4.148054122924805,
+ "step": 1471
+ },
+ {
+ "epoch": 20.44677137870855,
+ "grad_norm": 0.05158299580216408,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 4.1425018310546875,
+ "step": 1472
+ },
+ {
+ "epoch": 20.460732984293195,
+ "grad_norm": 0.052176836878061295,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 4.129712104797363,
+ "step": 1473
+ },
+ {
+ "epoch": 20.474694589877835,
+ "grad_norm": 0.052095964550971985,
+ "learning_rate": 0.000401595976318565,
+ "loss": 4.134920597076416,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48865619546248,
+ "grad_norm": 0.04944197088479996,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 4.165334701538086,
+ "step": 1475
+ },
+ {
+ "epoch": 20.50261780104712,
+ "grad_norm": 0.049351636320352554,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 4.144362926483154,
+ "step": 1476
+ },
+ {
+ "epoch": 20.516579406631763,
+ "grad_norm": 0.053657080978155136,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 4.164234161376953,
+ "step": 1477
+ },
+ {
+ "epoch": 20.530541012216403,
+ "grad_norm": 0.052578140050172806,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 4.131514549255371,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544502617801047,
+ "grad_norm": 0.049683794379234314,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 4.129961967468262,
+ "step": 1479
+ },
+ {
+ "epoch": 20.55846422338569,
+ "grad_norm": 0.05115316063165665,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 4.164663314819336,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57242582897033,
+ "grad_norm": 0.051620446145534515,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 4.161067485809326,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586387434554975,
+ "grad_norm": 0.04873501881957054,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 4.161318302154541,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600349040139616,
+ "grad_norm": 0.04968130216002464,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 4.151819229125977,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61431064572426,
+ "grad_norm": 0.05476681888103485,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 4.14247989654541,
+ "step": 1484
+ },
+ {
+ "epoch": 20.6282722513089,
+ "grad_norm": 0.05098642408847809,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 4.178083419799805,
+ "step": 1485
+ },
+ {
+ "epoch": 20.642233856893544,
+ "grad_norm": 0.047264356166124344,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 4.15170955657959,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656195462478184,
+ "grad_norm": 0.050382886081933975,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 4.166364669799805,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670157068062828,
+ "grad_norm": 0.04956239461898804,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 4.164124488830566,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68411867364747,
+ "grad_norm": 0.049608584493398666,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 4.130239486694336,
+ "step": 1489
+ },
+ {
+ "epoch": 20.698080279232112,
+ "grad_norm": 0.0489494763314724,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 4.163503646850586,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712041884816752,
+ "grad_norm": 0.04847714304924011,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 4.1744585037231445,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726003490401396,
+ "grad_norm": 0.050256017595529556,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 4.1502227783203125,
+ "step": 1492
+ },
+ {
+ "epoch": 20.739965095986037,
+ "grad_norm": 0.048107463866472244,
+ "learning_rate": 0.000396803974909638,
+ "loss": 4.1790971755981445,
+ "step": 1493
+ },
+ {
+ "epoch": 20.75392670157068,
+ "grad_norm": 0.04828944429755211,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 4.128883361816406,
+ "step": 1494
+ },
+ {
+ "epoch": 20.767888307155324,
+ "grad_norm": 0.04720750078558922,
+ "learning_rate": 0.000396297942393265,
+ "loss": 4.1661577224731445,
+ "step": 1495
+ },
+ {
+ "epoch": 20.781849912739965,
+ "grad_norm": 0.047148339450359344,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 4.172522068023682,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79581151832461,
+ "grad_norm": 0.04612366110086441,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 4.149024963378906,
+ "step": 1497
+ },
+ {
+ "epoch": 20.80977312390925,
+ "grad_norm": 0.048896901309490204,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 4.179235458374023,
+ "step": 1498
+ },
+ {
+ "epoch": 20.823734729493893,
+ "grad_norm": 0.04779870808124542,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 4.1649932861328125,
+ "step": 1499
+ },
+ {
+ "epoch": 20.837696335078533,
+ "grad_norm": 0.04949707165360451,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 4.17645263671875,
+ "step": 1500
+ },
+ {
+ "epoch": 20.851657940663177,
+ "grad_norm": 0.04645023122429848,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 4.216496467590332,
+ "step": 1501
+ },
+ {
+ "epoch": 20.865619546247817,
+ "grad_norm": 0.046520136296749115,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 4.152667999267578,
+ "step": 1502
+ },
+ {
+ "epoch": 20.87958115183246,
+ "grad_norm": 0.04882653057575226,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 4.187082767486572,
+ "step": 1503
+ },
+ {
+ "epoch": 20.8935427574171,
+ "grad_norm": 0.04669157415628433,
+ "learning_rate": 0.000394017098798633,
+ "loss": 4.172348499298096,
+ "step": 1504
+ },
+ {
+ "epoch": 20.907504363001745,
+ "grad_norm": 0.04595336690545082,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 4.180559158325195,
+ "step": 1505
+ },
+ {
+ "epoch": 20.921465968586386,
+ "grad_norm": 0.048478737473487854,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 4.142880439758301,
+ "step": 1506
+ },
+ {
+ "epoch": 20.93542757417103,
+ "grad_norm": 0.04609905555844307,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 4.1459641456604,
+ "step": 1507
+ },
+ {
+ "epoch": 20.949389179755673,
+ "grad_norm": 0.04524408280849457,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 4.185099124908447,
+ "step": 1508
+ },
+ {
+ "epoch": 20.963350785340314,
+ "grad_norm": 0.0475911870598793,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 4.2085161209106445,
+ "step": 1509
+ },
+ {
+ "epoch": 20.977312390924958,
+ "grad_norm": 0.047796521335840225,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 4.165589332580566,
+ "step": 1510
+ },
+ {
+ "epoch": 20.991273996509598,
+ "grad_norm": 0.047358278185129166,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 4.1634368896484375,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.03328625112771988,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 2.6023788452148438,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.5952173471450806,
+ "eval_runtime": 58.6816,
+ "eval_samples_per_second": 41.614,
+ "eval_steps_per_second": 0.665,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013961605584644,
+ "grad_norm": 0.04970651492476463,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 4.077130317687988,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027923211169284,
+ "grad_norm": 0.060882844030857086,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 4.074376106262207,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041884816753928,
+ "grad_norm": 0.06609825044870377,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 4.107635021209717,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05584642233857,
+ "grad_norm": 0.056066352874040604,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 4.068084716796875,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069808027923212,
+ "grad_norm": 0.053013063967227936,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 4.069967269897461,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083769633507853,
+ "grad_norm": 0.059536613523960114,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 4.0639142990112305,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097731239092496,
+ "grad_norm": 0.06084490567445755,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 4.093890190124512,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111692844677137,
+ "grad_norm": 0.05299340933561325,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 4.100567817687988,
+ "step": 1520
+ },
+ {
+ "epoch": 21.12565445026178,
+ "grad_norm": 0.05136161297559738,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 4.08171272277832,
+ "step": 1521
+ },
+ {
+ "epoch": 21.13961605584642,
+ "grad_norm": 0.05201329290866852,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 4.0869975090026855,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153577661431065,
+ "grad_norm": 0.053958263248205185,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 4.0928850173950195,
+ "step": 1523
+ },
+ {
+ "epoch": 21.167539267015705,
+ "grad_norm": 0.053465038537979126,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 4.097828388214111,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18150087260035,
+ "grad_norm": 0.050980277359485626,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 4.101456165313721,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195462478184993,
+ "grad_norm": 0.05415358766913414,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 4.09096097946167,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209424083769633,
+ "grad_norm": 0.05567048117518425,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 4.061287879943848,
+ "step": 1527
+ },
+ {
+ "epoch": 21.223385689354277,
+ "grad_norm": 0.052690375596284866,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 4.080180644989014,
+ "step": 1528
+ },
+ {
+ "epoch": 21.237347294938917,
+ "grad_norm": 0.05448971316218376,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 4.079794883728027,
+ "step": 1529
+ },
+ {
+ "epoch": 21.25130890052356,
+ "grad_norm": 0.055472876876592636,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 4.072025299072266,
+ "step": 1530
+ },
+ {
+ "epoch": 21.2652705061082,
+ "grad_norm": 0.05418326333165169,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 4.103858470916748,
+ "step": 1531
+ },
+ {
+ "epoch": 21.279232111692846,
+ "grad_norm": 0.0524221733212471,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 4.0808916091918945,
+ "step": 1532
+ },
+ {
+ "epoch": 21.293193717277486,
+ "grad_norm": 0.05224563553929329,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 4.090270042419434,
+ "step": 1533
+ },
+ {
+ "epoch": 21.30715532286213,
+ "grad_norm": 0.052917953580617905,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 4.081887245178223,
+ "step": 1534
+ },
+ {
+ "epoch": 21.32111692844677,
+ "grad_norm": 0.05619358271360397,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 4.089571475982666,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335078534031414,
+ "grad_norm": 0.05455326288938522,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 4.096992015838623,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349040139616054,
+ "grad_norm": 0.049311455339193344,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 4.114621639251709,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363001745200698,
+ "grad_norm": 0.05175276845693588,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 4.107684135437012,
+ "step": 1538
+ },
+ {
+ "epoch": 21.376963350785342,
+ "grad_norm": 0.05236738547682762,
+ "learning_rate": 0.00038509205478744,
+ "loss": 4.075709342956543,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390924956369982,
+ "grad_norm": 0.04899081960320473,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 4.108396530151367,
+ "step": 1540
+ },
+ {
+ "epoch": 21.404886561954626,
+ "grad_norm": 0.05129663646221161,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 4.1135711669921875,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418848167539267,
+ "grad_norm": 0.05478592589497566,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 4.11839485168457,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43280977312391,
+ "grad_norm": 0.055830392986536026,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 4.11065149307251,
+ "step": 1543
+ },
+ {
+ "epoch": 21.44677137870855,
+ "grad_norm": 0.05370553582906723,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 4.113103866577148,
+ "step": 1544
+ },
+ {
+ "epoch": 21.460732984293195,
+ "grad_norm": 0.04968908056616783,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 4.119443893432617,
+ "step": 1545
+ },
+ {
+ "epoch": 21.474694589877835,
+ "grad_norm": 0.056149110198020935,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 4.126479625701904,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48865619546248,
+ "grad_norm": 0.053045690059661865,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 4.138535499572754,
+ "step": 1547
+ },
+ {
+ "epoch": 21.50261780104712,
+ "grad_norm": 0.04864223673939705,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 4.099347114562988,
+ "step": 1548
+ },
+ {
+ "epoch": 21.516579406631763,
+ "grad_norm": 0.05234768986701965,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 4.112222194671631,
+ "step": 1549
+ },
+ {
+ "epoch": 21.530541012216403,
+ "grad_norm": 0.050486158579587936,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 4.107913970947266,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544502617801047,
+ "grad_norm": 0.048641227185726166,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 4.114014148712158,
+ "step": 1551
+ },
+ {
+ "epoch": 21.55846422338569,
+ "grad_norm": 0.047759316861629486,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 4.128541946411133,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57242582897033,
+ "grad_norm": 0.046098388731479645,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 4.12142276763916,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586387434554975,
+ "grad_norm": 0.04934769123792648,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 4.13035249710083,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600349040139616,
+ "grad_norm": 0.048911672085523605,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 4.111344337463379,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61431064572426,
+ "grad_norm": 0.049368683248758316,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 4.1345672607421875,
+ "step": 1556
+ },
+ {
+ "epoch": 21.6282722513089,
+ "grad_norm": 0.04890736937522888,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 4.1286821365356445,
+ "step": 1557
+ },
+ {
+ "epoch": 21.642233856893544,
+ "grad_norm": 0.04812583327293396,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 4.09993314743042,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656195462478184,
+ "grad_norm": 0.05025152117013931,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 4.150769233703613,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670157068062828,
+ "grad_norm": 0.05007397383451462,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 4.074525833129883,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68411867364747,
+ "grad_norm": 0.050168465822935104,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 4.0933732986450195,
+ "step": 1561
+ },
+ {
+ "epoch": 21.698080279232112,
+ "grad_norm": 0.04990607872605324,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 4.126741886138916,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712041884816752,
+ "grad_norm": 0.05023440346121788,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 4.120314598083496,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726003490401396,
+ "grad_norm": 0.048100546002388,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 4.1242356300354,
+ "step": 1564
+ },
+ {
+ "epoch": 21.739965095986037,
+ "grad_norm": 0.047999415546655655,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 4.127745628356934,
+ "step": 1565
+ },
+ {
+ "epoch": 21.75392670157068,
+ "grad_norm": 0.05180094763636589,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 4.112565994262695,
+ "step": 1566
+ },
+ {
+ "epoch": 21.767888307155324,
+ "grad_norm": 0.053367964923381805,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 4.109419822692871,
+ "step": 1567
+ },
+ {
+ "epoch": 21.781849912739965,
+ "grad_norm": 0.05205400660634041,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 4.111886024475098,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79581151832461,
+ "grad_norm": 0.0464242659509182,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 4.104648590087891,
+ "step": 1569
+ },
+ {
+ "epoch": 21.80977312390925,
+ "grad_norm": 0.047767721116542816,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 4.156681060791016,
+ "step": 1570
+ },
+ {
+ "epoch": 21.823734729493893,
+ "grad_norm": 0.05028790235519409,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 4.132200241088867,
+ "step": 1571
+ },
+ {
+ "epoch": 21.837696335078533,
+ "grad_norm": 0.04714380204677582,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 4.132688999176025,
+ "step": 1572
+ },
+ {
+ "epoch": 21.851657940663177,
+ "grad_norm": 0.04726320132613182,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 4.173277378082275,
+ "step": 1573
+ },
+ {
+ "epoch": 21.865619546247817,
+ "grad_norm": 0.049558594822883606,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 4.15473747253418,
+ "step": 1574
+ },
+ {
+ "epoch": 21.87958115183246,
+ "grad_norm": 0.04742267727851868,
+ "learning_rate": 0.000375827577564042,
+ "loss": 4.10634708404541,
+ "step": 1575
+ },
+ {
+ "epoch": 21.8935427574171,
+ "grad_norm": 0.048223357647657394,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 4.133949279785156,
+ "step": 1576
+ },
+ {
+ "epoch": 21.907504363001745,
+ "grad_norm": 0.046766068786382675,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 4.153807640075684,
+ "step": 1577
+ },
+ {
+ "epoch": 21.921465968586386,
+ "grad_norm": 0.04586656019091606,
+ "learning_rate": 0.000375051971546236,
+ "loss": 4.169903755187988,
+ "step": 1578
+ },
+ {
+ "epoch": 21.93542757417103,
+ "grad_norm": 0.047912538051605225,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 4.148102283477783,
+ "step": 1579
+ },
+ {
+ "epoch": 21.949389179755673,
+ "grad_norm": 0.048949215561151505,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 4.149707317352295,
+ "step": 1580
+ },
+ {
+ "epoch": 21.963350785340314,
+ "grad_norm": 0.04765952378511429,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 4.128008842468262,
+ "step": 1581
+ },
+ {
+ "epoch": 21.977312390924958,
+ "grad_norm": 0.04726482927799225,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 4.1268720626831055,
+ "step": 1582
+ },
+ {
+ "epoch": 21.991273996509598,
+ "grad_norm": 0.0468299426138401,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 4.1332926750183105,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.03415689244866371,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 2.5875940322875977,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.5966277122497559,
+ "eval_runtime": 59.7704,
+ "eval_samples_per_second": 40.856,
+ "eval_steps_per_second": 0.652,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013961605584644,
+ "grad_norm": 0.04954535514116287,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 4.054527759552002,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027923211169284,
+ "grad_norm": 0.0498427115380764,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 4.048842430114746,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041884816753928,
+ "grad_norm": 0.051013074815273285,
+ "learning_rate": 0.000372722041257003,
+ "loss": 4.036187171936035,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05584642233857,
+ "grad_norm": 0.04973271116614342,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 4.050318717956543,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069808027923212,
+ "grad_norm": 0.051144856959581375,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 4.036319732666016,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083769633507853,
+ "grad_norm": 0.051874030381441116,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 4.017857551574707,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097731239092496,
+ "grad_norm": 0.05181736871600151,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 4.0519819259643555,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111692844677137,
+ "grad_norm": 0.05467049032449722,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 4.064291954040527,
+ "step": 1592
+ },
+ {
+ "epoch": 22.12565445026178,
+ "grad_norm": 0.05251661688089371,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 4.056951522827148,
+ "step": 1593
+ },
+ {
+ "epoch": 22.13961605584642,
+ "grad_norm": 0.05278680846095085,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 4.037125587463379,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153577661431065,
+ "grad_norm": 0.05276667699217796,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 4.053675174713135,
+ "step": 1595
+ },
+ {
+ "epoch": 22.167539267015705,
+ "grad_norm": 0.05007271468639374,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 4.044122695922852,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18150087260035,
+ "grad_norm": 0.051073770970106125,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 4.0265960693359375,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195462478184993,
+ "grad_norm": 0.05045647919178009,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 4.060938835144043,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209424083769633,
+ "grad_norm": 0.05075158178806305,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 4.035902976989746,
+ "step": 1599
+ },
+ {
+ "epoch": 22.223385689354277,
+ "grad_norm": 0.04867244139313698,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 4.055220603942871,
+ "step": 1600
+ },
+ {
+ "epoch": 22.237347294938917,
+ "grad_norm": 0.049976419657468796,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 4.026124954223633,
+ "step": 1601
+ },
+ {
+ "epoch": 22.25130890052356,
+ "grad_norm": 0.04968957230448723,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 4.081721782684326,
+ "step": 1602
+ },
+ {
+ "epoch": 22.2652705061082,
+ "grad_norm": 0.0503058098256588,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 4.0384721755981445,
+ "step": 1603
+ },
+ {
+ "epoch": 22.279232111692846,
+ "grad_norm": 0.050643209367990494,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 4.049397945404053,
+ "step": 1604
+ },
+ {
+ "epoch": 22.293193717277486,
+ "grad_norm": 0.05297046899795532,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 4.046389579772949,
+ "step": 1605
+ },
+ {
+ "epoch": 22.30715532286213,
+ "grad_norm": 0.053165238350629807,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 4.090221405029297,
+ "step": 1606
+ },
+ {
+ "epoch": 22.32111692844677,
+ "grad_norm": 0.05104832351207733,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 4.023227691650391,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335078534031414,
+ "grad_norm": 0.048736415803432465,
+ "learning_rate": 0.000367268105959126,
+ "loss": 4.07478666305542,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349040139616054,
+ "grad_norm": 0.05190873518586159,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 4.068147659301758,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363001745200698,
+ "grad_norm": 0.05409928783774376,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 4.060364723205566,
+ "step": 1610
+ },
+ {
+ "epoch": 22.376963350785342,
+ "grad_norm": 0.05112519487738609,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 4.072710990905762,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390924956369982,
+ "grad_norm": 0.04818312078714371,
+ "learning_rate": 0.000366226612810024,
+ "loss": 4.042757034301758,
+ "step": 1612
+ },
+ {
+ "epoch": 22.404886561954626,
+ "grad_norm": 0.050733599811792374,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 4.064810752868652,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418848167539267,
+ "grad_norm": 0.054070211946964264,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 4.054926872253418,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43280977312391,
+ "grad_norm": 0.05459332838654518,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 4.0652923583984375,
+ "step": 1615
+ },
+ {
+ "epoch": 22.44677137870855,
+ "grad_norm": 0.05064154416322708,
+ "learning_rate": 0.000365184304613104,
+ "loss": 4.076216220855713,
+ "step": 1616
+ },
+ {
+ "epoch": 22.460732984293195,
+ "grad_norm": 0.05071895569562912,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 4.070466041564941,
+ "step": 1617
+ },
+ {
+ "epoch": 22.474694589877835,
+ "grad_norm": 0.05049987882375717,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 4.062691688537598,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48865619546248,
+ "grad_norm": 0.05174451321363449,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 4.07501220703125,
+ "step": 1619
+ },
+ {
+ "epoch": 22.50261780104712,
+ "grad_norm": 0.05360838770866394,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 4.054711818695068,
+ "step": 1620
+ },
+ {
+ "epoch": 22.516579406631763,
+ "grad_norm": 0.051809027791023254,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 4.075811386108398,
+ "step": 1621
+ },
+ {
+ "epoch": 22.530541012216403,
+ "grad_norm": 0.05586795508861542,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 4.0638203620910645,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544502617801047,
+ "grad_norm": 0.05571979284286499,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 4.089414596557617,
+ "step": 1623
+ },
+ {
+ "epoch": 22.55846422338569,
+ "grad_norm": 0.05098723992705345,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 4.091159820556641,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57242582897033,
+ "grad_norm": 0.05359277129173279,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 4.060613632202148,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586387434554975,
+ "grad_norm": 0.05281100794672966,
+ "learning_rate": 0.000362575056487632,
+ "loss": 4.08416748046875,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600349040139616,
+ "grad_norm": 0.053913842886686325,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 4.088703155517578,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61431064572426,
+ "grad_norm": 0.05476798862218857,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 4.064074516296387,
+ "step": 1628
+ },
+ {
+ "epoch": 22.6282722513089,
+ "grad_norm": 0.051888495683670044,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 4.116634368896484,
+ "step": 1629
+ },
+ {
+ "epoch": 22.642233856893544,
+ "grad_norm": 0.051054876297712326,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 4.0800018310546875,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656195462478184,
+ "grad_norm": 0.054063159972429276,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 4.112773418426514,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670157068062828,
+ "grad_norm": 0.05173034593462944,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 4.062599182128906,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68411867364747,
+ "grad_norm": 0.054618336260318756,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 4.110188961029053,
+ "step": 1633
+ },
+ {
+ "epoch": 22.698080279232112,
+ "grad_norm": 0.055602896958589554,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 4.140998840332031,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712041884816752,
+ "grad_norm": 0.05332222580909729,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 4.119804382324219,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726003490401396,
+ "grad_norm": 0.05162122845649719,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 4.0731282234191895,
+ "step": 1636
+ },
+ {
+ "epoch": 22.739965095986037,
+ "grad_norm": 0.0514204278588295,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 4.068007946014404,
+ "step": 1637
+ },
+ {
+ "epoch": 22.75392670157068,
+ "grad_norm": 0.05027637258172035,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 4.097658157348633,
+ "step": 1638
+ },
+ {
+ "epoch": 22.767888307155324,
+ "grad_norm": 0.051812347024679184,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 4.083312511444092,
+ "step": 1639
+ },
+ {
+ "epoch": 22.781849912739965,
+ "grad_norm": 0.05145815759897232,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 4.129471778869629,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79581151832461,
+ "grad_norm": 0.05093684792518616,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 4.091841220855713,
+ "step": 1641
+ },
+ {
+ "epoch": 22.80977312390925,
+ "grad_norm": 0.05100737139582634,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 4.076615810394287,
+ "step": 1642
+ },
+ {
+ "epoch": 22.823734729493893,
+ "grad_norm": 0.05079149082303047,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 4.067737579345703,
+ "step": 1643
+ },
+ {
+ "epoch": 22.837696335078533,
+ "grad_norm": 0.05044509842991829,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 4.118824481964111,
+ "step": 1644
+ },
+ {
+ "epoch": 22.851657940663177,
+ "grad_norm": 0.05010584369301796,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 4.113405227661133,
+ "step": 1645
+ },
+ {
+ "epoch": 22.865619546247817,
+ "grad_norm": 0.052232302725315094,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 4.101553440093994,
+ "step": 1646
+ },
+ {
+ "epoch": 22.87958115183246,
+ "grad_norm": 0.05080661550164223,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 4.116809844970703,
+ "step": 1647
+ },
+ {
+ "epoch": 22.8935427574171,
+ "grad_norm": 0.05051033943891525,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 4.128484725952148,
+ "step": 1648
+ },
+ {
+ "epoch": 22.907504363001745,
+ "grad_norm": 0.051927294582128525,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 4.092285633087158,
+ "step": 1649
+ },
+ {
+ "epoch": 22.921465968586386,
+ "grad_norm": 0.049063827842473984,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 4.088473796844482,
+ "step": 1650
+ },
+ {
+ "epoch": 22.93542757417103,
+ "grad_norm": 0.05048854276537895,
+ "learning_rate": 0.000356031397755195,
+ "loss": 4.081521987915039,
+ "step": 1651
+ },
+ {
+ "epoch": 22.949389179755673,
+ "grad_norm": 0.049307651817798615,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 4.122158050537109,
+ "step": 1652
+ },
+ {
+ "epoch": 22.963350785340314,
+ "grad_norm": 0.048311345279216766,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 4.100488662719727,
+ "step": 1653
+ },
+ {
+ "epoch": 22.977312390924958,
+ "grad_norm": 0.05016734451055527,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 4.088587760925293,
+ "step": 1654
+ },
+ {
+ "epoch": 22.991273996509598,
+ "grad_norm": 0.04944450408220291,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 4.114686012268066,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.03592754527926445,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 2.545376777648926,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.5981184840202332,
+ "eval_runtime": 59.9558,
+ "eval_samples_per_second": 40.73,
+ "eval_steps_per_second": 0.65,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013961605584644,
+ "grad_norm": 0.05112785845994949,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 4.00495719909668,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027923211169284,
+ "grad_norm": 0.05807499587535858,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 4.015654563903809,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041884816753928,
+ "grad_norm": 0.05383148789405823,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.979349136352539,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05584642233857,
+ "grad_norm": 0.05221369490027428,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 4.009493827819824,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069808027923212,
+ "grad_norm": 0.05144324526190758,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 4.045680999755859,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083769633507853,
+ "grad_norm": 0.051422782242298126,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.994011878967285,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097731239092496,
+ "grad_norm": 0.05142863467335701,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 4.03209114074707,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111692844677137,
+ "grad_norm": 0.050954073667526245,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 4.00391960144043,
+ "step": 1664
+ },
+ {
+ "epoch": 23.12565445026178,
+ "grad_norm": 0.05365024879574776,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.9839367866516113,
+ "step": 1665
+ },
+ {
+ "epoch": 23.13961605584642,
+ "grad_norm": 0.051181286573410034,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 4.017707824707031,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153577661431065,
+ "grad_norm": 0.054279860109090805,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.989722490310669,
+ "step": 1667
+ },
+ {
+ "epoch": 23.167539267015705,
+ "grad_norm": 0.052196286618709564,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 4.014932155609131,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18150087260035,
+ "grad_norm": 0.05098043382167816,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 4.026510238647461,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195462478184993,
+ "grad_norm": 0.05241888761520386,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.979708433151245,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209424083769633,
+ "grad_norm": 0.051801104098558426,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 4.029015064239502,
+ "step": 1671
+ },
+ {
+ "epoch": 23.223385689354277,
+ "grad_norm": 0.052666619420051575,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.9932432174682617,
+ "step": 1672
+ },
+ {
+ "epoch": 23.237347294938917,
+ "grad_norm": 0.054418303072452545,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 4.05676794052124,
+ "step": 1673
+ },
+ {
+ "epoch": 23.25130890052356,
+ "grad_norm": 0.055074021220207214,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.999929666519165,
+ "step": 1674
+ },
+ {
+ "epoch": 23.2652705061082,
+ "grad_norm": 0.05626136437058449,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 4.013473987579346,
+ "step": 1675
+ },
+ {
+ "epoch": 23.279232111692846,
+ "grad_norm": 0.054028257727622986,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 4.015802383422852,
+ "step": 1676
+ },
+ {
+ "epoch": 23.293193717277486,
+ "grad_norm": 0.054044198244810104,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 4.020464897155762,
+ "step": 1677
+ },
+ {
+ "epoch": 23.30715532286213,
+ "grad_norm": 0.05465259402990341,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 4.010867118835449,
+ "step": 1678
+ },
+ {
+ "epoch": 23.32111692844677,
+ "grad_norm": 0.05545585975050926,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 4.054412841796875,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335078534031414,
+ "grad_norm": 0.055265430361032486,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 4.020756244659424,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349040139616054,
+ "grad_norm": 0.053363267332315445,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 4.033337116241455,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363001745200698,
+ "grad_norm": 0.05347830802202225,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 4.024794578552246,
+ "step": 1682
+ },
+ {
+ "epoch": 23.376963350785342,
+ "grad_norm": 0.05496951565146446,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 4.013302803039551,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390924956369982,
+ "grad_norm": 0.05179804563522339,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 4.030914306640625,
+ "step": 1684
+ },
+ {
+ "epoch": 23.404886561954626,
+ "grad_norm": 0.054958634078502655,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 4.012727737426758,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418848167539267,
+ "grad_norm": 0.05533262714743614,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 4.0543365478515625,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43280977312391,
+ "grad_norm": 0.052329424768686295,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 4.031486511230469,
+ "step": 1687
+ },
+ {
+ "epoch": 23.44677137870855,
+ "grad_norm": 0.05273434892296791,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 4.040292739868164,
+ "step": 1688
+ },
+ {
+ "epoch": 23.460732984293195,
+ "grad_norm": 0.05445481091737747,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 4.025630950927734,
+ "step": 1689
+ },
+ {
+ "epoch": 23.474694589877835,
+ "grad_norm": 0.052102912217378616,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 4.049562454223633,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48865619546248,
+ "grad_norm": 0.051272716373205185,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 4.0431365966796875,
+ "step": 1691
+ },
+ {
+ "epoch": 23.50261780104712,
+ "grad_norm": 0.051311176270246506,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 4.042119026184082,
+ "step": 1692
+ },
+ {
+ "epoch": 23.516579406631763,
+ "grad_norm": 0.05042742192745209,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 4.044881820678711,
+ "step": 1693
+ },
+ {
+ "epoch": 23.530541012216403,
+ "grad_norm": 0.05103316903114319,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 4.033669471740723,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544502617801047,
+ "grad_norm": 0.05441150814294815,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 4.076536178588867,
+ "step": 1695
+ },
+ {
+ "epoch": 23.55846422338569,
+ "grad_norm": 0.05529651418328285,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 4.073678016662598,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57242582897033,
+ "grad_norm": 0.0526033453643322,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 4.067898750305176,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586387434554975,
+ "grad_norm": 0.052817948162555695,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 4.017403602600098,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600349040139616,
+ "grad_norm": 0.052461300045251846,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 4.026311874389648,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61431064572426,
+ "grad_norm": 0.05097689852118492,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 4.076074600219727,
+ "step": 1700
+ },
+ {
+ "epoch": 23.6282722513089,
+ "grad_norm": 0.05387786403298378,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 4.022763729095459,
+ "step": 1701
+ },
+ {
+ "epoch": 23.642233856893544,
+ "grad_norm": 0.05132855102419853,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 4.077712535858154,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656195462478184,
+ "grad_norm": 0.05380886048078537,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 4.038937091827393,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670157068062828,
+ "grad_norm": 0.053299430757761,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 4.063364028930664,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68411867364747,
+ "grad_norm": 0.051913514733314514,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 4.05495548248291,
+ "step": 1705
+ },
+ {
+ "epoch": 23.698080279232112,
+ "grad_norm": 0.05660073831677437,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 4.057949066162109,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712041884816752,
+ "grad_norm": 0.054757390171289444,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 4.05377721786499,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726003490401396,
+ "grad_norm": 0.0517287515103817,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 4.035928726196289,
+ "step": 1708
+ },
+ {
+ "epoch": 23.739965095986037,
+ "grad_norm": 0.05187627300620079,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 4.043362617492676,
+ "step": 1709
+ },
+ {
+ "epoch": 23.75392670157068,
+ "grad_norm": 0.053502220660448074,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 4.007593154907227,
+ "step": 1710
+ },
+ {
+ "epoch": 23.767888307155324,
+ "grad_norm": 0.05755629390478134,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 4.076575756072998,
+ "step": 1711
+ },
+ {
+ "epoch": 23.781849912739965,
+ "grad_norm": 0.05619359388947487,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 4.072240352630615,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79581151832461,
+ "grad_norm": 0.05604318529367447,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 4.0486650466918945,
+ "step": 1713
+ },
+ {
+ "epoch": 23.80977312390925,
+ "grad_norm": 0.05630200728774071,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 4.086282730102539,
+ "step": 1714
+ },
+ {
+ "epoch": 23.823734729493893,
+ "grad_norm": 0.053262822329998016,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 4.089612007141113,
+ "step": 1715
+ },
+ {
+ "epoch": 23.837696335078533,
+ "grad_norm": 0.05699621140956879,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 4.053062438964844,
+ "step": 1716
+ },
+ {
+ "epoch": 23.851657940663177,
+ "grad_norm": 0.05963658541440964,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 4.072884559631348,
+ "step": 1717
+ },
+ {
+ "epoch": 23.865619546247817,
+ "grad_norm": 0.0547492615878582,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 4.066218852996826,
+ "step": 1718
+ },
+ {
+ "epoch": 23.87958115183246,
+ "grad_norm": 0.054908450692892075,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 4.035344123840332,
+ "step": 1719
+ },
+ {
+ "epoch": 23.8935427574171,
+ "grad_norm": 0.055191367864608765,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 4.08471155166626,
+ "step": 1720
+ },
+ {
+ "epoch": 23.907504363001745,
+ "grad_norm": 0.05334226414561272,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 4.076174736022949,
+ "step": 1721
+ },
+ {
+ "epoch": 23.921465968586386,
+ "grad_norm": 0.05579005926847458,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 4.05291223526001,
+ "step": 1722
+ },
+ {
+ "epoch": 23.93542757417103,
+ "grad_norm": 0.05464969575405121,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 4.07335090637207,
+ "step": 1723
+ },
+ {
+ "epoch": 23.949389179755673,
+ "grad_norm": 0.05240754038095474,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 4.0995073318481445,
+ "step": 1724
+ },
+ {
+ "epoch": 23.963350785340314,
+ "grad_norm": 0.05416056886315346,
+ "learning_rate": 0.000336518346307424,
+ "loss": 4.072793960571289,
+ "step": 1725
+ },
+ {
+ "epoch": 23.977312390924958,
+ "grad_norm": 0.05496834218502045,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 4.059562683105469,
+ "step": 1726
+ },
+ {
+ "epoch": 23.991273996509598,
+ "grad_norm": 0.05537887290120125,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 4.050772666931152,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.036677468568086624,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 2.5407257080078125,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.5995977520942688,
+ "eval_runtime": 60.2704,
+ "eval_samples_per_second": 40.517,
+ "eval_steps_per_second": 0.647,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013961605584644,
+ "grad_norm": 0.05583682656288147,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.9873995780944824,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027923211169284,
+ "grad_norm": 0.0605769157409668,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.999206066131592,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041884816753928,
+ "grad_norm": 0.057975225150585175,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.9923095703125,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05584642233857,
+ "grad_norm": 0.056104276329278946,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.946317672729492,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069808027923212,
+ "grad_norm": 0.05481491982936859,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.9973483085632324,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083769633507853,
+ "grad_norm": 0.05460227280855179,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.984617233276367,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097731239092496,
+ "grad_norm": 0.058481909334659576,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 4.0138773918151855,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111692844677137,
+ "grad_norm": 0.05672750994563103,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.9683098793029785,
+ "step": 1736
+ },
+ {
+ "epoch": 24.12565445026178,
+ "grad_norm": 0.05315157398581505,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.9909560680389404,
+ "step": 1737
+ },
+ {
+ "epoch": 24.13961605584642,
+ "grad_norm": 0.05565650761127472,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.954531192779541,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153577661431065,
+ "grad_norm": 0.05550679191946983,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.97672700881958,
+ "step": 1739
+ },
+ {
+ "epoch": 24.167539267015705,
+ "grad_norm": 0.05608576908707619,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.987941265106201,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18150087260035,
+ "grad_norm": 0.05266846343874931,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.97526216506958,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195462478184993,
+ "grad_norm": 0.05439139902591705,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 4.012054920196533,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209424083769633,
+ "grad_norm": 0.05473192036151886,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 4.022945404052734,
+ "step": 1743
+ },
+ {
+ "epoch": 24.223385689354277,
+ "grad_norm": 0.054737839847803116,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.981072425842285,
+ "step": 1744
+ },
+ {
+ "epoch": 24.237347294938917,
+ "grad_norm": 0.05696691572666168,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.9708170890808105,
+ "step": 1745
+ },
+ {
+ "epoch": 24.25130890052356,
+ "grad_norm": 0.05677401274442673,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.985995292663574,
+ "step": 1746
+ },
+ {
+ "epoch": 24.2652705061082,
+ "grad_norm": 0.05895088613033295,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.991887092590332,
+ "step": 1747
+ },
+ {
+ "epoch": 24.279232111692846,
+ "grad_norm": 0.05758961662650108,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.9695687294006348,
+ "step": 1748
+ },
+ {
+ "epoch": 24.293193717277486,
+ "grad_norm": 0.05174405500292778,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.959138870239258,
+ "step": 1749
+ },
+ {
+ "epoch": 24.30715532286213,
+ "grad_norm": 0.055284611880779266,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 4.012698173522949,
+ "step": 1750
+ },
+ {
+ "epoch": 24.32111692844677,
+ "grad_norm": 0.06042446941137314,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.9892425537109375,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335078534031414,
+ "grad_norm": 0.05489135906100273,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.9962222576141357,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349040139616054,
+ "grad_norm": 0.0542326495051384,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.9787094593048096,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363001745200698,
+ "grad_norm": 0.05612313374876976,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.9881339073181152,
+ "step": 1754
+ },
+ {
+ "epoch": 24.376963350785342,
+ "grad_norm": 0.05661709979176521,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 4.009440898895264,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390924956369982,
+ "grad_norm": 0.055537596344947815,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 4.00221061706543,
+ "step": 1756
+ },
+ {
+ "epoch": 24.404886561954626,
+ "grad_norm": 0.054210033267736435,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.9807987213134766,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418848167539267,
+ "grad_norm": 0.05847020074725151,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.960838794708252,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43280977312391,
+ "grad_norm": 0.05386386811733246,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 4.017860412597656,
+ "step": 1759
+ },
+ {
+ "epoch": 24.44677137870855,
+ "grad_norm": 0.056131865829229355,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.980543613433838,
+ "step": 1760
+ },
+ {
+ "epoch": 24.460732984293195,
+ "grad_norm": 0.05648239329457283,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.9948508739471436,
+ "step": 1761
+ },
+ {
+ "epoch": 24.474694589877835,
+ "grad_norm": 0.054138120263814926,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 4.002998352050781,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48865619546248,
+ "grad_norm": 0.05725586414337158,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 4.01864767074585,
+ "step": 1763
+ },
+ {
+ "epoch": 24.50261780104712,
+ "grad_norm": 0.05680665746331215,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 4.023143768310547,
+ "step": 1764
+ },
+ {
+ "epoch": 24.516579406631763,
+ "grad_norm": 0.052627697587013245,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 4.006389617919922,
+ "step": 1765
+ },
+ {
+ "epoch": 24.530541012216403,
+ "grad_norm": 0.053234782069921494,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 4.007419586181641,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544502617801047,
+ "grad_norm": 0.05756475403904915,
+ "learning_rate": 0.000325372061241796,
+ "loss": 4.0226945877075195,
+ "step": 1767
+ },
+ {
+ "epoch": 24.55846422338569,
+ "grad_norm": 0.055328067392110825,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 4.022767066955566,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57242582897033,
+ "grad_norm": 0.05378621071577072,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 4.005022048950195,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586387434554975,
+ "grad_norm": 0.056049644947052,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 4.025008201599121,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600349040139616,
+ "grad_norm": 0.057860117405653,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 4.003593444824219,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61431064572426,
+ "grad_norm": 0.056685276329517365,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.997868299484253,
+ "step": 1772
+ },
+ {
+ "epoch": 24.6282722513089,
+ "grad_norm": 0.053294967859983444,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 4.0211992263793945,
+ "step": 1773
+ },
+ {
+ "epoch": 24.642233856893544,
+ "grad_norm": 0.053967639803886414,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 4.022367477416992,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656195462478184,
+ "grad_norm": 0.052357546985149384,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 4.021792411804199,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670157068062828,
+ "grad_norm": 0.0547296479344368,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 4.008403778076172,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68411867364747,
+ "grad_norm": 0.056104280054569244,
+ "learning_rate": 0.000322712903929477,
+ "loss": 4.033827781677246,
+ "step": 1777
+ },
+ {
+ "epoch": 24.698080279232112,
+ "grad_norm": 0.055415406823158264,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 4.015371322631836,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712041884816752,
+ "grad_norm": 0.054506298154592514,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 4.0000152587890625,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726003490401396,
+ "grad_norm": 0.05468682199716568,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.9905481338500977,
+ "step": 1780
+ },
+ {
+ "epoch": 24.739965095986037,
+ "grad_norm": 0.05323269963264465,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 4.040729999542236,
+ "step": 1781
+ },
+ {
+ "epoch": 24.75392670157068,
+ "grad_norm": 0.052569612860679626,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 4.028299331665039,
+ "step": 1782
+ },
+ {
+ "epoch": 24.767888307155324,
+ "grad_norm": 0.052494753152132034,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 4.028860569000244,
+ "step": 1783
+ },
+ {
+ "epoch": 24.781849912739965,
+ "grad_norm": 0.05346379801630974,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 4.000349044799805,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79581151832461,
+ "grad_norm": 0.051543332636356354,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 4.016782760620117,
+ "step": 1785
+ },
+ {
+ "epoch": 24.80977312390925,
+ "grad_norm": 0.05330873280763626,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 4.024371147155762,
+ "step": 1786
+ },
+ {
+ "epoch": 24.823734729493893,
+ "grad_norm": 0.05306831747293472,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 4.0101213455200195,
+ "step": 1787
+ },
+ {
+ "epoch": 24.837696335078533,
+ "grad_norm": 0.05148417130112648,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 4.020768642425537,
+ "step": 1788
+ },
+ {
+ "epoch": 24.851657940663177,
+ "grad_norm": 0.050848908722400665,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 4.030246734619141,
+ "step": 1789
+ },
+ {
+ "epoch": 24.865619546247817,
+ "grad_norm": 0.05279014632105827,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 4.04910135269165,
+ "step": 1790
+ },
+ {
+ "epoch": 24.87958115183246,
+ "grad_norm": 0.0514594204723835,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 4.025290489196777,
+ "step": 1791
+ },
+ {
+ "epoch": 24.8935427574171,
+ "grad_norm": 0.05200032889842987,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 4.019560813903809,
+ "step": 1792
+ },
+ {
+ "epoch": 24.907504363001745,
+ "grad_norm": 0.052474118769168854,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 4.035177230834961,
+ "step": 1793
+ },
+ {
+ "epoch": 24.921465968586386,
+ "grad_norm": 0.05411219224333763,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.9920859336853027,
+ "step": 1794
+ },
+ {
+ "epoch": 24.93542757417103,
+ "grad_norm": 0.0543992817401886,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 4.017391204833984,
+ "step": 1795
+ },
+ {
+ "epoch": 24.949389179755673,
+ "grad_norm": 0.05187678337097168,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 4.042266368865967,
+ "step": 1796
+ },
+ {
+ "epoch": 24.963350785340314,
+ "grad_norm": 0.053762417286634445,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 4.029013633728027,
+ "step": 1797
+ },
+ {
+ "epoch": 24.977312390924958,
+ "grad_norm": 0.051723796874284744,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 4.038886070251465,
+ "step": 1798
+ },
+ {
+ "epoch": 24.991273996509598,
+ "grad_norm": 0.05242295190691948,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 4.056063652038574,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.03998562693595886,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 2.534770965576172,
+ "step": 1800
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 7.68609738597335e+17,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-1800/training_args.bin b/runs/l2r90-ssl/checkpoint-1800/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..103048067ef8897d0386013ed89ce6764bad5521
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-1800/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f320fb44ea3a4a93198f1a6c2405447428148ce142cd94ee283a3c077a327c2d
+size 4792
diff --git a/runs/l2r90-ssl/checkpoint-2160/chat_template.jinja b/runs/l2r90-ssl/checkpoint-2160/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2160/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-ssl/checkpoint-2160/config.json b/runs/l2r90-ssl/checkpoint-2160/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2160/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-ssl/checkpoint-2160/generation_config.json b/runs/l2r90-ssl/checkpoint-2160/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2160/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-ssl/checkpoint-2160/model.safetensors b/runs/l2r90-ssl/checkpoint-2160/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..e6964efa658598a2e406208bc2f540c4c757c561
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2160/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ef43a748cb673c9c5cc7e7fa38247e660a28df13f3e34bbda36a9315052b58ab
+size 583356232
diff --git a/runs/l2r90-ssl/checkpoint-2160/optimizer.pt b/runs/l2r90-ssl/checkpoint-2160/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..b56923eedd5054998fc87c3f465c212bf721fd48
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2160/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:fd2649ee535b692aa4da4c82a069618c6cd44aa33bbc3c7daa8a391b337dd746
+size 1166827898
diff --git a/runs/l2r90-ssl/checkpoint-2160/rng_state_0.pth b/runs/l2r90-ssl/checkpoint-2160/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..f7b0f80bddd9d1575a028e4ed08ac5261f9a78cc
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2160/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:36272cc63a644540024f08ffb529f51f3af1b1c7c616e1e8762def8137bd7cc5
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-2160/rng_state_1.pth b/runs/l2r90-ssl/checkpoint-2160/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..7b62eb37975baa72a3c8fce7a6ede595f7ae575b
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2160/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f87b906c8d2b07e01c0df3c22ac835949a28469a6ca239354ae46dbeab434b1f
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-2160/scheduler.pt b/runs/l2r90-ssl/checkpoint-2160/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..35aefe7762df83336d8bfddab210c09c85e6b3aa
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2160/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a7d6a0a5c94f665a32f55d7b034b4fc829f9f44c468d851ff5b24b22d8388e63
+size 1064
diff --git a/runs/l2r90-ssl/checkpoint-2160/tokenizer.json b/runs/l2r90-ssl/checkpoint-2160/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2160/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-ssl/checkpoint-2160/tokenizer_config.json b/runs/l2r90-ssl/checkpoint-2160/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2160/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-2160/trainer_state.json b/runs/l2r90-ssl/checkpoint-2160/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..193a68b19b9f4c042971383d2b581a797a899bb6
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2160/trainer_state.json
@@ -0,0 +1,15378 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 30.0,
+ "eval_steps": 500,
+ "global_step": 2160,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.2588357925415039,
+ "learning_rate": 0.0,
+ "loss": 3.9982247352600098,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.25455471873283386,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.998683452606201,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.2555367052555084,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.9821574687957764,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.2383691370487213,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.9531917572021484,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2023179680109024,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.9148876667022705,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.17691479623317719,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.875638961791992,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.14372293651103973,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.836796283721924,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1074419692158699,
+ "learning_rate": 4.2e-05,
+ "loss": 3.8036489486694336,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.08204353600740433,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.774808406829834,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.0632256343960762,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.7503671646118164,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.047985248267650604,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.7345473766326904,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.03675895184278488,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.721525192260742,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.02908993512392044,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.713454246520996,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.02443614974617958,
+ "learning_rate": 7.8e-05,
+ "loss": 3.7078704833984375,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02109951712191105,
+ "learning_rate": 8.4e-05,
+ "loss": 3.7034401893615723,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.018616363406181335,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.7009854316711426,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.016399752348661423,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.697768211364746,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.014256286434829235,
+ "learning_rate": 0.000102,
+ "loss": 3.6974306106567383,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.013020666316151619,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.694406509399414,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.011879238300025463,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.691807270050049,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.010290366597473621,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.6937897205352783,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.009482803754508495,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.6902384757995605,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.008425266481935978,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.686815023422241,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.007676936220377684,
+ "learning_rate": 0.000138,
+ "loss": 3.685192823410034,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.007491032127290964,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.6839442253112793,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.007428477052599192,
+ "learning_rate": 0.00015,
+ "loss": 3.6783409118652344,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.006965892389416695,
+ "learning_rate": 0.000156,
+ "loss": 3.6813433170318604,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.008106694556772709,
+ "learning_rate": 0.000162,
+ "loss": 3.6741726398468018,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.0076780058443546295,
+ "learning_rate": 0.000168,
+ "loss": 3.67673921585083,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.007904314436018467,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.6724467277526855,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.008422375656664371,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.673013925552368,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00811714306473732,
+ "learning_rate": 0.000186,
+ "loss": 3.6698389053344727,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.009063836187124252,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.668990135192871,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.01878332532942295,
+ "learning_rate": 0.000198,
+ "loss": 3.6658849716186523,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.01812782697379589,
+ "learning_rate": 0.000204,
+ "loss": 3.6697864532470703,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.020587686449289322,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.662757158279419,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.01515690516680479,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.6558732986450195,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.02510508894920349,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.6495113372802734,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.020487826317548752,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.638803243637085,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.017446046695113182,
+ "learning_rate": 0.000234,
+ "loss": 3.633307456970215,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.030667584389448166,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.6277506351470947,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.023789944127202034,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.6189804077148438,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.02660694345831871,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.610522747039795,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.023814303800463676,
+ "learning_rate": 0.000258,
+ "loss": 3.615910291671753,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.021863561123609543,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.612008571624756,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03705216199159622,
+ "learning_rate": 0.00027,
+ "loss": 3.6066887378692627,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.024143919348716736,
+ "learning_rate": 0.000276,
+ "loss": 3.5987331867218018,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.03251276910305023,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.595367431640625,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06619632989168167,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.59018611907959,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.042714253067970276,
+ "learning_rate": 0.000294,
+ "loss": 3.58981990814209,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.018893254920840263,
+ "learning_rate": 0.0003,
+ "loss": 3.578472375869751,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0749685987830162,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.5828654766082764,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.12142857164144516,
+ "learning_rate": 0.000312,
+ "loss": 3.5840892791748047,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.06944665312767029,
+ "learning_rate": 0.000318,
+ "loss": 3.580961227416992,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.04545000568032265,
+ "learning_rate": 0.000324,
+ "loss": 3.5674569606781006,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.061073049902915955,
+ "learning_rate": 0.00033,
+ "loss": 3.569991111755371,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.04133498668670654,
+ "learning_rate": 0.000336,
+ "loss": 3.5657706260681152,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.055982090532779694,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.559979200363159,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.05259208381175995,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.558039426803589,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.029976001009345055,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.5491275787353516,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.04998327046632767,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.5406670570373535,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.03230690956115723,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.5391530990600586,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.06102178618311882,
+ "learning_rate": 0.000372,
+ "loss": 3.5311684608459473,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.031715717166662216,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.518899440765381,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.06841139495372772,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.514200210571289,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.10323675721883774,
+ "learning_rate": 0.00039,
+ "loss": 3.513646125793457,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.1448228508234024,
+ "learning_rate": 0.000396,
+ "loss": 3.5110087394714355,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.12129563093185425,
+ "learning_rate": 0.000402,
+ "loss": 3.501497507095337,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.030224842950701714,
+ "learning_rate": 0.000408,
+ "loss": 3.4870352745056152,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.07492101192474365,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.482234001159668,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.03942760080099106,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.4746525287628174,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.04967200756072998,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.604069232940674,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8671882152557373,
+ "eval_runtime": 42.6602,
+ "eval_samples_per_second": 57.243,
+ "eval_steps_per_second": 0.469,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.05385325849056244,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.462052822113037,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.04870929941534996,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.459104537963867,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.08701689541339874,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.4533209800720215,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.12219952791929245,
+ "learning_rate": 0.00045,
+ "loss": 3.4558796882629395,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.0823654904961586,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.443049192428589,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.12765446305274963,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.4373486042022705,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.14215047657489777,
+ "learning_rate": 0.000468,
+ "loss": 3.449429988861084,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.03649887070059776,
+ "learning_rate": 0.000474,
+ "loss": 3.42877197265625,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09551949054002762,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.4352874755859375,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.030483486130833626,
+ "learning_rate": 0.000486,
+ "loss": 3.4125874042510986,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.0745147094130516,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.4153835773468018,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.06620422005653381,
+ "learning_rate": 0.000498,
+ "loss": 3.4053854942321777,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.06098725646734238,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.404346466064453,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.047781601548194885,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.3991761207580566,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.04409855231642723,
+ "learning_rate": 0.000516,
+ "loss": 3.396144390106201,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.04355994611978531,
+ "learning_rate": 0.000522,
+ "loss": 3.3928911685943604,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.0670008510351181,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.3831443786621094,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.0812622532248497,
+ "learning_rate": 0.000534,
+ "loss": 3.382413148880005,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.0632157102227211,
+ "learning_rate": 0.00054,
+ "loss": 3.3784892559051514,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.028686825186014175,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.3673207759857178,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.034818850457668304,
+ "learning_rate": 0.000552,
+ "loss": 3.3648972511291504,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.04509204626083374,
+ "learning_rate": 0.000558,
+ "loss": 3.35561203956604,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.054211195558309555,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.3612167835235596,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.03752049803733826,
+ "learning_rate": 0.00057,
+ "loss": 3.357103109359741,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.03620241954922676,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.3469700813293457,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.02454194985330105,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.3429036140441895,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.041030462831258774,
+ "learning_rate": 0.000588,
+ "loss": 3.338247299194336,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.03266424685716629,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.332688331604004,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.04730091616511345,
+ "learning_rate": 0.0006,
+ "loss": 3.335902452468872,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.065738245844841,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.3289852142333984,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.09062394499778748,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.3278567790985107,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.1482805460691452,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.332685947418213,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.20066532492637634,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.363180637359619,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.11500677466392517,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.3308753967285156,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.11754138022661209,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.3383703231811523,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.09672239422798157,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.3267624378204346,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.07190799713134766,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.317737579345703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.0815449133515358,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.3198161125183105,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09612327069044113,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.315814971923828,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.14795289933681488,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.317605495452881,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06667589396238327,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.3127357959747314,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.058537885546684265,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.319807767868042,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.06428155303001404,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.297541618347168,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.03881775587797165,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.297539710998535,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.061801470816135406,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.290297508239746,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.07642851769924164,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.284428596496582,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08123870939016342,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.2820234298706055,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07174684852361679,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.2697296142578125,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.04030010476708412,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.264328956604004,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.05662153288722038,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.2699337005615234,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.03761111572384834,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.255704879760742,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.043027132749557495,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.2579030990600586,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.044409338384866714,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.248547077178955,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.0239730142056942,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.249307632446289,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03628705441951752,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.2451741695404053,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.029564054682850838,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.238445997238159,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.032036975026130676,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.2320892810821533,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.030085794627666473,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.231468915939331,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.025232303887605667,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.232199192047119,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.027053389698266983,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.210745334625244,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.028464114293456078,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.210076332092285,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04616783559322357,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.2103490829467773,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.039081938564777374,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.202662467956543,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.03428936004638672,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.1997475624084473,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.04351917654275894,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.19844913482666,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.0840945914387703,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.2013347148895264,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.36558300256729126,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.26310396194458,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.09432770311832428,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.2881736755371094,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.061970267444849014,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.2840280532836914,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.05112256482243538,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.2632040977478027,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06797203421592712,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.254606246948242,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.051026541739702225,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.429918050765991,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.8117228746414185,
+ "eval_runtime": 42.8572,
+ "eval_samples_per_second": 56.98,
+ "eval_steps_per_second": 0.467,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.0608622208237648,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.2348179817199707,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.05581629276275635,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.22896146774292,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.09162136912345886,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.2343664169311523,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.13750384747982025,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.2507717609405518,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.0872517079114914,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.2440056800842285,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.07161762565374374,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.229494571685791,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04082738608121872,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.2183239459991455,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.04972497373819351,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.218075752258301,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.049254074692726135,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.213379383087158,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.056671254336833954,
+ "learning_rate": 0.000599663984012462,
+ "loss": 3.209568977355957,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.05541909113526344,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.207760810852051,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.03819117695093155,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 3.199732780456543,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.0529443584382534,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.1945419311523438,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.043010979890823364,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.185936212539673,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.07311078161001205,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.189894914627075,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06110258400440216,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 3.1911773681640625,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.06880149245262146,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 3.187437057495117,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.051326874643564224,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 3.1826558113098145,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.03363822400569916,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 3.175527572631836,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.10268989205360413,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 3.172367572784424,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.09723010659217834,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 3.1873984336853027,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.04129317030310631,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 3.165799617767334,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.05292229354381561,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 3.171780824661255,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.03873143345117569,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 3.1680874824523926,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.047018859535455704,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 3.161402940750122,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.03942955285310745,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 3.1530532836914062,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.039801858365535736,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 3.1465344429016113,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.03795155510306358,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 3.1364030838012695,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.0342850424349308,
+ "learning_rate": 0.000599379982511273,
+ "loss": 3.1391525268554688,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.03243280574679375,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 3.1331253051757812,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.029679352417588234,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 3.1346120834350586,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.030230293050408363,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 3.132615804672241,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.02845083735883236,
+ "learning_rate": 0.000599309205504528,
+ "loss": 3.1242547035217285,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.03208369016647339,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 3.119110584259033,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.040268607437610626,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 3.113833427429199,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.06951889395713806,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 3.1186914443969727,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.08879823237657547,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 3.1140987873077393,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07374681532382965,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 3.1147828102111816,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.038190409541130066,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 3.1161253452301025,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.03297388181090355,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 3.1006083488464355,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.05905177444219589,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 3.106564521789551,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.037567321211099625,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 3.0884456634521484,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.02725658379495144,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 3.09000825881958,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.04272424429655075,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 3.090528964996338,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.027587950229644775,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 3.085131883621216,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.03282274678349495,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 3.0880112648010254,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.04064026474952698,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 3.079259157180786,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.023374097421765327,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 3.0747299194335938,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.025003870949149132,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 3.0721259117126465,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.030445197597146034,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 3.0733323097229004,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.025935636833310127,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 3.070709228515625,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.021337421610951424,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 3.0627551078796387,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.040481697767972946,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 3.0529394149780273,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.047608256340026855,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 3.041020631790161,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.03356955572962761,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 3.0532307624816895,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.02101682312786579,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 3.0492985248565674,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.030098581686615944,
+ "learning_rate": 0.000598804365316586,
+ "loss": 3.0527801513671875,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.040166422724723816,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 3.0505762100219727,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.051579318940639496,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 3.0422110557556152,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.06485334783792496,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 3.0435101985931396,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.055590901523828506,
+ "learning_rate": 0.000598706872384039,
+ "loss": 3.0448739528656006,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.03183475881814957,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 3.0416858196258545,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.03995629772543907,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 3.0255980491638184,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.03907827287912369,
+ "learning_rate": 0.000598631251036868,
+ "loss": 3.032339096069336,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.027681387960910797,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 3.025439739227295,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.03470184653997421,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 3.02883243560791,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.031407181173563004,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 3.0289883613586426,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.0371522456407547,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 3.013608932495117,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.03651345893740654,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 3.0120253562927246,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.06004595756530762,
+ "learning_rate": 0.000598473577921366,
+ "loss": 3.017958402633667,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.09979518502950668,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 3.023041248321533,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.10138462483882904,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.275729179382324,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7571707367897034,
+ "eval_runtime": 42.8283,
+ "eval_samples_per_second": 57.018,
+ "eval_steps_per_second": 0.467,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.04982660710811615,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 3.016209602355957,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.06532539427280426,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 3.015043020248413,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.04369623214006424,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 3.0096778869628906,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.05180002748966217,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 3.0121543407440186,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.0440233014523983,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.9937305450439453,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.05645671859383583,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 3.0023865699768066,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.03990716487169266,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 3.0003232955932617,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.03818608075380325,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.994137763977051,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.035388972610235214,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.993170738220215,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.03162284567952156,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.9922614097595215,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.029497336596250534,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.9852986335754395,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.02586926706135273,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.9798672199249268,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.027355050668120384,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.985158920288086,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.0341978520154953,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.982581615447998,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.030727429315447807,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.9701411724090576,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.04429870471358299,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.9789254665374756,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.05775132030248642,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.9758143424987793,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.05635962635278702,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.981480121612549,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.03639192506670952,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.968752861022949,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.028748342767357826,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.9584972858428955,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.035653006285429,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.967545747756958,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.03697885945439339,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.9578099250793457,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.03518654406070709,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.9680328369140625,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.034762222319841385,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.9670262336730957,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03813016042113304,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.957522392272949,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.04372904822230339,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.9540514945983887,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.06646446138620377,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.951028823852539,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.07831092923879623,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.95898699760437,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.05779588967561722,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.965884208679199,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.04414296895265579,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.9510560035705566,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.04797815904021263,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.955803394317627,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.04482145980000496,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.954059362411499,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.037104446440935135,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.94792103767395,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.03616701066493988,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.9500932693481445,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.03882797807455063,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.9558305740356445,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.04356204345822334,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.9237029552459717,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.04535257816314697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.925853729248047,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.04519587755203247,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.924984931945801,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.055922407656908035,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.932561159133911,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.06330762803554535,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.9388322830200195,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.0517115443944931,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.9272942543029785,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.057369694113731384,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.9201431274414062,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.05687880516052246,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.922886371612549,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.055525410920381546,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.9211201667785645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04503652825951576,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.9164583683013916,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.03385302051901817,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.9145395755767822,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03564384952187538,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.916003704071045,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.04188712686300278,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.915839433670044,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03657370060682297,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.903120279312134,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.03126626834273338,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.895824432373047,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.04196290299296379,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.9093422889709473,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.1706087589263916,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.9183316230773926,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.05810421332716942,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.911245107650757,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.061330314725637436,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.9159231185913086,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.06062281131744385,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.9211535453796387,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.05590800568461418,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.9140496253967285,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.04496239125728607,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.9040920734405518,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.03835335746407509,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.9026548862457275,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.05137286335229874,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.889634132385254,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.04091101884841919,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.886838674545288,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.04076292738318443,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.892918109893799,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.045437734574079514,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.893305778503418,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.041114673018455505,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.88517165184021,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.03600773215293884,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.8779795169830322,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.04311220720410347,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.87247896194458,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.04818756878376007,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.889103412628174,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.053460877388715744,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.8899683952331543,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.06763825565576553,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.8823537826538086,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.07304801046848297,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.877340316772461,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.061337269842624664,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.8820111751556396,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.04660911113023758,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.868886709213257,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.053929202258586884,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.159337043762207,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.721762478351593,
+ "eval_runtime": 42.7977,
+ "eval_samples_per_second": 57.059,
+ "eval_steps_per_second": 0.467,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.04259369149804115,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.859579086303711,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.044425517320632935,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.8681139945983887,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.04301869496703148,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.853911876678467,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.04574296995997429,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.8724427223205566,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.050247322767972946,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.8528172969818115,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.046683840453624725,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.853405714035034,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.05966169387102127,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.8605127334594727,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.08935574442148209,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.862339973449707,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.08019179105758667,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.8655967712402344,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.04753892868757248,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.8657939434051514,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06868117302656174,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.861976146697998,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.07040470093488693,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.8562135696411133,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.05683179199695587,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.8576629161834717,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.04918001592159271,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.861943244934082,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.04921915754675865,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.8495378494262695,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.037651967257261276,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.8428618907928467,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.03522057458758354,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.8372769355773926,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.03111186809837818,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.8377151489257812,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.03906773030757904,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.846196174621582,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.02546556107699871,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.8288016319274902,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.03119821846485138,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.8334617614746094,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.030583377927541733,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.8209362030029297,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.029693076387047768,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.8267486095428467,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.028978517279028893,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.819244861602783,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.027871523052453995,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.8289647102355957,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.029287064447999,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.8235747814178467,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.024990517646074295,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.819016218185425,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.030980689451098442,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.828155040740967,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03721801936626434,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.813418388366699,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03738148882985115,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.808328866958618,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.03580411896109581,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.812070846557617,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.039036594331264496,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.807159423828125,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.044666968286037445,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.799992561340332,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.055589817464351654,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.804142951965332,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.06139020994305611,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.8182730674743652,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.07769318670034409,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.8166518211364746,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.06521332263946533,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.8142249584198,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.09083619713783264,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.8166933059692383,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.09737467020750046,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.8111555576324463,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.07148188352584839,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.8168413639068604,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.055516090244054794,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.8183553218841553,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.05338391661643982,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.8127028942108154,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.040785472840070724,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.8124775886535645,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.037135496735572815,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.8065357208251953,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.032455578446388245,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.798459053039551,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.03465662896633148,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.798649549484253,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.032026033848524094,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.7906997203826904,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.026019956916570663,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.791969060897827,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.030298635363578796,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.784593105316162,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.028147678822278976,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.782454013824463,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.023933131247758865,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.787851333618164,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.026789527386426926,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.788067579269409,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.024554884061217308,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.777033567428589,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.028679484501481056,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.786342144012451,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.026410218328237534,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.7790610790252686,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.027413051575422287,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.778944969177246,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.02908242493867874,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.763850212097168,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.029632994905114174,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.7672982215881348,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.027458716183900833,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.7702178955078125,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.03263958916068077,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.7581803798675537,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04427146911621094,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.768141508102417,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.07142394036054611,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.75819730758667,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.08833584934473038,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.782716751098633,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.08168833702802658,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.7763240337371826,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07253874838352203,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.7717690467834473,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.08186902105808258,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.767303466796875,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.05854064226150513,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.761544704437256,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.04602917283773422,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.7669177055358887,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.03920970857143402,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.766085624694824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.04551195353269577,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.7623450756073,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.040861159563064575,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.7471840381622314,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.026907291263341904,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 2.0710244178771973,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6934371590614319,
+ "eval_runtime": 42.6015,
+ "eval_samples_per_second": 57.322,
+ "eval_steps_per_second": 0.469,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.03801899403333664,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.739086627960205,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.03707238286733627,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.741638422012329,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.0387551411986351,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.740777015686035,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.028010359033942223,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.7353010177612305,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.026474518701434135,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.73842453956604,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.02920856513082981,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.736295461654663,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.026074763387441635,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.7295751571655273,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.026862047612667084,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.723040819168091,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.026300370693206787,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.7199485301971436,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.026030950248241425,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.7276217937469482,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.028342435136437416,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.7193045616149902,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.032135091722011566,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.7143118381500244,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.03068603202700615,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.7191050052642822,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.035828713327646255,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.718085765838623,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.03343002125620842,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.707131862640381,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.035051796585321426,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.7091879844665527,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.03395913913846016,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6980862617492676,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.034198079258203506,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.699704647064209,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.0322839580476284,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.702791213989258,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.041531339287757874,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.7018051147460938,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.040806882083415985,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.7096598148345947,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.04156940430402756,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6956775188446045,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.05608797073364258,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.70127272605896,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0729481428861618,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.708817720413208,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.0905836969614029,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.718085289001465,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.08345075696706772,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.7235910892486572,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.08599896728992462,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.7178783416748047,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.07301691174507141,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.714029312133789,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.08231015503406525,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.7196826934814453,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.08286412060260773,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.7176082134246826,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.08590883016586304,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.7242164611816406,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.07531397044658661,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.7225024700164795,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.04711190238595009,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.695887327194214,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.05076327919960022,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.7029781341552734,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.04495669901371002,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.7057900428771973,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.04312257468700409,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.697537899017334,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.043580975383520126,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.67315673828125,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.03360886126756668,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.6883413791656494,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.030360296368598938,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.700927257537842,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.0310810673981905,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.6917803287506104,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.02703232690691948,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.6667580604553223,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.02756045199930668,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.6702821254730225,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.02611083723604679,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.669790744781494,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.02615324780344963,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.6672821044921875,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.026000119745731354,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.6671881675720215,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.02882794849574566,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.67586088180542,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03327213600277901,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.6618690490722656,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.032920945435762405,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.665444850921631,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.02812962979078293,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.6680707931518555,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.025886602699756622,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.652898073196411,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.025222789496183395,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.6560399532318115,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.023131893947720528,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.66127347946167,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.02513662725687027,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.6463732719421387,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.027243830263614655,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.6507630348205566,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.025437965989112854,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.6485800743103027,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.02770340070128441,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.657241106033325,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.03594067320227623,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.6281137466430664,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04486451297998428,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.6340909004211426,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.05102410912513733,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.6463394165039062,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04514491558074951,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.636554002761841,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04469453915953636,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.6453680992126465,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.05421523004770279,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.646010398864746,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04779462888836861,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.6315622329711914,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.06697724014520645,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.6416730880737305,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.07520416378974915,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.641632318496704,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.055692676454782486,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.6289877891540527,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.04539895057678223,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.6363251209259033,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.04829065501689911,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.641608476638794,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.041314542293548584,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.6269643306732178,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.04100198298692703,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.6243174076080322,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.04105962812900543,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.626585006713867,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.03172122687101364,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9550899267196655,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6614077091217041,
+ "eval_runtime": 42.9009,
+ "eval_samples_per_second": 56.922,
+ "eval_steps_per_second": 0.466,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.03258531913161278,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.612126588821411,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.040250636637210846,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.6025328636169434,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.034234337508678436,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.6085972785949707,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.03635786101222038,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.6126222610473633,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.039404451847076416,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.6035327911376953,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.038504838943481445,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5985798835754395,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.03744236007332802,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5873115062713623,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.034184787422418594,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.5899314880371094,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.03766429424285889,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.602734327316284,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.03741588443517685,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.592207193374634,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.03187853842973709,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5850372314453125,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.038447409868240356,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.592654228210449,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.04046263173222542,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.5881710052490234,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.03872602432966232,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5881471633911133,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.03960775211453438,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5826945304870605,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.033132150769233704,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.584038734436035,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.029504230245947838,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.584357261657715,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.03665253147482872,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5677168369293213,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.037208545953035355,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.560795783996582,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.038244519382715225,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.591458559036255,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.043307431042194366,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.5642588138580322,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.055119726806879044,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.57377290725708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.053323689848184586,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5767312049865723,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.0502864345908165,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.562190055847168,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.05568842589855194,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.565556049346924,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.061149828135967255,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5883288383483887,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.059933699667453766,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.5710184574127197,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.06041933596134186,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.572603225708008,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.05482695251703262,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.5667271614074707,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.04712968319654465,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.554041624069214,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.04842432215809822,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.5660672187805176,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.04273047298192978,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.554839849472046,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.05356749892234802,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.5521483421325684,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.06062676012516022,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.5551886558532715,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.05431462824344635,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5654542446136475,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.0505850687623024,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5648036003112793,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.04696856811642647,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.5559377670288086,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.042585279792547226,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.5660319328308105,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.04026484116911888,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.551206588745117,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03382421284914017,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.556361198425293,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.031177924945950508,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.5413894653320312,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.028217196464538574,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.5600497722625732,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.027062954381108284,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5447568893432617,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.03001212701201439,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.540614604949951,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.028869854286313057,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.544583797454834,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.02869918756186962,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.5471596717834473,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03359716013073921,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5416982173919678,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.04000182822346687,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.534921407699585,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.043030042201280594,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5368564128875732,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.04329848289489746,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.539764881134033,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.03740967437624931,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.5246665477752686,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.03733935207128525,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.529362916946411,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.039609868079423904,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.542544364929199,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.03937261551618576,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5366649627685547,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.03261927142739296,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.5250535011291504,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.03904294595122337,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.520721435546875,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.03913239762187004,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.5122013092041016,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.03610851988196373,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.5212767124176025,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.03944548964500427,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.537186622619629,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04762452840805054,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.5204620361328125,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04427091404795647,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5286149978637695,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04366641491651535,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.5398361682891846,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.04644155502319336,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.519331932067871,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.04372382536530495,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.531022787094116,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.04404324293136597,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.521758794784546,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.042115990072488785,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.5146074295043945,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.03796013817191124,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.5040054321289062,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.04327977076172829,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.5209312438964844,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04150385782122612,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.515871286392212,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.04465346783399582,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.5247764587402344,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.05690346285700798,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.501911163330078,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04907930642366409,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8968371152877808,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6358620524406433,
+ "eval_runtime": 42.9991,
+ "eval_samples_per_second": 56.792,
+ "eval_steps_per_second": 0.465,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.03755255416035652,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.486576795578003,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.040234558284282684,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.501087188720703,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.041416723281145096,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4942233562469482,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.039655014872550964,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.503453254699707,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.041707683354616165,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.490978717803955,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.04545537009835243,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4757118225097656,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.04328389838337898,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.504516839981079,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.043312299996614456,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.4920215606689453,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.046466976404190063,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.494760513305664,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.041762981563806534,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4875552654266357,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.03713781014084816,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.495652675628662,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.040603458881378174,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4821407794952393,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.037969645112752914,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.484981060028076,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.03652074560523033,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4736409187316895,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.03783264383673668,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.483346700668335,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.03404324874281883,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.48856782913208,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.03294725716114044,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.483706474304199,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.0347001850605011,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.474888324737549,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.0359107106924057,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.478990077972412,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.03738103806972504,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.4709420204162598,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.03722188249230385,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.465726852416992,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.034907322376966476,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.482231855392456,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.035695310682058334,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4785828590393066,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.04009179025888443,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.4795875549316406,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.038996778428554535,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.478696823120117,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04082407429814339,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.465200901031494,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04856136813759804,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4605681896209717,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.05671004578471184,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.47277569770813,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.0481606163084507,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.476151466369629,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.037689510732889175,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.466142177581787,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.05005146935582161,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.4798035621643066,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.04745829105377197,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.46163272857666,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.04638158157467842,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4728071689605713,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.04530330374836922,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.4621877670288086,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.03971488028764725,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.474224090576172,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.04521876201033592,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.46122670173645,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03923020884394646,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.4603679180145264,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03986494988203049,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.4589664936065674,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.04280020296573639,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.4624319076538086,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.03930043429136276,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.47080135345459,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.0314876064658165,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.458728790283203,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.03189660236239433,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4467148780822754,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03264275938272476,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4613699913024902,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03021484799683094,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.4535279273986816,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.030740290880203247,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.4621875286102295,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03096810355782509,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.458500385284424,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03105205111205578,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.461958646774292,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.03330068290233612,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.449777364730835,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.032831475138664246,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.4616191387176514,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.0387776717543602,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.445256471633911,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.041803423315286636,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.4545960426330566,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.04148292914032936,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.433408737182617,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.04648579657077789,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.449134349822998,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.04186409339308739,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.4449565410614014,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.038569554686546326,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.443675994873047,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03908300772309303,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4626526832580566,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.038897737860679626,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.438591957092285,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.0360480472445488,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.436077117919922,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0429830439388752,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4384970664978027,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.04465422406792641,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.458029270172119,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.037096526473760605,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.436816930770874,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.047823466360569,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.44439959526062,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.05929310992360115,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.4599215984344482,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.053512319922447205,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4387903213500977,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.042625587433576584,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4359588623046875,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.056042976677417755,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.4441447257995605,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05813189223408699,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4399144649505615,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.06313478201627731,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.453447103500366,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.05281798169016838,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.4428963661193848,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.03991910442709923,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4359219074249268,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.047981832176446915,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4506607055664062,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.03646523132920265,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8225739002227783,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6196661591529846,
+ "eval_runtime": 43.8106,
+ "eval_samples_per_second": 55.74,
+ "eval_steps_per_second": 0.457,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.03673036769032478,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.4227209091186523,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.044547371566295624,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.4073169231414795,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.04089285805821419,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4279446601867676,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.03538106009364128,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.4234392642974854,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.03152978792786598,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4214606285095215,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.03324202075600624,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.4085447788238525,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.03352442383766174,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.417776107788086,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.03582802042365074,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.411144256591797,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.0351114459335804,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.405872344970703,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.038898173719644547,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4029338359832764,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.03926656022667885,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4100089073181152,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.04285634309053421,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.410360336303711,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.04179581627249718,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4093637466430664,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.03356701880693436,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4118056297302246,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.038140635937452316,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.4154677391052246,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.04468780383467674,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416846513748169,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.062421441078186035,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4157638549804688,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06317313015460968,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.418593406677246,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.049498945474624634,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4204845428466797,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.046322405338287354,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.403489589691162,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.053081002086400986,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.4148740768432617,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05222945287823677,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.416757583618164,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04893574118614197,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4241795539855957,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04951956123113632,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.401386260986328,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04019852355122566,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.4168622493743896,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.03625848889350891,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4026846885681152,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.03464345261454582,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4021215438842773,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.03750818222761154,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4113571643829346,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.0360100194811821,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.3977622985839844,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.03464623540639877,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.413273334503174,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.03193575143814087,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.423659324645996,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.02710454724729061,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.396085262298584,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.02830432914197445,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.420255422592163,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.029717663303017616,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.3958120346069336,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.030005933716893196,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3908543586730957,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.029054716229438782,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.3873138427734375,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.031556472182273865,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3892760276794434,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03518872708082199,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4113478660583496,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.03471249341964722,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.384857654571533,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03223155066370964,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3840765953063965,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.030012644827365875,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.388247489929199,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.03162285313010216,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.3987512588500977,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.03700442612171173,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.3993430137634277,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.03710822016000748,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.387359380722046,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.028411181643605232,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3925249576568604,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031246855854988098,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.38845157623291,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.034664880484342575,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.392436981201172,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.03354701027274132,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.390890121459961,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.03527764603495598,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.3954997062683105,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03325049206614494,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3847177028656006,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.028021875768899918,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.385910987854004,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.02964954823255539,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3787484169006348,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.03068520873785019,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.400784969329834,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03416873514652252,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.391033411026001,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.03534875437617302,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.3796474933624268,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.03732474148273468,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3831934928894043,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.03648678958415985,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3785600662231445,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03801294043660164,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.387587070465088,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.04186025261878967,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.3893089294433594,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.04416101053357124,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3852195739746094,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.04137178510427475,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.391308307647705,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.05010323226451874,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.3906192779541016,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.049587685614824295,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.3846192359924316,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.04125971347093582,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.384507179260254,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.039676666259765625,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.38167142868042,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.038775283843278885,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.3789167404174805,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.03634221479296684,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.383060932159424,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03464275971055031,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.388552665710449,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.031778786331415176,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.3903298377990723,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.027904309332370758,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.377493381500244,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.028964119032025337,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.381959915161133,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.0233962032943964,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7810559272766113,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6070130467414856,
+ "eval_runtime": 50.871,
+ "eval_samples_per_second": 48.004,
+ "eval_steps_per_second": 0.393,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.030209315940737724,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.3569841384887695,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.04112662002444267,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.376680374145508,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.055858466774225235,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.3621644973754883,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.06746986508369446,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3607563972473145,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.06736797094345093,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3548707962036133,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.0562627948820591,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.364654541015625,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.051833927631378174,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.3748884201049805,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.04824110120534897,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.3502087593078613,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.05518793687224388,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3687515258789062,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.06564056873321533,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.367131233215332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.0605909489095211,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.3589868545532227,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.05744732916355133,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.370117664337158,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.06430432945489883,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.3703484535217285,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.06186915934085846,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3720293045043945,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.05977998673915863,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.3512699604034424,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.06413097679615021,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.377257823944092,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.053935080766677856,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.350717544555664,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.049110863357782364,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.3688902854919434,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.04313219338655472,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3374459743499756,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.039556123316287994,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3482556343078613,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.03796547278761864,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3547937870025635,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.03646104037761688,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.344686269760132,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.03355357423424721,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.3588008880615234,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.03004707582294941,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.352043628692627,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.030186176300048828,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3500709533691406,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.029748762026429176,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.3523058891296387,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.03238696977496147,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3635129928588867,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.032599564641714096,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.346269130706787,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.028626957908272743,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.357151746749878,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.02909751422703266,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3557677268981934,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.029951099306344986,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.34753680229187,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.029768310487270355,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3408985137939453,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.03005973994731903,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.362542152404785,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03360462561249733,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.352529287338257,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.034002047032117844,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.3334460258483887,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.03292316198348999,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3446991443634033,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.0337117575109005,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.350605010986328,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03428212180733681,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.357408046722412,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.03169601410627365,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.349088191986084,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.03389411047101021,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.361576795578003,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.036497049033641815,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.35420560836792,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03574785590171814,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3678526878356934,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.030757993459701538,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3520517349243164,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.0289121363312006,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.337864637374878,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.028012607246637344,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.353999614715576,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.027558492496609688,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.342132568359375,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.02871510572731495,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3425655364990234,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03070886805653572,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3401665687561035,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.03056657873094082,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.345277786254883,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.030907969921827316,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.346494674682617,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0318334735929966,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.3390543460845947,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.029700733721256256,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.3398728370666504,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.029687359929084778,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.3474364280700684,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.027234600856900215,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3457484245300293,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.02823043428361416,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.337700366973877,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03002646192908287,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.335954189300537,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.027246717363595963,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.330207347869873,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.029703574255108833,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.328280448913574,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.028796236962080002,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3448386192321777,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.027944380417466164,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.320234537124634,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.030626650899648666,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.334698438644409,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.029164662584662437,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3450326919555664,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03034045360982418,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.3420588970184326,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.02856968529522419,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3396072387695312,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.02892562374472618,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.340027332305908,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.02938242256641388,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.33550763130188,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.029261089861392975,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3246421813964844,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.030340535566210747,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.3465514183044434,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.033335570245981216,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.3467206954956055,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.03452218323945999,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.343702793121338,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.03326255828142166,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.337631940841675,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025433633476495743,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7402067184448242,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03175730258226395,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.62623405456543,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.04780422896146774,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.623239517211914,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.0504758283495903,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.6369194984436035,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.04481920227408409,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.57640266418457,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0440552793443203,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.630992889404297,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0503336526453495,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.627910137176514,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.053885314613580704,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.637897491455078,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.05481956899166107,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.610715866088867,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.050510745495557785,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.639884948730469,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.04579266160726547,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.590188980102539,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.04817410930991173,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.636295318603516,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.058862220495939255,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.614317417144775,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.06328638643026352,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.655928611755371,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.058752477169036865,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.652004241943359,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.06049015372991562,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.6513800621032715,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.058941151946783066,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.657688140869141,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.05542639642953873,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.637376308441162,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06024057790637016,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.63592529296875,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.06040216609835625,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.6275787353515625,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.059600673615932465,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.652995586395264,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.05379730463027954,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.597358226776123,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05355260521173477,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.6106367111206055,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.06649646162986755,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.654443740844727,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.058309994637966156,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.6869659423828125,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.05403570458292961,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.654750823974609,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04678157716989517,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.653955459594727,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04297948628664017,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.644915580749512,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.04235142841935158,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.662282943725586,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.0405130498111248,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.64569091796875,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.03351007401943207,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.653165817260742,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.03182631731033325,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.606291770935059,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.031552448868751526,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.642643928527832,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027941036969423294,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.623664379119873,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.027744563296437263,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.605886459350586,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02962736040353775,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.606762886047363,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.027251560240983963,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.620492935180664,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.027894044294953346,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.613903999328613,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.02778882533311844,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.634457588195801,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.028082096949219704,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.610980033874512,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.026210976764559746,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.612090110778809,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.026537755504250526,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.581171035766602,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.028709523379802704,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.628799915313721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.028233813121914864,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.586182117462158,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.02966046892106533,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.62320613861084,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.03198636695742607,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.6018524169921875,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.033404890447854996,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.610121726989746,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.029217196628451347,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.601312160491943,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.029335757717490196,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.620820999145508,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.031007781624794006,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.626503944396973,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.028453458100557327,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.589323997497559,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02526312880218029,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.598350524902344,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.027317512780427933,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.62388277053833,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02710605598986149,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.606523513793945,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.027491388842463493,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.586644172668457,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.027158567681908607,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.606039047241211,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.0298276599496603,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.593776702880859,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.026622479781508446,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.59510612487793,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.028238043189048767,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.5899505615234375,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.029481008648872375,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.6092071533203125,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.030138323083519936,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.600069046020508,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.02951825223863125,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.600125789642334,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.029242202639579773,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.590723037719727,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.02941613830626011,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.623342514038086,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.029764151200652122,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.631185531616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.031503792852163315,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.605094909667969,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.03184918314218521,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.592767715454102,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.03223618119955063,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.5930585861206055,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.0340263694524765,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.609347343444824,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.03245621174573898,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.598440170288086,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.03204867243766785,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.619702339172363,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.030866602435708046,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.609165191650391,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.01960863545536995,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8655829429626465,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.593852162361145,
+ "eval_runtime": 60.0971,
+ "eval_samples_per_second": 40.634,
+ "eval_steps_per_second": 0.649,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03552858531475067,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.539240837097168,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.0536109134554863,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.52609395980835,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.061411239206790924,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.536109447479248,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.06352879852056503,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.546850204467773,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.05683520436286926,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.563828468322754,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.05522637069225311,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.550929069519043,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.058990392833948135,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.554411888122559,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.057443976402282715,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.545921325683594,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.05383576825261116,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.5481414794921875,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04797440022230148,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.55556583404541,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04923664405941963,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.545513153076172,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.056191347539424896,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.536705017089844,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.06088989973068237,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.570624351501465,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.04802761599421501,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.526930809020996,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.04018320515751839,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.562432289123535,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.04280981421470642,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.559737205505371,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.04258878529071808,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.561749458312988,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.04411563277244568,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.562010765075684,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.041077401489019394,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.573984622955322,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.04125386103987694,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.550564289093018,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.041974026709795,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.552739143371582,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03689907118678093,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.540107727050781,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03586242347955704,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.554999351501465,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.036602675914764404,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.546337127685547,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.03707926720380783,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.543377876281738,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.03322125971317291,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.574325084686279,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03359243646264076,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.5613603591918945,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.0314965583384037,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.530994415283203,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.031806789338588715,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.56138801574707,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.034988485276699066,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.553149700164795,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.037627559155225754,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.555401802062988,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03723530471324921,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.52452278137207,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.036220453679561615,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.54799747467041,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03228041157126427,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.574814319610596,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.030872676521539688,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.539666652679443,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.033935241401195526,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.579252243041992,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.034338828176259995,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.538324356079102,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.03441951423883438,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.540358543395996,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03328141197562218,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.520340919494629,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03224223107099533,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.5646209716796875,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.02999890223145485,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.504915714263916,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.03014362044632435,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.516615867614746,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.029469173401594162,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.560475826263428,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.027887821197509766,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.586333751678467,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.029970254749059677,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.530396461486816,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.031648170202970505,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.554404258728027,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.0325581431388855,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.544102668762207,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.03555625304579735,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.543701171875,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.033053912222385406,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.508375644683838,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.03059076890349388,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.545750617980957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.029640182852745056,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.54715633392334,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030139951035380363,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.533708572387695,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029753828421235085,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.543603897094727,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.029481034725904465,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.534200668334961,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.028991524130105972,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.5683674812316895,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.028145765885710716,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.543053150177002,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.02999844029545784,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.5482354164123535,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03134117275476456,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.5360107421875,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.029277140274643898,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.5729827880859375,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028516866266727448,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.543548583984375,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028387080878019333,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.552616596221924,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.028977863490581512,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.5456695556640625,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.028010815382003784,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.534921646118164,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.030222605913877487,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.561526298522949,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.03140007704496384,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.500124931335449,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.029269184917211533,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.536653995513916,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.02935507707297802,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.521398544311523,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02801917865872383,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.527442455291748,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.027995768934488297,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.556139945983887,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.028441084548830986,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.555442810058594,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.029100807383656502,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.532083511352539,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020385215058922768,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.8488006591796875,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5902590751647949,
+ "eval_runtime": 60.1424,
+ "eval_samples_per_second": 40.604,
+ "eval_steps_per_second": 0.648,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03513244539499283,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.486500263214111,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.052679311484098434,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.489691734313965,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.06171489134430885,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.471555709838867,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.06164781004190445,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.486891746520996,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.06410388648509979,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.494929313659668,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06399523466825485,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.5334320068359375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.0636502280831337,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.5039381980896,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.06139931082725525,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.512765884399414,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.07460282742977142,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.477229118347168,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.075187548995018,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.484212398529053,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.07584068924188614,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.495598793029785,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.07192158699035645,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.507598400115967,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.05366869643330574,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.514532089233398,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.06397061794996262,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.512920379638672,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.06391509622335434,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.465702056884766,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.059490568935871124,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.5351881980896,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.057501111179590225,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.510270595550537,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.056947533041238785,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.509241104125977,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.050393324345350266,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.50428581237793,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.057608287781476974,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.474919319152832,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.054470717906951904,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.521528244018555,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.054405540227890015,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.510211944580078,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.051328495144844055,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.480720520019531,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.045930180698633194,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.503024101257324,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.04298274591565132,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.49046516418457,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.0406254380941391,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.544064521789551,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.03557496517896652,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.496966361999512,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.0364493802189827,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.5047502517700195,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.034521300345659256,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.485365867614746,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.03305799141526222,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.484884738922119,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.033053625375032425,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.478784561157227,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03426017984747887,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.482428073883057,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.03638783469796181,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.433219909667969,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.03597250208258629,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.522584915161133,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.03086097538471222,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.500803470611572,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.029642771929502487,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.444099426269531,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.03190716728568077,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.493034362792969,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.0312873050570488,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.487887382507324,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03201943263411522,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.479177474975586,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.033088769763708115,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.5020246505737305,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.02975946106016636,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.455426216125488,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03213953226804733,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.501615047454834,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.03262908384203911,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.485121250152588,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.031102091073989868,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.488893508911133,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.02945449948310852,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.490121841430664,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.033680472522974014,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.51833438873291,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.03277525305747986,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.480165481567383,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.03303041681647301,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.501155376434326,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.034927863627672195,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.513175010681152,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.03098093718290329,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.503798484802246,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.029513906687498093,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.49006986618042,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029630785807967186,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.522514343261719,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.030888330191373825,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.4895734786987305,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.032392196357250214,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.471953868865967,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.03215661272406578,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.48297643661499,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.03346327692270279,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.518160820007324,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.031311362981796265,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.484991550445557,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.03056488186120987,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.525470733642578,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.03043230250477791,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.510650634765625,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.02916652522981167,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.496840953826904,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.030673664063215256,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.494868278503418,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030033906921744347,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.479182720184326,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.028828797861933708,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.471983909606934,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028316056355834007,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.5049147605896,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.03158092871308327,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.449270248413086,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.03041706420481205,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.508289813995361,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.02967553399503231,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.465579032897949,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030721619725227356,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.471003532409668,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03066832385957241,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.488987445831299,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.032297998666763306,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.491018295288086,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03221398591995239,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.471237659454346,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.021120507270097733,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.8080053329467773,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5888046026229858,
+ "eval_runtime": 59.1638,
+ "eval_samples_per_second": 41.275,
+ "eval_steps_per_second": 0.659,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.03799329698085785,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.401787757873535,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.05785121023654938,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.408820629119873,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.05964238941669464,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.427263259887695,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05466582626104355,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.403415679931641,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.05613672360777855,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.413750648498535,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.05396881327033043,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.41916561126709,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.05689029023051262,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.395590782165527,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.056090328842401505,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.421253204345703,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.0473211295902729,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.385158061981201,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.04464765638113022,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.431746959686279,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.046421997249126434,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.428816795349121,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.05039866268634796,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.452417850494385,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.05491664633154869,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.447483062744141,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05866196006536484,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.434992790222168,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.05532094091176987,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.450716018676758,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.051083534955978394,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.402632713317871,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.03815016150474548,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.433640956878662,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04203395918011665,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.415888786315918,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.043729111552238464,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.423260688781738,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.039426498115062714,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.428949356079102,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.03837858885526657,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.434070587158203,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.038315922021865845,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.471246719360352,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.03963236138224602,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.4434990882873535,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04027022793889046,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.436992645263672,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.0409243181347847,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.426936149597168,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03404631093144417,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.43462610244751,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.034393735229969025,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.437409400939941,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03635180741548538,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.451286315917969,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03349733352661133,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.419209003448486,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.03238677605986595,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.456454277038574,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.034473925828933716,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.3988728523254395,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.03455805033445358,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.434884071350098,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.037486057728528976,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.466741561889648,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.03617125004529953,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.429305076599121,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.033646389842033386,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.407780170440674,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03401118144392967,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.42977237701416,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.036534134298563004,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.451094627380371,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.0384049266576767,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.414777755737305,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03460627421736717,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.420644760131836,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.033746346831321716,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.461175918579102,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.037153810262680054,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.444699287414551,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03283539041876793,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.431022644042969,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03248969465494156,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.447571277618408,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03327852487564087,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.42236328125,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.031114818528294563,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.418244361877441,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.03231207653880119,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.455004692077637,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.030702251940965652,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.483610153198242,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031027289107441902,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.4470109939575195,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.03129073604941368,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.452799320220947,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.03634301945567131,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4661664962768555,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03718913346529007,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.412006855010986,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.034018341451883316,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.436755180358887,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032502319663763046,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.452866554260254,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.032902952283620834,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.449224472045898,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.0340626984834671,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.445425510406494,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.033442508429288864,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.432218551635742,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.03489428386092186,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.4205827713012695,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.031868621706962585,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.414196968078613,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.032986633479595184,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.46116304397583,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.03348355367779732,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.424936771392822,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03174544870853424,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.434438705444336,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.03293656185269356,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.440502166748047,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.035124607384204865,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.440943717956543,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03438971936702728,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.431408882141113,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.03537093102931976,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.43867301940918,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.03267325833439827,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.462226867675781,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.030995232984423637,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.433740615844727,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.03384929150342941,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.508179664611816,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.032025814056396484,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.407378196716309,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03288578242063522,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.455531120300293,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.032061848789453506,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.452212333679199,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022320276126265526,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7567386627197266,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5878658294677734,
+ "eval_runtime": 59.5054,
+ "eval_samples_per_second": 41.038,
+ "eval_steps_per_second": 0.655,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.039114587008953094,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.347412586212158,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.05164389684796333,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.391279220581055,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055882278829813004,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.355602264404297,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.05625041201710701,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3934855461120605,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05861108750104904,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.364226341247559,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.052431970834732056,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.363334655761719,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.06147193908691406,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.384820938110352,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.06701470166444778,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.373311996459961,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.06714873015880585,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.32745361328125,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.06684907525777817,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.373660087585449,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.05738082900643349,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.362325668334961,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05306725949048996,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.351685047149658,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.05831029266119003,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.374813079833984,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05547972396016121,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.361842632293701,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.048057056963443756,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.383205413818359,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.04682513326406479,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.365846157073975,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.05001377314329147,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.362151145935059,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.05140263959765434,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.376064300537109,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.04640384018421173,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.363801002502441,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04175020009279251,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.3477983474731445,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.04134776070713997,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.337469100952148,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.042178574949502945,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.351490020751953,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.03989115729928017,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.40725040435791,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04086652770638466,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.341119766235352,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.03931201249361038,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.3786468505859375,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.03855905309319496,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.3701605796813965,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.03803952783346176,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.401841163635254,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.036908265203237534,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.410647392272949,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.0398186556994915,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.380431175231934,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03899035602807999,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.374293804168701,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.03946344554424286,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.396918773651123,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.03738604858517647,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.404016017913818,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.03579043596982956,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.396914958953857,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.037139154970645905,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.407739639282227,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.03581099584698677,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.411419868469238,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.03548453748226166,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.332690238952637,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.037561237812042236,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.381314277648926,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.035827480256557465,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.371536731719971,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03538314625620842,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.395467758178711,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.036672890186309814,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.420590400695801,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.038981080055236816,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.395296096801758,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.038234200328588486,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.415799140930176,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.038830097764730453,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.413394927978516,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03990919142961502,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.39515495300293,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.036424681544303894,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.39730978012085,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03488154709339142,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.403696060180664,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.0363851860165596,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.368661403656006,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.035461440682411194,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.396518707275391,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03647598251700401,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.440667152404785,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.03667188808321953,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.3781352043151855,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034516457468271255,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.414007186889648,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.03536032512784004,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.42198371887207,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.037454646080732346,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.409174919128418,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03632275015115738,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3547682762146,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.038922134786844254,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.39993143081665,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.03562692925333977,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.409564971923828,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.033203717321157455,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.421782493591309,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03830202296376228,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.430092811584473,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.03835884481668472,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.379505634307861,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.037440430372953415,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.403937339782715,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03732272610068321,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.440683364868164,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03325957432389259,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.3949294090271,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.035377148538827896,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.400777816772461,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.03725374862551689,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.422494411468506,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03601110726594925,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.4008941650390625,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03193027526140213,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.400361061096191,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.034382354468107224,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.408535003662109,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.03747732564806938,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.412189483642578,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03641800582408905,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.381453037261963,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.0315779373049736,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.4169769287109375,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03583509847521782,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.385746955871582,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02574268914759159,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7456769943237305,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.588273823261261,
+ "eval_runtime": 59.9356,
+ "eval_samples_per_second": 40.744,
+ "eval_steps_per_second": 0.651,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013961605584642,
+ "grad_norm": 0.037516895681619644,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 4.33476448059082,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027923211169284,
+ "grad_norm": 0.048450272530317307,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 4.296454429626465,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041884816753926,
+ "grad_norm": 0.05368480458855629,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 4.314815998077393,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055846422338568,
+ "grad_norm": 0.05520009621977806,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 4.330798625946045,
+ "step": 1084
+ },
+ {
+ "epoch": 15.06980802792321,
+ "grad_norm": 0.05928800627589226,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 4.317663192749023,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083769633507853,
+ "grad_norm": 0.06315027177333832,
+ "learning_rate": 0.000491285979372924,
+ "loss": 4.329221725463867,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097731239092496,
+ "grad_norm": 0.06342213600873947,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 4.32546329498291,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111692844677139,
+ "grad_norm": 0.0667002946138382,
+ "learning_rate": 0.000490873338088198,
+ "loss": 4.309828758239746,
+ "step": 1088
+ },
+ {
+ "epoch": 15.12565445026178,
+ "grad_norm": 0.05264135077595711,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 4.307073593139648,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139616055846423,
+ "grad_norm": 0.051895879209041595,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 4.333526611328125,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153577661431065,
+ "grad_norm": 0.05508396774530411,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 4.330171585083008,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167539267015707,
+ "grad_norm": 0.057272691279649734,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 4.329268455505371,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181500872600349,
+ "grad_norm": 0.058956146240234375,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 4.341983795166016,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195462478184991,
+ "grad_norm": 0.0557059645652771,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 4.348274230957031,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209424083769633,
+ "grad_norm": 0.05661974847316742,
+ "learning_rate": 0.000489424334303338,
+ "loss": 4.347631454467773,
+ "step": 1095
+ },
+ {
+ "epoch": 15.223385689354275,
+ "grad_norm": 0.05338657647371292,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 4.334639072418213,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237347294938917,
+ "grad_norm": 0.05091148987412453,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 4.355465412139893,
+ "step": 1097
+ },
+ {
+ "epoch": 15.25130890052356,
+ "grad_norm": 0.04996965080499649,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 4.346457004547119,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265270506108202,
+ "grad_norm": 0.05573750659823418,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 4.346520900726318,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279232111692846,
+ "grad_norm": 0.05776917189359665,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 4.359007835388184,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293193717277488,
+ "grad_norm": 0.050820041447877884,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 4.331710338592529,
+ "step": 1101
+ },
+ {
+ "epoch": 15.30715532286213,
+ "grad_norm": 0.04441820830106735,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 4.3126420974731445,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321116928446772,
+ "grad_norm": 0.04582618921995163,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 4.312925338745117,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335078534031414,
+ "grad_norm": 0.043592359870672226,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 4.345674514770508,
+ "step": 1104
+ },
+ {
+ "epoch": 15.349040139616056,
+ "grad_norm": 0.04069037735462189,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 4.334713935852051,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363001745200698,
+ "grad_norm": 0.04083041846752167,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 4.358171463012695,
+ "step": 1106
+ },
+ {
+ "epoch": 15.37696335078534,
+ "grad_norm": 0.04119594022631645,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 4.317891597747803,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390924956369982,
+ "grad_norm": 0.042060643434524536,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 4.355810165405273,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404886561954624,
+ "grad_norm": 0.04721749201416969,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 4.353546619415283,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418848167539267,
+ "grad_norm": 0.04657730832695961,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 4.341683387756348,
+ "step": 1110
+ },
+ {
+ "epoch": 15.432809773123909,
+ "grad_norm": 0.04096715524792671,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 4.357192039489746,
+ "step": 1111
+ },
+ {
+ "epoch": 15.44677137870855,
+ "grad_norm": 0.03904155269265175,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 4.354775428771973,
+ "step": 1112
+ },
+ {
+ "epoch": 15.460732984293193,
+ "grad_norm": 0.03828521445393562,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 4.356499671936035,
+ "step": 1113
+ },
+ {
+ "epoch": 15.474694589877837,
+ "grad_norm": 0.03688685595989227,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 4.339485168457031,
+ "step": 1114
+ },
+ {
+ "epoch": 15.488656195462479,
+ "grad_norm": 0.03945618495345116,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 4.363500118255615,
+ "step": 1115
+ },
+ {
+ "epoch": 15.502617801047121,
+ "grad_norm": 0.03821462020277977,
+ "learning_rate": 0.000485033394934934,
+ "loss": 4.353529930114746,
+ "step": 1116
+ },
+ {
+ "epoch": 15.516579406631763,
+ "grad_norm": 0.036265499889850616,
+ "learning_rate": 0.000484822676912974,
+ "loss": 4.337663173675537,
+ "step": 1117
+ },
+ {
+ "epoch": 15.530541012216405,
+ "grad_norm": 0.03803930804133415,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 4.34660530090332,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544502617801047,
+ "grad_norm": 0.038425102829933167,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 4.373246192932129,
+ "step": 1119
+ },
+ {
+ "epoch": 15.55846422338569,
+ "grad_norm": 0.036852262914180756,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 4.331226825714111,
+ "step": 1120
+ },
+ {
+ "epoch": 15.572425828970331,
+ "grad_norm": 0.03700809180736542,
+ "learning_rate": 0.000483978341612154,
+ "loss": 4.375818729400635,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586387434554974,
+ "grad_norm": 0.03628664091229439,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 4.32705020904541,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600349040139616,
+ "grad_norm": 0.036831192672252655,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 4.337127208709717,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614310645724258,
+ "grad_norm": 0.03736541047692299,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 4.36502742767334,
+ "step": 1124
+ },
+ {
+ "epoch": 15.6282722513089,
+ "grad_norm": 0.03658579662442207,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 4.359844207763672,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642233856893542,
+ "grad_norm": 0.03688633441925049,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 4.3355793952941895,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656195462478184,
+ "grad_norm": 0.03477008268237114,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 4.333405494689941,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670157068062828,
+ "grad_norm": 0.03655015677213669,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 4.338570594787598,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68411867364747,
+ "grad_norm": 0.03664765506982803,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 4.352496147155762,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698080279232112,
+ "grad_norm": 0.037019114941358566,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 4.335735321044922,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712041884816754,
+ "grad_norm": 0.03949085995554924,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 4.342585563659668,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726003490401396,
+ "grad_norm": 0.04084350913763046,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 4.329622268676758,
+ "step": 1132
+ },
+ {
+ "epoch": 15.739965095986038,
+ "grad_norm": 0.038439832627773285,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 4.360318660736084,
+ "step": 1133
+ },
+ {
+ "epoch": 15.75392670157068,
+ "grad_norm": 0.03932987526059151,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 4.343446254730225,
+ "step": 1134
+ },
+ {
+ "epoch": 15.767888307155323,
+ "grad_norm": 0.041595373302698135,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 4.351495742797852,
+ "step": 1135
+ },
+ {
+ "epoch": 15.781849912739965,
+ "grad_norm": 0.03917490318417549,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 4.380537033081055,
+ "step": 1136
+ },
+ {
+ "epoch": 15.795811518324607,
+ "grad_norm": 0.038597021251916885,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 4.3527116775512695,
+ "step": 1137
+ },
+ {
+ "epoch": 15.809773123909249,
+ "grad_norm": 0.03651711344718933,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 4.369256496429443,
+ "step": 1138
+ },
+ {
+ "epoch": 15.823734729493891,
+ "grad_norm": 0.035118963569402695,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 4.350616455078125,
+ "step": 1139
+ },
+ {
+ "epoch": 15.837696335078533,
+ "grad_norm": 0.034945692867040634,
+ "learning_rate": 0.000479936075927473,
+ "loss": 4.365446090698242,
+ "step": 1140
+ },
+ {
+ "epoch": 15.851657940663177,
+ "grad_norm": 0.03519820421934128,
+ "learning_rate": 0.000479721889242305,
+ "loss": 4.357390403747559,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86561954624782,
+ "grad_norm": 0.03407098725438118,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 4.350573539733887,
+ "step": 1142
+ },
+ {
+ "epoch": 15.879581151832461,
+ "grad_norm": 0.036511123180389404,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 4.328500270843506,
+ "step": 1143
+ },
+ {
+ "epoch": 15.893542757417103,
+ "grad_norm": 0.039194293320178986,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 4.346471786499023,
+ "step": 1144
+ },
+ {
+ "epoch": 15.907504363001745,
+ "grad_norm": 0.03777911514043808,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 4.358243942260742,
+ "step": 1145
+ },
+ {
+ "epoch": 15.921465968586388,
+ "grad_norm": 0.0354655385017395,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 4.368768692016602,
+ "step": 1146
+ },
+ {
+ "epoch": 15.93542757417103,
+ "grad_norm": 0.03710418567061424,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 4.3482184410095215,
+ "step": 1147
+ },
+ {
+ "epoch": 15.949389179755672,
+ "grad_norm": 0.03545419126749039,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 4.360692501068115,
+ "step": 1148
+ },
+ {
+ "epoch": 15.963350785340314,
+ "grad_norm": 0.035963475704193115,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 4.3592119216918945,
+ "step": 1149
+ },
+ {
+ "epoch": 15.977312390924956,
+ "grad_norm": 0.03425297513604164,
+ "learning_rate": 0.000477787822648007,
+ "loss": 4.329504013061523,
+ "step": 1150
+ },
+ {
+ "epoch": 15.991273996509598,
+ "grad_norm": 0.03453255817294121,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 4.298142433166504,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.025766054168343544,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 2.7026405334472656,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.5888971090316772,
+ "eval_runtime": 58.9721,
+ "eval_samples_per_second": 41.409,
+ "eval_steps_per_second": 0.661,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013961605584644,
+ "grad_norm": 0.04088004305958748,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 4.268601417541504,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027923211169284,
+ "grad_norm": 0.05194641277194023,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 4.267483711242676,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041884816753928,
+ "grad_norm": 0.05488700792193413,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 4.2527570724487305,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05584642233857,
+ "grad_norm": 0.059292279183864594,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 4.283864974975586,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069808027923212,
+ "grad_norm": 0.0567227303981781,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 4.291148662567139,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083769633507853,
+ "grad_norm": 0.05582835525274277,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 4.250524520874023,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097731239092496,
+ "grad_norm": 0.055709999054670334,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 4.259716033935547,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111692844677137,
+ "grad_norm": 0.053324732929468155,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 4.247231483459473,
+ "step": 1160
+ },
+ {
+ "epoch": 16.12565445026178,
+ "grad_norm": 0.05367910861968994,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 4.263905048370361,
+ "step": 1161
+ },
+ {
+ "epoch": 16.13961605584642,
+ "grad_norm": 0.057094886898994446,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 4.266510009765625,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153577661431065,
+ "grad_norm": 0.05921962857246399,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 4.278281211853027,
+ "step": 1163
+ },
+ {
+ "epoch": 16.167539267015705,
+ "grad_norm": 0.060246821492910385,
+ "learning_rate": 0.000474756648870413,
+ "loss": 4.2848052978515625,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18150087260035,
+ "grad_norm": 0.05686050280928612,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 4.286434173583984,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195462478184993,
+ "grad_norm": 0.05207265168428421,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 4.246121406555176,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209424083769633,
+ "grad_norm": 0.0477963425219059,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 4.287411689758301,
+ "step": 1167
+ },
+ {
+ "epoch": 16.223385689354277,
+ "grad_norm": 0.05586806684732437,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 4.2721333503723145,
+ "step": 1168
+ },
+ {
+ "epoch": 16.237347294938917,
+ "grad_norm": 0.05427060276269913,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 4.2751641273498535,
+ "step": 1169
+ },
+ {
+ "epoch": 16.25130890052356,
+ "grad_norm": 0.0518609918653965,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 4.287444114685059,
+ "step": 1170
+ },
+ {
+ "epoch": 16.2652705061082,
+ "grad_norm": 0.053606316447257996,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 4.287125587463379,
+ "step": 1171
+ },
+ {
+ "epoch": 16.279232111692846,
+ "grad_norm": 0.05118813365697861,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 4.291893005371094,
+ "step": 1172
+ },
+ {
+ "epoch": 16.293193717277486,
+ "grad_norm": 0.04774491861462593,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 4.325250148773193,
+ "step": 1173
+ },
+ {
+ "epoch": 16.30715532286213,
+ "grad_norm": 0.04634870961308479,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 4.311764717102051,
+ "step": 1174
+ },
+ {
+ "epoch": 16.32111692844677,
+ "grad_norm": 0.04411927983164787,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 4.277068138122559,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335078534031414,
+ "grad_norm": 0.043819356709718704,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 4.309638977050781,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349040139616054,
+ "grad_norm": 0.0418904684484005,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 4.306234836578369,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363001745200698,
+ "grad_norm": 0.04327298328280449,
+ "learning_rate": 0.000471698352726896,
+ "loss": 4.296874046325684,
+ "step": 1178
+ },
+ {
+ "epoch": 16.376963350785342,
+ "grad_norm": 0.045296184718608856,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 4.288957595825195,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390924956369982,
+ "grad_norm": 0.04447923228144646,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 4.313834190368652,
+ "step": 1180
+ },
+ {
+ "epoch": 16.404886561954626,
+ "grad_norm": 0.04548390209674835,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 4.284562587738037,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418848167539267,
+ "grad_norm": 0.04567817226052284,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 4.295994758605957,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43280977312391,
+ "grad_norm": 0.04277439042925835,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 4.262106895446777,
+ "step": 1183
+ },
+ {
+ "epoch": 16.44677137870855,
+ "grad_norm": 0.043597545474767685,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 4.294536590576172,
+ "step": 1184
+ },
+ {
+ "epoch": 16.460732984293195,
+ "grad_norm": 0.04424033313989639,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 4.297497749328613,
+ "step": 1185
+ },
+ {
+ "epoch": 16.474694589877835,
+ "grad_norm": 0.04354134574532509,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 4.321508884429932,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48865619546248,
+ "grad_norm": 0.04241178184747696,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 4.322726249694824,
+ "step": 1187
+ },
+ {
+ "epoch": 16.50261780104712,
+ "grad_norm": 0.039317622780799866,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 4.311155319213867,
+ "step": 1188
+ },
+ {
+ "epoch": 16.516579406631763,
+ "grad_norm": 0.04126536473631859,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 4.306636810302734,
+ "step": 1189
+ },
+ {
+ "epoch": 16.530541012216403,
+ "grad_norm": 0.040116794407367706,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 4.314671516418457,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544502617801047,
+ "grad_norm": 0.04045999422669411,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 4.301012992858887,
+ "step": 1191
+ },
+ {
+ "epoch": 16.55846422338569,
+ "grad_norm": 0.04013422504067421,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 4.29913330078125,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57242582897033,
+ "grad_norm": 0.03878502920269966,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 4.351600646972656,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586387434554975,
+ "grad_norm": 0.038935888558626175,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 4.27667236328125,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600349040139616,
+ "grad_norm": 0.04055892676115036,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 4.29019021987915,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61431064572426,
+ "grad_norm": 0.03959144279360771,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 4.327725887298584,
+ "step": 1196
+ },
+ {
+ "epoch": 16.6282722513089,
+ "grad_norm": 0.039878539741039276,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 4.297952651977539,
+ "step": 1197
+ },
+ {
+ "epoch": 16.642233856893544,
+ "grad_norm": 0.041209928691387177,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 4.313030242919922,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656195462478184,
+ "grad_norm": 0.04346393793821335,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 4.274901390075684,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670157068062828,
+ "grad_norm": 0.04090269282460213,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 4.330657958984375,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68411867364747,
+ "grad_norm": 0.04004506766796112,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 4.323106288909912,
+ "step": 1201
+ },
+ {
+ "epoch": 16.698080279232112,
+ "grad_norm": 0.041541460901498795,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 4.280947685241699,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712041884816752,
+ "grad_norm": 0.04033653065562248,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 4.341434955596924,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726003490401396,
+ "grad_norm": 0.04104280099272728,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 4.2931671142578125,
+ "step": 1204
+ },
+ {
+ "epoch": 16.739965095986037,
+ "grad_norm": 0.039398156106472015,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 4.316840648651123,
+ "step": 1205
+ },
+ {
+ "epoch": 16.75392670157068,
+ "grad_norm": 0.03980560600757599,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 4.325345039367676,
+ "step": 1206
+ },
+ {
+ "epoch": 16.767888307155324,
+ "grad_norm": 0.040656186640262604,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 4.32675838470459,
+ "step": 1207
+ },
+ {
+ "epoch": 16.781849912739965,
+ "grad_norm": 0.03775986284017563,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 4.314423561096191,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79581151832461,
+ "grad_norm": 0.04043284431099892,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 4.344202041625977,
+ "step": 1209
+ },
+ {
+ "epoch": 16.80977312390925,
+ "grad_norm": 0.04074503108859062,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 4.350123405456543,
+ "step": 1210
+ },
+ {
+ "epoch": 16.823734729493893,
+ "grad_norm": 0.04117420315742493,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 4.350917816162109,
+ "step": 1211
+ },
+ {
+ "epoch": 16.837696335078533,
+ "grad_norm": 0.03799610957503319,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 4.289987564086914,
+ "step": 1212
+ },
+ {
+ "epoch": 16.851657940663177,
+ "grad_norm": 0.04071265831589699,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 4.306744575500488,
+ "step": 1213
+ },
+ {
+ "epoch": 16.865619546247817,
+ "grad_norm": 0.04109715297818184,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 4.318753242492676,
+ "step": 1214
+ },
+ {
+ "epoch": 16.87958115183246,
+ "grad_norm": 0.03910145163536072,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 4.281325340270996,
+ "step": 1215
+ },
+ {
+ "epoch": 16.8935427574171,
+ "grad_norm": 0.03752733767032623,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 4.3231048583984375,
+ "step": 1216
+ },
+ {
+ "epoch": 16.907504363001745,
+ "grad_norm": 0.03752103075385094,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 4.351336479187012,
+ "step": 1217
+ },
+ {
+ "epoch": 16.921465968586386,
+ "grad_norm": 0.03675422817468643,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 4.282256126403809,
+ "step": 1218
+ },
+ {
+ "epoch": 16.93542757417103,
+ "grad_norm": 0.03850258141756058,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 4.322597503662109,
+ "step": 1219
+ },
+ {
+ "epoch": 16.949389179755673,
+ "grad_norm": 0.03692392632365227,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 4.320333957672119,
+ "step": 1220
+ },
+ {
+ "epoch": 16.963350785340314,
+ "grad_norm": 0.035366240888834,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 4.31341552734375,
+ "step": 1221
+ },
+ {
+ "epoch": 16.977312390924958,
+ "grad_norm": 0.037120968103408813,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 4.321141242980957,
+ "step": 1222
+ },
+ {
+ "epoch": 16.991273996509598,
+ "grad_norm": 0.03721754252910614,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 4.305757522583008,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.026698043569922447,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 2.716372489929199,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.5894472599029541,
+ "eval_runtime": 58.927,
+ "eval_samples_per_second": 41.441,
+ "eval_steps_per_second": 0.662,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013961605584644,
+ "grad_norm": 0.044532276690006256,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 4.243244171142578,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027923211169284,
+ "grad_norm": 0.050529688596725464,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 4.221382141113281,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041884816753928,
+ "grad_norm": 0.05690870061516762,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 4.240285873413086,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05584642233857,
+ "grad_norm": 0.054803814738988876,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 4.213047027587891,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069808027923212,
+ "grad_norm": 0.051825862377882004,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 4.226880073547363,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083769633507853,
+ "grad_norm": 0.049802035093307495,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 4.25303840637207,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097731239092496,
+ "grad_norm": 0.05290328338742256,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 4.199422836303711,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111692844677137,
+ "grad_norm": 0.0531814880669117,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 4.201648712158203,
+ "step": 1232
+ },
+ {
+ "epoch": 17.12565445026178,
+ "grad_norm": 0.048972588032484055,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 4.207578182220459,
+ "step": 1233
+ },
+ {
+ "epoch": 17.13961605584642,
+ "grad_norm": 0.051345713436603546,
+ "learning_rate": 0.000459203665939781,
+ "loss": 4.206205368041992,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153577661431065,
+ "grad_norm": 0.059012677520513535,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 4.228360176086426,
+ "step": 1235
+ },
+ {
+ "epoch": 17.167539267015705,
+ "grad_norm": 0.05984492972493172,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 4.234362602233887,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18150087260035,
+ "grad_norm": 0.060886185616254807,
+ "learning_rate": 0.000458522886595554,
+ "loss": 4.253525257110596,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195462478184993,
+ "grad_norm": 0.06476911902427673,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 4.249889373779297,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209424083769633,
+ "grad_norm": 0.059773124754428864,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 4.225249290466309,
+ "step": 1239
+ },
+ {
+ "epoch": 17.223385689354277,
+ "grad_norm": 0.06063077598810196,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 4.25465202331543,
+ "step": 1240
+ },
+ {
+ "epoch": 17.237347294938917,
+ "grad_norm": 0.059922896325588226,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 4.252101898193359,
+ "step": 1241
+ },
+ {
+ "epoch": 17.25130890052356,
+ "grad_norm": 0.05849695950746536,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 4.268218040466309,
+ "step": 1242
+ },
+ {
+ "epoch": 17.2652705061082,
+ "grad_norm": 0.05769657716155052,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 4.2395782470703125,
+ "step": 1243
+ },
+ {
+ "epoch": 17.279232111692846,
+ "grad_norm": 0.058189719915390015,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 4.257606506347656,
+ "step": 1244
+ },
+ {
+ "epoch": 17.293193717277486,
+ "grad_norm": 0.059679049998521805,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 4.278588771820068,
+ "step": 1245
+ },
+ {
+ "epoch": 17.30715532286213,
+ "grad_norm": 0.0579286590218544,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 4.25511360168457,
+ "step": 1246
+ },
+ {
+ "epoch": 17.32111692844677,
+ "grad_norm": 0.05269965901970863,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 4.241326332092285,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335078534031414,
+ "grad_norm": 0.04828358441591263,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 4.243182182312012,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349040139616054,
+ "grad_norm": 0.050172045826911926,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 4.295144081115723,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363001745200698,
+ "grad_norm": 0.047297943383455276,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 4.250682830810547,
+ "step": 1250
+ },
+ {
+ "epoch": 17.376963350785342,
+ "grad_norm": 0.04574039950966835,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 4.269444465637207,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390924956369982,
+ "grad_norm": 0.042310427874326706,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 4.255361557006836,
+ "step": 1252
+ },
+ {
+ "epoch": 17.404886561954626,
+ "grad_norm": 0.04268861189484596,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 4.252292156219482,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418848167539267,
+ "grad_norm": 0.0424286350607872,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 4.24439811706543,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43280977312391,
+ "grad_norm": 0.042049869894981384,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 4.236266136169434,
+ "step": 1255
+ },
+ {
+ "epoch": 17.44677137870855,
+ "grad_norm": 0.04185241088271141,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 4.268826484680176,
+ "step": 1256
+ },
+ {
+ "epoch": 17.460732984293195,
+ "grad_norm": 0.04423609375953674,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 4.240240573883057,
+ "step": 1257
+ },
+ {
+ "epoch": 17.474694589877835,
+ "grad_norm": 0.041972566395998,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 4.269453048706055,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48865619546248,
+ "grad_norm": 0.04493343085050583,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 4.270544052124023,
+ "step": 1259
+ },
+ {
+ "epoch": 17.50261780104712,
+ "grad_norm": 0.04502921551465988,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 4.271880149841309,
+ "step": 1260
+ },
+ {
+ "epoch": 17.516579406631763,
+ "grad_norm": 0.044518254697322845,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 4.249786376953125,
+ "step": 1261
+ },
+ {
+ "epoch": 17.530541012216403,
+ "grad_norm": 0.044482357800006866,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 4.300779819488525,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544502617801047,
+ "grad_norm": 0.0424671471118927,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 4.248597621917725,
+ "step": 1263
+ },
+ {
+ "epoch": 17.55846422338569,
+ "grad_norm": 0.04136480763554573,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 4.276763916015625,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57242582897033,
+ "grad_norm": 0.04094843938946724,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 4.263628005981445,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586387434554975,
+ "grad_norm": 0.04415388032793999,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 4.257946014404297,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600349040139616,
+ "grad_norm": 0.04177384823560715,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 4.289669036865234,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61431064572426,
+ "grad_norm": 0.040863484144210815,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 4.320903778076172,
+ "step": 1268
+ },
+ {
+ "epoch": 17.6282722513089,
+ "grad_norm": 0.04025879129767418,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 4.268717288970947,
+ "step": 1269
+ },
+ {
+ "epoch": 17.642233856893544,
+ "grad_norm": 0.04161325842142105,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 4.275268077850342,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656195462478184,
+ "grad_norm": 0.04260355979204178,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 4.263361930847168,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670157068062828,
+ "grad_norm": 0.043366871774196625,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 4.254581928253174,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68411867364747,
+ "grad_norm": 0.04299113526940346,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 4.268922328948975,
+ "step": 1273
+ },
+ {
+ "epoch": 17.698080279232112,
+ "grad_norm": 0.041865330189466476,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 4.290097236633301,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712041884816752,
+ "grad_norm": 0.04515853524208069,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 4.261376857757568,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726003490401396,
+ "grad_norm": 0.046030085533857346,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 4.290409564971924,
+ "step": 1276
+ },
+ {
+ "epoch": 17.739965095986037,
+ "grad_norm": 0.045034267008304596,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 4.28201961517334,
+ "step": 1277
+ },
+ {
+ "epoch": 17.75392670157068,
+ "grad_norm": 0.04293077066540718,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 4.282614707946777,
+ "step": 1278
+ },
+ {
+ "epoch": 17.767888307155324,
+ "grad_norm": 0.043866172432899475,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 4.2777557373046875,
+ "step": 1279
+ },
+ {
+ "epoch": 17.781849912739965,
+ "grad_norm": 0.04693951830267906,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 4.272597312927246,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79581151832461,
+ "grad_norm": 0.04072774201631546,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 4.284282684326172,
+ "step": 1281
+ },
+ {
+ "epoch": 17.80977312390925,
+ "grad_norm": 0.04241872951388359,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 4.254406452178955,
+ "step": 1282
+ },
+ {
+ "epoch": 17.823734729493893,
+ "grad_norm": 0.04614942893385887,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 4.297187805175781,
+ "step": 1283
+ },
+ {
+ "epoch": 17.837696335078533,
+ "grad_norm": 0.04326155036687851,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 4.292263984680176,
+ "step": 1284
+ },
+ {
+ "epoch": 17.851657940663177,
+ "grad_norm": 0.04053817689418793,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 4.2925920486450195,
+ "step": 1285
+ },
+ {
+ "epoch": 17.865619546247817,
+ "grad_norm": 0.040867388248443604,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 4.276965141296387,
+ "step": 1286
+ },
+ {
+ "epoch": 17.87958115183246,
+ "grad_norm": 0.041000183671712875,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 4.257638454437256,
+ "step": 1287
+ },
+ {
+ "epoch": 17.8935427574171,
+ "grad_norm": 0.04014866054058075,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 4.2734479904174805,
+ "step": 1288
+ },
+ {
+ "epoch": 17.907504363001745,
+ "grad_norm": 0.041237372905015945,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 4.263461112976074,
+ "step": 1289
+ },
+ {
+ "epoch": 17.921465968586386,
+ "grad_norm": 0.039109110832214355,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 4.295286655426025,
+ "step": 1290
+ },
+ {
+ "epoch": 17.93542757417103,
+ "grad_norm": 0.03911379352211952,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 4.254448413848877,
+ "step": 1291
+ },
+ {
+ "epoch": 17.949389179755673,
+ "grad_norm": 0.03997721150517464,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 4.284714698791504,
+ "step": 1292
+ },
+ {
+ "epoch": 17.963350785340314,
+ "grad_norm": 0.04011402651667595,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 4.278304100036621,
+ "step": 1293
+ },
+ {
+ "epoch": 17.977312390924958,
+ "grad_norm": 0.04040595516562462,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 4.288183212280273,
+ "step": 1294
+ },
+ {
+ "epoch": 17.991273996509598,
+ "grad_norm": 0.038857635110616684,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 4.274664402008057,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.028110472485423088,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 2.6625723838806152,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.5912551283836365,
+ "eval_runtime": 60.273,
+ "eval_samples_per_second": 40.516,
+ "eval_steps_per_second": 0.647,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013961605584644,
+ "grad_norm": 0.041634928435087204,
+ "learning_rate": 0.000444677103403854,
+ "loss": 4.181324005126953,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027923211169284,
+ "grad_norm": 0.051596369594335556,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 4.204722881317139,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041884816753928,
+ "grad_norm": 0.058580417186021805,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 4.207996845245361,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05584642233857,
+ "grad_norm": 0.05986522138118744,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 4.184011459350586,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069808027923212,
+ "grad_norm": 0.051045969128608704,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 4.148118019104004,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083769633507853,
+ "grad_norm": 0.045477766543626785,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 4.207469940185547,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097731239092496,
+ "grad_norm": 0.0488731674849987,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 4.201406478881836,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111692844677137,
+ "grad_norm": 0.05013517662882805,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 4.192239761352539,
+ "step": 1304
+ },
+ {
+ "epoch": 18.12565445026178,
+ "grad_norm": 0.05091651529073715,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 4.188252925872803,
+ "step": 1305
+ },
+ {
+ "epoch": 18.13961605584642,
+ "grad_norm": 0.051825884729623795,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 4.157929420471191,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153577661431065,
+ "grad_norm": 0.049625713378190994,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 4.208776473999023,
+ "step": 1307
+ },
+ {
+ "epoch": 18.167539267015705,
+ "grad_norm": 0.04816461354494095,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 4.180496692657471,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18150087260035,
+ "grad_norm": 0.04720531404018402,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 4.178536415100098,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195462478184993,
+ "grad_norm": 0.0475568026304245,
+ "learning_rate": 0.000441621628895411,
+ "loss": 4.203497886657715,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209424083769633,
+ "grad_norm": 0.04613947868347168,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 4.226914405822754,
+ "step": 1311
+ },
+ {
+ "epoch": 18.223385689354277,
+ "grad_norm": 0.04585929960012436,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 4.175119400024414,
+ "step": 1312
+ },
+ {
+ "epoch": 18.237347294938917,
+ "grad_norm": 0.045777060091495514,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 4.181760311126709,
+ "step": 1313
+ },
+ {
+ "epoch": 18.25130890052356,
+ "grad_norm": 0.04583602771162987,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 4.179868698120117,
+ "step": 1314
+ },
+ {
+ "epoch": 18.2652705061082,
+ "grad_norm": 0.04706503078341484,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 4.203325271606445,
+ "step": 1315
+ },
+ {
+ "epoch": 18.279232111692846,
+ "grad_norm": 0.045122336596250534,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 4.187152862548828,
+ "step": 1316
+ },
+ {
+ "epoch": 18.293193717277486,
+ "grad_norm": 0.04495825991034508,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 4.2080535888671875,
+ "step": 1317
+ },
+ {
+ "epoch": 18.30715532286213,
+ "grad_norm": 0.04767543077468872,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 4.207225322723389,
+ "step": 1318
+ },
+ {
+ "epoch": 18.32111692844677,
+ "grad_norm": 0.053396135568618774,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 4.227478981018066,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335078534031414,
+ "grad_norm": 0.0512089841067791,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 4.1904754638671875,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349040139616054,
+ "grad_norm": 0.04967977851629257,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 4.2079925537109375,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363001745200698,
+ "grad_norm": 0.04800970479846001,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 4.206037998199463,
+ "step": 1322
+ },
+ {
+ "epoch": 18.376963350785342,
+ "grad_norm": 0.04993556812405586,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 4.210054397583008,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390924956369982,
+ "grad_norm": 0.0480492040514946,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 4.229897499084473,
+ "step": 1324
+ },
+ {
+ "epoch": 18.404886561954626,
+ "grad_norm": 0.048699282109737396,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 4.17500114440918,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418848167539267,
+ "grad_norm": 0.04681490734219551,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 4.196526527404785,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43280977312391,
+ "grad_norm": 0.04674449935555458,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 4.246790409088135,
+ "step": 1327
+ },
+ {
+ "epoch": 18.44677137870855,
+ "grad_norm": 0.04650607705116272,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 4.225878715515137,
+ "step": 1328
+ },
+ {
+ "epoch": 18.460732984293195,
+ "grad_norm": 0.047200269997119904,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 4.20685338973999,
+ "step": 1329
+ },
+ {
+ "epoch": 18.474694589877835,
+ "grad_norm": 0.04322655498981476,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 4.212920665740967,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48865619546248,
+ "grad_norm": 0.04489128664135933,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 4.266653060913086,
+ "step": 1331
+ },
+ {
+ "epoch": 18.50261780104712,
+ "grad_norm": 0.048838935792446136,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 4.2364501953125,
+ "step": 1332
+ },
+ {
+ "epoch": 18.516579406631763,
+ "grad_norm": 0.04893286153674126,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 4.202394485473633,
+ "step": 1333
+ },
+ {
+ "epoch": 18.530541012216403,
+ "grad_norm": 0.046900101006031036,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 4.209832191467285,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544502617801047,
+ "grad_norm": 0.044056762009859085,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 4.229903221130371,
+ "step": 1335
+ },
+ {
+ "epoch": 18.55846422338569,
+ "grad_norm": 0.04681328684091568,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 4.22783899307251,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57242582897033,
+ "grad_norm": 0.046087440103292465,
+ "learning_rate": 0.000435215814386134,
+ "loss": 4.2078986167907715,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586387434554975,
+ "grad_norm": 0.04363090172410011,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 4.218686103820801,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600349040139616,
+ "grad_norm": 0.046255532652139664,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 4.2475128173828125,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61431064572426,
+ "grad_norm": 0.045545510947704315,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 4.233272552490234,
+ "step": 1340
+ },
+ {
+ "epoch": 18.6282722513089,
+ "grad_norm": 0.045259326696395874,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 4.252043724060059,
+ "step": 1341
+ },
+ {
+ "epoch": 18.642233856893544,
+ "grad_norm": 0.042144861072301865,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 4.247762203216553,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656195462478184,
+ "grad_norm": 0.04298558086156845,
+ "learning_rate": 0.000433781621332752,
+ "loss": 4.222329616546631,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670157068062828,
+ "grad_norm": 0.044070884585380554,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 4.222488880157471,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68411867364747,
+ "grad_norm": 0.044758375734090805,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 4.238160133361816,
+ "step": 1345
+ },
+ {
+ "epoch": 18.698080279232112,
+ "grad_norm": 0.04324857518076897,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 4.220003128051758,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712041884816752,
+ "grad_norm": 0.04349793121218681,
+ "learning_rate": 0.000432823382695327,
+ "loss": 4.251608848571777,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726003490401396,
+ "grad_norm": 0.04483242705464363,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 4.243319511413574,
+ "step": 1348
+ },
+ {
+ "epoch": 18.739965095986037,
+ "grad_norm": 0.04391994699835777,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 4.237479209899902,
+ "step": 1349
+ },
+ {
+ "epoch": 18.75392670157068,
+ "grad_norm": 0.042584922164678574,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 4.2309980392456055,
+ "step": 1350
+ },
+ {
+ "epoch": 18.767888307155324,
+ "grad_norm": 0.04363991320133209,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 4.216980934143066,
+ "step": 1351
+ },
+ {
+ "epoch": 18.781849912739965,
+ "grad_norm": 0.04662707448005676,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 4.257450103759766,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79581151832461,
+ "grad_norm": 0.0462074875831604,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 4.222329139709473,
+ "step": 1353
+ },
+ {
+ "epoch": 18.80977312390925,
+ "grad_norm": 0.043543554842472076,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 4.231960773468018,
+ "step": 1354
+ },
+ {
+ "epoch": 18.823734729493893,
+ "grad_norm": 0.04444567486643791,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 4.225714683532715,
+ "step": 1355
+ },
+ {
+ "epoch": 18.837696335078533,
+ "grad_norm": 0.04507977142930031,
+ "learning_rate": 0.000430661245733797,
+ "loss": 4.234722137451172,
+ "step": 1356
+ },
+ {
+ "epoch": 18.851657940663177,
+ "grad_norm": 0.043786536902189255,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 4.274998664855957,
+ "step": 1357
+ },
+ {
+ "epoch": 18.865619546247817,
+ "grad_norm": 0.04188325256109238,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 4.260570526123047,
+ "step": 1358
+ },
+ {
+ "epoch": 18.87958115183246,
+ "grad_norm": 0.0439927875995636,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 4.250492572784424,
+ "step": 1359
+ },
+ {
+ "epoch": 18.8935427574171,
+ "grad_norm": 0.04520147293806076,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 4.250911712646484,
+ "step": 1360
+ },
+ {
+ "epoch": 18.907504363001745,
+ "grad_norm": 0.04442249983549118,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 4.255832672119141,
+ "step": 1361
+ },
+ {
+ "epoch": 18.921465968586386,
+ "grad_norm": 0.044756144285202026,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 4.226641654968262,
+ "step": 1362
+ },
+ {
+ "epoch": 18.93542757417103,
+ "grad_norm": 0.04647127911448479,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 4.2110915184021,
+ "step": 1363
+ },
+ {
+ "epoch": 18.949389179755673,
+ "grad_norm": 0.045986004173755646,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 4.243263244628906,
+ "step": 1364
+ },
+ {
+ "epoch": 18.963350785340314,
+ "grad_norm": 0.04533308371901512,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 4.253750801086426,
+ "step": 1365
+ },
+ {
+ "epoch": 18.977312390924958,
+ "grad_norm": 0.04307471215724945,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 4.233024597167969,
+ "step": 1366
+ },
+ {
+ "epoch": 18.991273996509598,
+ "grad_norm": 0.0416078194975853,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 4.264965534210205,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.030926328152418137,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 2.651331901550293,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.592528223991394,
+ "eval_runtime": 59.67,
+ "eval_samples_per_second": 40.925,
+ "eval_steps_per_second": 0.654,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013961605584644,
+ "grad_norm": 0.04524356126785278,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 4.153796672821045,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027923211169284,
+ "grad_norm": 0.05105067789554596,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 4.154603004455566,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041884816753928,
+ "grad_norm": 0.052823506295681,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 4.1599273681640625,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05584642233857,
+ "grad_norm": 0.05688423290848732,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 4.124622344970703,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069808027923212,
+ "grad_norm": 0.053178343921899796,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 4.138644218444824,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083769633507853,
+ "grad_norm": 0.05473732203245163,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 4.138207912445068,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097731239092496,
+ "grad_norm": 0.055207233875989914,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 4.155735015869141,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111692844677137,
+ "grad_norm": 0.05484927073121071,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 4.140205383300781,
+ "step": 1376
+ },
+ {
+ "epoch": 19.12565445026178,
+ "grad_norm": 0.05589596927165985,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 4.160840034484863,
+ "step": 1377
+ },
+ {
+ "epoch": 19.13961605584642,
+ "grad_norm": 0.05222902446985245,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 4.154781341552734,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153577661431065,
+ "grad_norm": 0.051824700087308884,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 4.134243488311768,
+ "step": 1379
+ },
+ {
+ "epoch": 19.167539267015705,
+ "grad_norm": 0.0517590269446373,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 4.125945568084717,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18150087260035,
+ "grad_norm": 0.05399726703763008,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 4.146993160247803,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195462478184993,
+ "grad_norm": 0.054430052638053894,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 4.1310906410217285,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209424083769633,
+ "grad_norm": 0.0551968477666378,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 4.170640468597412,
+ "step": 1383
+ },
+ {
+ "epoch": 19.223385689354277,
+ "grad_norm": 0.06050962209701538,
+ "learning_rate": 0.000423881957237287,
+ "loss": 4.166876316070557,
+ "step": 1384
+ },
+ {
+ "epoch": 19.237347294938917,
+ "grad_norm": 0.055587392300367355,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 4.150339603424072,
+ "step": 1385
+ },
+ {
+ "epoch": 19.25130890052356,
+ "grad_norm": 0.05193328112363815,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 4.171754360198975,
+ "step": 1386
+ },
+ {
+ "epoch": 19.2652705061082,
+ "grad_norm": 0.05272402986884117,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 4.165149688720703,
+ "step": 1387
+ },
+ {
+ "epoch": 19.279232111692846,
+ "grad_norm": 0.056283943355083466,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 4.193273544311523,
+ "step": 1388
+ },
+ {
+ "epoch": 19.293193717277486,
+ "grad_norm": 0.05825977027416229,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 4.192652225494385,
+ "step": 1389
+ },
+ {
+ "epoch": 19.30715532286213,
+ "grad_norm": 0.05539379641413689,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 4.187044143676758,
+ "step": 1390
+ },
+ {
+ "epoch": 19.32111692844677,
+ "grad_norm": 0.05300498753786087,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 4.168220043182373,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335078534031414,
+ "grad_norm": 0.05425756797194481,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 4.185193061828613,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349040139616054,
+ "grad_norm": 0.05527182295918465,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 4.1618332862854,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363001745200698,
+ "grad_norm": 0.054422613233327866,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 4.195295810699463,
+ "step": 1394
+ },
+ {
+ "epoch": 19.376963350785342,
+ "grad_norm": 0.052717432379722595,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 4.143290996551514,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390924956369982,
+ "grad_norm": 0.05154350399971008,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 4.115989685058594,
+ "step": 1396
+ },
+ {
+ "epoch": 19.404886561954626,
+ "grad_norm": 0.0489451140165329,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 4.169506072998047,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418848167539267,
+ "grad_norm": 0.049093786627054214,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 4.164122104644775,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43280977312391,
+ "grad_norm": 0.04690805822610855,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 4.186679840087891,
+ "step": 1399
+ },
+ {
+ "epoch": 19.44677137870855,
+ "grad_norm": 0.046708665788173676,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 4.184325695037842,
+ "step": 1400
+ },
+ {
+ "epoch": 19.460732984293195,
+ "grad_norm": 0.04617927595973015,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 4.201678276062012,
+ "step": 1401
+ },
+ {
+ "epoch": 19.474694589877835,
+ "grad_norm": 0.05017530545592308,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 4.18928337097168,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48865619546248,
+ "grad_norm": 0.05197977274656296,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 4.188801288604736,
+ "step": 1403
+ },
+ {
+ "epoch": 19.50261780104712,
+ "grad_norm": 0.049691297113895416,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 4.209467887878418,
+ "step": 1404
+ },
+ {
+ "epoch": 19.516579406631763,
+ "grad_norm": 0.049709953367710114,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 4.19733190536499,
+ "step": 1405
+ },
+ {
+ "epoch": 19.530541012216403,
+ "grad_norm": 0.05274137482047081,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 4.221798896789551,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544502617801047,
+ "grad_norm": 0.05015392601490021,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 4.210565567016602,
+ "step": 1407
+ },
+ {
+ "epoch": 19.55846422338569,
+ "grad_norm": 0.05023058131337166,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 4.184835910797119,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57242582897033,
+ "grad_norm": 0.0479239858686924,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 4.199846267700195,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586387434554975,
+ "grad_norm": 0.0481228232383728,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 4.185157775878906,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600349040139616,
+ "grad_norm": 0.04909738525748253,
+ "learning_rate": 0.000417272427439646,
+ "loss": 4.207550525665283,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61431064572426,
+ "grad_norm": 0.0475567989051342,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 4.213225841522217,
+ "step": 1412
+ },
+ {
+ "epoch": 19.6282722513089,
+ "grad_norm": 0.048073310405015945,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 4.212669372558594,
+ "step": 1413
+ },
+ {
+ "epoch": 19.642233856893544,
+ "grad_norm": 0.04523796588182449,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 4.1950860023498535,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656195462478184,
+ "grad_norm": 0.04359745234251022,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 4.2136640548706055,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670157068062828,
+ "grad_norm": 0.047282423824071884,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 4.209680557250977,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68411867364747,
+ "grad_norm": 0.0474061481654644,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 4.188494682312012,
+ "step": 1417
+ },
+ {
+ "epoch": 19.698080279232112,
+ "grad_norm": 0.04402102902531624,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 4.207236289978027,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712041884816752,
+ "grad_norm": 0.04398655146360397,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 4.210899353027344,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726003490401396,
+ "grad_norm": 0.045767925679683685,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 4.171738624572754,
+ "step": 1420
+ },
+ {
+ "epoch": 19.739965095986037,
+ "grad_norm": 0.04498986527323723,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 4.222414970397949,
+ "step": 1421
+ },
+ {
+ "epoch": 19.75392670157068,
+ "grad_norm": 0.04282551258802414,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 4.202138900756836,
+ "step": 1422
+ },
+ {
+ "epoch": 19.767888307155324,
+ "grad_norm": 0.04392262175679207,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 4.184439659118652,
+ "step": 1423
+ },
+ {
+ "epoch": 19.781849912739965,
+ "grad_norm": 0.04343154653906822,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 4.204674243927002,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79581151832461,
+ "grad_norm": 0.04496419429779053,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 4.23316764831543,
+ "step": 1425
+ },
+ {
+ "epoch": 19.80977312390925,
+ "grad_norm": 0.04488370940089226,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 4.201659202575684,
+ "step": 1426
+ },
+ {
+ "epoch": 19.823734729493893,
+ "grad_norm": 0.045598775148391724,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 4.187412261962891,
+ "step": 1427
+ },
+ {
+ "epoch": 19.837696335078533,
+ "grad_norm": 0.043846048414707184,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 4.215444564819336,
+ "step": 1428
+ },
+ {
+ "epoch": 19.851657940663177,
+ "grad_norm": 0.043062254786491394,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 4.2074055671691895,
+ "step": 1429
+ },
+ {
+ "epoch": 19.865619546247817,
+ "grad_norm": 0.04544302448630333,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 4.184919357299805,
+ "step": 1430
+ },
+ {
+ "epoch": 19.87958115183246,
+ "grad_norm": 0.046144865453243256,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 4.2175750732421875,
+ "step": 1431
+ },
+ {
+ "epoch": 19.8935427574171,
+ "grad_norm": 0.04465445503592491,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 4.1851115226745605,
+ "step": 1432
+ },
+ {
+ "epoch": 19.907504363001745,
+ "grad_norm": 0.045913565903902054,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 4.2226152420043945,
+ "step": 1433
+ },
+ {
+ "epoch": 19.921465968586386,
+ "grad_norm": 0.04401300847530365,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 4.216772556304932,
+ "step": 1434
+ },
+ {
+ "epoch": 19.93542757417103,
+ "grad_norm": 0.042624711990356445,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 4.2133684158325195,
+ "step": 1435
+ },
+ {
+ "epoch": 19.949389179755673,
+ "grad_norm": 0.04533256217837334,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 4.218449592590332,
+ "step": 1436
+ },
+ {
+ "epoch": 19.963350785340314,
+ "grad_norm": 0.04764309898018837,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 4.191402912139893,
+ "step": 1437
+ },
+ {
+ "epoch": 19.977312390924958,
+ "grad_norm": 0.04689318686723709,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 4.2113037109375,
+ "step": 1438
+ },
+ {
+ "epoch": 19.991273996509598,
+ "grad_norm": 0.04479913413524628,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 4.194989204406738,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.03151386231184006,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 2.6285438537597656,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 0.5938133597373962,
+ "eval_runtime": 58.9656,
+ "eval_samples_per_second": 41.414,
+ "eval_steps_per_second": 0.661,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013961605584644,
+ "grad_norm": 0.04884875938296318,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 4.1257734298706055,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027923211169284,
+ "grad_norm": 0.05968257039785385,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 4.091641426086426,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041884816753928,
+ "grad_norm": 0.06437838077545166,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 4.122465133666992,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05584642233857,
+ "grad_norm": 0.057831212878227234,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 4.129654884338379,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069808027923212,
+ "grad_norm": 0.049407217651605606,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 4.116935729980469,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083769633507853,
+ "grad_norm": 0.051957421004772186,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 4.125039577484131,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097731239092496,
+ "grad_norm": 0.05425718426704407,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 4.113485336303711,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111692844677137,
+ "grad_norm": 0.055318959057331085,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 4.119422912597656,
+ "step": 1448
+ },
+ {
+ "epoch": 20.12565445026178,
+ "grad_norm": 0.0548059344291687,
+ "learning_rate": 0.000407857329622967,
+ "loss": 4.097405910491943,
+ "step": 1449
+ },
+ {
+ "epoch": 20.13961605584642,
+ "grad_norm": 0.051046356558799744,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 4.115051746368408,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153577661431065,
+ "grad_norm": 0.05079492926597595,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 4.126070499420166,
+ "step": 1451
+ },
+ {
+ "epoch": 20.167539267015705,
+ "grad_norm": 0.04777265712618828,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 4.118952751159668,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18150087260035,
+ "grad_norm": 0.04895703122019768,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 4.095477104187012,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195462478184993,
+ "grad_norm": 0.051196496933698654,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 4.116866111755371,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209424083769633,
+ "grad_norm": 0.05625293031334877,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 4.141980171203613,
+ "step": 1455
+ },
+ {
+ "epoch": 20.223385689354277,
+ "grad_norm": 0.05852660909295082,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 4.126821517944336,
+ "step": 1456
+ },
+ {
+ "epoch": 20.237347294938917,
+ "grad_norm": 0.05419151484966278,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 4.10935640335083,
+ "step": 1457
+ },
+ {
+ "epoch": 20.25130890052356,
+ "grad_norm": 0.05251913517713547,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 4.151289463043213,
+ "step": 1458
+ },
+ {
+ "epoch": 20.2652705061082,
+ "grad_norm": 0.05223092809319496,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 4.149935722351074,
+ "step": 1459
+ },
+ {
+ "epoch": 20.279232111692846,
+ "grad_norm": 0.05314765125513077,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 4.106821060180664,
+ "step": 1460
+ },
+ {
+ "epoch": 20.293193717277486,
+ "grad_norm": 0.05472502112388611,
+ "learning_rate": 0.000404858275823277,
+ "loss": 4.120498180389404,
+ "step": 1461
+ },
+ {
+ "epoch": 20.30715532286213,
+ "grad_norm": 0.05233139917254448,
+ "learning_rate": 0.000404607816444578,
+ "loss": 4.1237335205078125,
+ "step": 1462
+ },
+ {
+ "epoch": 20.32111692844677,
+ "grad_norm": 0.05210232734680176,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 4.1271209716796875,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335078534031414,
+ "grad_norm": 0.049779925495386124,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 4.121427536010742,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349040139616054,
+ "grad_norm": 0.05232279747724533,
+ "learning_rate": 0.000403855947934478,
+ "loss": 4.136429786682129,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363001745200698,
+ "grad_norm": 0.05235981196165085,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 4.1808600425720215,
+ "step": 1466
+ },
+ {
+ "epoch": 20.376963350785342,
+ "grad_norm": 0.04902610555291176,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 4.138566017150879,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390924956369982,
+ "grad_norm": 0.04923798143863678,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 4.135126113891602,
+ "step": 1468
+ },
+ {
+ "epoch": 20.404886561954626,
+ "grad_norm": 0.05059202015399933,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 4.1548285484313965,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418848167539267,
+ "grad_norm": 0.05058307573199272,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 4.149041652679443,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43280977312391,
+ "grad_norm": 0.050172705203294754,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 4.148054122924805,
+ "step": 1471
+ },
+ {
+ "epoch": 20.44677137870855,
+ "grad_norm": 0.05158299580216408,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 4.1425018310546875,
+ "step": 1472
+ },
+ {
+ "epoch": 20.460732984293195,
+ "grad_norm": 0.052176836878061295,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 4.129712104797363,
+ "step": 1473
+ },
+ {
+ "epoch": 20.474694589877835,
+ "grad_norm": 0.052095964550971985,
+ "learning_rate": 0.000401595976318565,
+ "loss": 4.134920597076416,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48865619546248,
+ "grad_norm": 0.04944197088479996,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 4.165334701538086,
+ "step": 1475
+ },
+ {
+ "epoch": 20.50261780104712,
+ "grad_norm": 0.049351636320352554,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 4.144362926483154,
+ "step": 1476
+ },
+ {
+ "epoch": 20.516579406631763,
+ "grad_norm": 0.053657080978155136,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 4.164234161376953,
+ "step": 1477
+ },
+ {
+ "epoch": 20.530541012216403,
+ "grad_norm": 0.052578140050172806,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 4.131514549255371,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544502617801047,
+ "grad_norm": 0.049683794379234314,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 4.129961967468262,
+ "step": 1479
+ },
+ {
+ "epoch": 20.55846422338569,
+ "grad_norm": 0.05115316063165665,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 4.164663314819336,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57242582897033,
+ "grad_norm": 0.051620446145534515,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 4.161067485809326,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586387434554975,
+ "grad_norm": 0.04873501881957054,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 4.161318302154541,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600349040139616,
+ "grad_norm": 0.04968130216002464,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 4.151819229125977,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61431064572426,
+ "grad_norm": 0.05476681888103485,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 4.14247989654541,
+ "step": 1484
+ },
+ {
+ "epoch": 20.6282722513089,
+ "grad_norm": 0.05098642408847809,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 4.178083419799805,
+ "step": 1485
+ },
+ {
+ "epoch": 20.642233856893544,
+ "grad_norm": 0.047264356166124344,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 4.15170955657959,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656195462478184,
+ "grad_norm": 0.050382886081933975,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 4.166364669799805,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670157068062828,
+ "grad_norm": 0.04956239461898804,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 4.164124488830566,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68411867364747,
+ "grad_norm": 0.049608584493398666,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 4.130239486694336,
+ "step": 1489
+ },
+ {
+ "epoch": 20.698080279232112,
+ "grad_norm": 0.0489494763314724,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 4.163503646850586,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712041884816752,
+ "grad_norm": 0.04847714304924011,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 4.1744585037231445,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726003490401396,
+ "grad_norm": 0.050256017595529556,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 4.1502227783203125,
+ "step": 1492
+ },
+ {
+ "epoch": 20.739965095986037,
+ "grad_norm": 0.048107463866472244,
+ "learning_rate": 0.000396803974909638,
+ "loss": 4.1790971755981445,
+ "step": 1493
+ },
+ {
+ "epoch": 20.75392670157068,
+ "grad_norm": 0.04828944429755211,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 4.128883361816406,
+ "step": 1494
+ },
+ {
+ "epoch": 20.767888307155324,
+ "grad_norm": 0.04720750078558922,
+ "learning_rate": 0.000396297942393265,
+ "loss": 4.1661577224731445,
+ "step": 1495
+ },
+ {
+ "epoch": 20.781849912739965,
+ "grad_norm": 0.047148339450359344,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 4.172522068023682,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79581151832461,
+ "grad_norm": 0.04612366110086441,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 4.149024963378906,
+ "step": 1497
+ },
+ {
+ "epoch": 20.80977312390925,
+ "grad_norm": 0.048896901309490204,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 4.179235458374023,
+ "step": 1498
+ },
+ {
+ "epoch": 20.823734729493893,
+ "grad_norm": 0.04779870808124542,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 4.1649932861328125,
+ "step": 1499
+ },
+ {
+ "epoch": 20.837696335078533,
+ "grad_norm": 0.04949707165360451,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 4.17645263671875,
+ "step": 1500
+ },
+ {
+ "epoch": 20.851657940663177,
+ "grad_norm": 0.04645023122429848,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 4.216496467590332,
+ "step": 1501
+ },
+ {
+ "epoch": 20.865619546247817,
+ "grad_norm": 0.046520136296749115,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 4.152667999267578,
+ "step": 1502
+ },
+ {
+ "epoch": 20.87958115183246,
+ "grad_norm": 0.04882653057575226,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 4.187082767486572,
+ "step": 1503
+ },
+ {
+ "epoch": 20.8935427574171,
+ "grad_norm": 0.04669157415628433,
+ "learning_rate": 0.000394017098798633,
+ "loss": 4.172348499298096,
+ "step": 1504
+ },
+ {
+ "epoch": 20.907504363001745,
+ "grad_norm": 0.04595336690545082,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 4.180559158325195,
+ "step": 1505
+ },
+ {
+ "epoch": 20.921465968586386,
+ "grad_norm": 0.048478737473487854,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 4.142880439758301,
+ "step": 1506
+ },
+ {
+ "epoch": 20.93542757417103,
+ "grad_norm": 0.04609905555844307,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 4.1459641456604,
+ "step": 1507
+ },
+ {
+ "epoch": 20.949389179755673,
+ "grad_norm": 0.04524408280849457,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 4.185099124908447,
+ "step": 1508
+ },
+ {
+ "epoch": 20.963350785340314,
+ "grad_norm": 0.0475911870598793,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 4.2085161209106445,
+ "step": 1509
+ },
+ {
+ "epoch": 20.977312390924958,
+ "grad_norm": 0.047796521335840225,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 4.165589332580566,
+ "step": 1510
+ },
+ {
+ "epoch": 20.991273996509598,
+ "grad_norm": 0.047358278185129166,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 4.1634368896484375,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.03328625112771988,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 2.6023788452148438,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.5952173471450806,
+ "eval_runtime": 58.6816,
+ "eval_samples_per_second": 41.614,
+ "eval_steps_per_second": 0.665,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013961605584644,
+ "grad_norm": 0.04970651492476463,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 4.077130317687988,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027923211169284,
+ "grad_norm": 0.060882844030857086,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 4.074376106262207,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041884816753928,
+ "grad_norm": 0.06609825044870377,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 4.107635021209717,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05584642233857,
+ "grad_norm": 0.056066352874040604,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 4.068084716796875,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069808027923212,
+ "grad_norm": 0.053013063967227936,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 4.069967269897461,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083769633507853,
+ "grad_norm": 0.059536613523960114,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 4.0639142990112305,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097731239092496,
+ "grad_norm": 0.06084490567445755,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 4.093890190124512,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111692844677137,
+ "grad_norm": 0.05299340933561325,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 4.100567817687988,
+ "step": 1520
+ },
+ {
+ "epoch": 21.12565445026178,
+ "grad_norm": 0.05136161297559738,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 4.08171272277832,
+ "step": 1521
+ },
+ {
+ "epoch": 21.13961605584642,
+ "grad_norm": 0.05201329290866852,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 4.0869975090026855,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153577661431065,
+ "grad_norm": 0.053958263248205185,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 4.0928850173950195,
+ "step": 1523
+ },
+ {
+ "epoch": 21.167539267015705,
+ "grad_norm": 0.053465038537979126,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 4.097828388214111,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18150087260035,
+ "grad_norm": 0.050980277359485626,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 4.101456165313721,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195462478184993,
+ "grad_norm": 0.05415358766913414,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 4.09096097946167,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209424083769633,
+ "grad_norm": 0.05567048117518425,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 4.061287879943848,
+ "step": 1527
+ },
+ {
+ "epoch": 21.223385689354277,
+ "grad_norm": 0.052690375596284866,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 4.080180644989014,
+ "step": 1528
+ },
+ {
+ "epoch": 21.237347294938917,
+ "grad_norm": 0.05448971316218376,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 4.079794883728027,
+ "step": 1529
+ },
+ {
+ "epoch": 21.25130890052356,
+ "grad_norm": 0.055472876876592636,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 4.072025299072266,
+ "step": 1530
+ },
+ {
+ "epoch": 21.2652705061082,
+ "grad_norm": 0.05418326333165169,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 4.103858470916748,
+ "step": 1531
+ },
+ {
+ "epoch": 21.279232111692846,
+ "grad_norm": 0.0524221733212471,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 4.0808916091918945,
+ "step": 1532
+ },
+ {
+ "epoch": 21.293193717277486,
+ "grad_norm": 0.05224563553929329,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 4.090270042419434,
+ "step": 1533
+ },
+ {
+ "epoch": 21.30715532286213,
+ "grad_norm": 0.052917953580617905,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 4.081887245178223,
+ "step": 1534
+ },
+ {
+ "epoch": 21.32111692844677,
+ "grad_norm": 0.05619358271360397,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 4.089571475982666,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335078534031414,
+ "grad_norm": 0.05455326288938522,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 4.096992015838623,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349040139616054,
+ "grad_norm": 0.049311455339193344,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 4.114621639251709,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363001745200698,
+ "grad_norm": 0.05175276845693588,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 4.107684135437012,
+ "step": 1538
+ },
+ {
+ "epoch": 21.376963350785342,
+ "grad_norm": 0.05236738547682762,
+ "learning_rate": 0.00038509205478744,
+ "loss": 4.075709342956543,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390924956369982,
+ "grad_norm": 0.04899081960320473,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 4.108396530151367,
+ "step": 1540
+ },
+ {
+ "epoch": 21.404886561954626,
+ "grad_norm": 0.05129663646221161,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 4.1135711669921875,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418848167539267,
+ "grad_norm": 0.05478592589497566,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 4.11839485168457,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43280977312391,
+ "grad_norm": 0.055830392986536026,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 4.11065149307251,
+ "step": 1543
+ },
+ {
+ "epoch": 21.44677137870855,
+ "grad_norm": 0.05370553582906723,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 4.113103866577148,
+ "step": 1544
+ },
+ {
+ "epoch": 21.460732984293195,
+ "grad_norm": 0.04968908056616783,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 4.119443893432617,
+ "step": 1545
+ },
+ {
+ "epoch": 21.474694589877835,
+ "grad_norm": 0.056149110198020935,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 4.126479625701904,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48865619546248,
+ "grad_norm": 0.053045690059661865,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 4.138535499572754,
+ "step": 1547
+ },
+ {
+ "epoch": 21.50261780104712,
+ "grad_norm": 0.04864223673939705,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 4.099347114562988,
+ "step": 1548
+ },
+ {
+ "epoch": 21.516579406631763,
+ "grad_norm": 0.05234768986701965,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 4.112222194671631,
+ "step": 1549
+ },
+ {
+ "epoch": 21.530541012216403,
+ "grad_norm": 0.050486158579587936,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 4.107913970947266,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544502617801047,
+ "grad_norm": 0.048641227185726166,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 4.114014148712158,
+ "step": 1551
+ },
+ {
+ "epoch": 21.55846422338569,
+ "grad_norm": 0.047759316861629486,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 4.128541946411133,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57242582897033,
+ "grad_norm": 0.046098388731479645,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 4.12142276763916,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586387434554975,
+ "grad_norm": 0.04934769123792648,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 4.13035249710083,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600349040139616,
+ "grad_norm": 0.048911672085523605,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 4.111344337463379,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61431064572426,
+ "grad_norm": 0.049368683248758316,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 4.1345672607421875,
+ "step": 1556
+ },
+ {
+ "epoch": 21.6282722513089,
+ "grad_norm": 0.04890736937522888,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 4.1286821365356445,
+ "step": 1557
+ },
+ {
+ "epoch": 21.642233856893544,
+ "grad_norm": 0.04812583327293396,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 4.09993314743042,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656195462478184,
+ "grad_norm": 0.05025152117013931,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 4.150769233703613,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670157068062828,
+ "grad_norm": 0.05007397383451462,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 4.074525833129883,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68411867364747,
+ "grad_norm": 0.050168465822935104,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 4.0933732986450195,
+ "step": 1561
+ },
+ {
+ "epoch": 21.698080279232112,
+ "grad_norm": 0.04990607872605324,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 4.126741886138916,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712041884816752,
+ "grad_norm": 0.05023440346121788,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 4.120314598083496,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726003490401396,
+ "grad_norm": 0.048100546002388,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 4.1242356300354,
+ "step": 1564
+ },
+ {
+ "epoch": 21.739965095986037,
+ "grad_norm": 0.047999415546655655,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 4.127745628356934,
+ "step": 1565
+ },
+ {
+ "epoch": 21.75392670157068,
+ "grad_norm": 0.05180094763636589,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 4.112565994262695,
+ "step": 1566
+ },
+ {
+ "epoch": 21.767888307155324,
+ "grad_norm": 0.053367964923381805,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 4.109419822692871,
+ "step": 1567
+ },
+ {
+ "epoch": 21.781849912739965,
+ "grad_norm": 0.05205400660634041,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 4.111886024475098,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79581151832461,
+ "grad_norm": 0.0464242659509182,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 4.104648590087891,
+ "step": 1569
+ },
+ {
+ "epoch": 21.80977312390925,
+ "grad_norm": 0.047767721116542816,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 4.156681060791016,
+ "step": 1570
+ },
+ {
+ "epoch": 21.823734729493893,
+ "grad_norm": 0.05028790235519409,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 4.132200241088867,
+ "step": 1571
+ },
+ {
+ "epoch": 21.837696335078533,
+ "grad_norm": 0.04714380204677582,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 4.132688999176025,
+ "step": 1572
+ },
+ {
+ "epoch": 21.851657940663177,
+ "grad_norm": 0.04726320132613182,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 4.173277378082275,
+ "step": 1573
+ },
+ {
+ "epoch": 21.865619546247817,
+ "grad_norm": 0.049558594822883606,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 4.15473747253418,
+ "step": 1574
+ },
+ {
+ "epoch": 21.87958115183246,
+ "grad_norm": 0.04742267727851868,
+ "learning_rate": 0.000375827577564042,
+ "loss": 4.10634708404541,
+ "step": 1575
+ },
+ {
+ "epoch": 21.8935427574171,
+ "grad_norm": 0.048223357647657394,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 4.133949279785156,
+ "step": 1576
+ },
+ {
+ "epoch": 21.907504363001745,
+ "grad_norm": 0.046766068786382675,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 4.153807640075684,
+ "step": 1577
+ },
+ {
+ "epoch": 21.921465968586386,
+ "grad_norm": 0.04586656019091606,
+ "learning_rate": 0.000375051971546236,
+ "loss": 4.169903755187988,
+ "step": 1578
+ },
+ {
+ "epoch": 21.93542757417103,
+ "grad_norm": 0.047912538051605225,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 4.148102283477783,
+ "step": 1579
+ },
+ {
+ "epoch": 21.949389179755673,
+ "grad_norm": 0.048949215561151505,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 4.149707317352295,
+ "step": 1580
+ },
+ {
+ "epoch": 21.963350785340314,
+ "grad_norm": 0.04765952378511429,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 4.128008842468262,
+ "step": 1581
+ },
+ {
+ "epoch": 21.977312390924958,
+ "grad_norm": 0.04726482927799225,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 4.1268720626831055,
+ "step": 1582
+ },
+ {
+ "epoch": 21.991273996509598,
+ "grad_norm": 0.0468299426138401,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 4.1332926750183105,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.03415689244866371,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 2.5875940322875977,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.5966277122497559,
+ "eval_runtime": 59.7704,
+ "eval_samples_per_second": 40.856,
+ "eval_steps_per_second": 0.652,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013961605584644,
+ "grad_norm": 0.04954535514116287,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 4.054527759552002,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027923211169284,
+ "grad_norm": 0.0498427115380764,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 4.048842430114746,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041884816753928,
+ "grad_norm": 0.051013074815273285,
+ "learning_rate": 0.000372722041257003,
+ "loss": 4.036187171936035,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05584642233857,
+ "grad_norm": 0.04973271116614342,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 4.050318717956543,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069808027923212,
+ "grad_norm": 0.051144856959581375,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 4.036319732666016,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083769633507853,
+ "grad_norm": 0.051874030381441116,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 4.017857551574707,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097731239092496,
+ "grad_norm": 0.05181736871600151,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 4.0519819259643555,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111692844677137,
+ "grad_norm": 0.05467049032449722,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 4.064291954040527,
+ "step": 1592
+ },
+ {
+ "epoch": 22.12565445026178,
+ "grad_norm": 0.05251661688089371,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 4.056951522827148,
+ "step": 1593
+ },
+ {
+ "epoch": 22.13961605584642,
+ "grad_norm": 0.05278680846095085,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 4.037125587463379,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153577661431065,
+ "grad_norm": 0.05276667699217796,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 4.053675174713135,
+ "step": 1595
+ },
+ {
+ "epoch": 22.167539267015705,
+ "grad_norm": 0.05007271468639374,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 4.044122695922852,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18150087260035,
+ "grad_norm": 0.051073770970106125,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 4.0265960693359375,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195462478184993,
+ "grad_norm": 0.05045647919178009,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 4.060938835144043,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209424083769633,
+ "grad_norm": 0.05075158178806305,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 4.035902976989746,
+ "step": 1599
+ },
+ {
+ "epoch": 22.223385689354277,
+ "grad_norm": 0.04867244139313698,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 4.055220603942871,
+ "step": 1600
+ },
+ {
+ "epoch": 22.237347294938917,
+ "grad_norm": 0.049976419657468796,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 4.026124954223633,
+ "step": 1601
+ },
+ {
+ "epoch": 22.25130890052356,
+ "grad_norm": 0.04968957230448723,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 4.081721782684326,
+ "step": 1602
+ },
+ {
+ "epoch": 22.2652705061082,
+ "grad_norm": 0.0503058098256588,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 4.0384721755981445,
+ "step": 1603
+ },
+ {
+ "epoch": 22.279232111692846,
+ "grad_norm": 0.050643209367990494,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 4.049397945404053,
+ "step": 1604
+ },
+ {
+ "epoch": 22.293193717277486,
+ "grad_norm": 0.05297046899795532,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 4.046389579772949,
+ "step": 1605
+ },
+ {
+ "epoch": 22.30715532286213,
+ "grad_norm": 0.053165238350629807,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 4.090221405029297,
+ "step": 1606
+ },
+ {
+ "epoch": 22.32111692844677,
+ "grad_norm": 0.05104832351207733,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 4.023227691650391,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335078534031414,
+ "grad_norm": 0.048736415803432465,
+ "learning_rate": 0.000367268105959126,
+ "loss": 4.07478666305542,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349040139616054,
+ "grad_norm": 0.05190873518586159,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 4.068147659301758,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363001745200698,
+ "grad_norm": 0.05409928783774376,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 4.060364723205566,
+ "step": 1610
+ },
+ {
+ "epoch": 22.376963350785342,
+ "grad_norm": 0.05112519487738609,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 4.072710990905762,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390924956369982,
+ "grad_norm": 0.04818312078714371,
+ "learning_rate": 0.000366226612810024,
+ "loss": 4.042757034301758,
+ "step": 1612
+ },
+ {
+ "epoch": 22.404886561954626,
+ "grad_norm": 0.050733599811792374,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 4.064810752868652,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418848167539267,
+ "grad_norm": 0.054070211946964264,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 4.054926872253418,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43280977312391,
+ "grad_norm": 0.05459332838654518,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 4.0652923583984375,
+ "step": 1615
+ },
+ {
+ "epoch": 22.44677137870855,
+ "grad_norm": 0.05064154416322708,
+ "learning_rate": 0.000365184304613104,
+ "loss": 4.076216220855713,
+ "step": 1616
+ },
+ {
+ "epoch": 22.460732984293195,
+ "grad_norm": 0.05071895569562912,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 4.070466041564941,
+ "step": 1617
+ },
+ {
+ "epoch": 22.474694589877835,
+ "grad_norm": 0.05049987882375717,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 4.062691688537598,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48865619546248,
+ "grad_norm": 0.05174451321363449,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 4.07501220703125,
+ "step": 1619
+ },
+ {
+ "epoch": 22.50261780104712,
+ "grad_norm": 0.05360838770866394,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 4.054711818695068,
+ "step": 1620
+ },
+ {
+ "epoch": 22.516579406631763,
+ "grad_norm": 0.051809027791023254,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 4.075811386108398,
+ "step": 1621
+ },
+ {
+ "epoch": 22.530541012216403,
+ "grad_norm": 0.05586795508861542,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 4.0638203620910645,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544502617801047,
+ "grad_norm": 0.05571979284286499,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 4.089414596557617,
+ "step": 1623
+ },
+ {
+ "epoch": 22.55846422338569,
+ "grad_norm": 0.05098723992705345,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 4.091159820556641,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57242582897033,
+ "grad_norm": 0.05359277129173279,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 4.060613632202148,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586387434554975,
+ "grad_norm": 0.05281100794672966,
+ "learning_rate": 0.000362575056487632,
+ "loss": 4.08416748046875,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600349040139616,
+ "grad_norm": 0.053913842886686325,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 4.088703155517578,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61431064572426,
+ "grad_norm": 0.05476798862218857,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 4.064074516296387,
+ "step": 1628
+ },
+ {
+ "epoch": 22.6282722513089,
+ "grad_norm": 0.051888495683670044,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 4.116634368896484,
+ "step": 1629
+ },
+ {
+ "epoch": 22.642233856893544,
+ "grad_norm": 0.051054876297712326,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 4.0800018310546875,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656195462478184,
+ "grad_norm": 0.054063159972429276,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 4.112773418426514,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670157068062828,
+ "grad_norm": 0.05173034593462944,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 4.062599182128906,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68411867364747,
+ "grad_norm": 0.054618336260318756,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 4.110188961029053,
+ "step": 1633
+ },
+ {
+ "epoch": 22.698080279232112,
+ "grad_norm": 0.055602896958589554,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 4.140998840332031,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712041884816752,
+ "grad_norm": 0.05332222580909729,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 4.119804382324219,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726003490401396,
+ "grad_norm": 0.05162122845649719,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 4.0731282234191895,
+ "step": 1636
+ },
+ {
+ "epoch": 22.739965095986037,
+ "grad_norm": 0.0514204278588295,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 4.068007946014404,
+ "step": 1637
+ },
+ {
+ "epoch": 22.75392670157068,
+ "grad_norm": 0.05027637258172035,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 4.097658157348633,
+ "step": 1638
+ },
+ {
+ "epoch": 22.767888307155324,
+ "grad_norm": 0.051812347024679184,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 4.083312511444092,
+ "step": 1639
+ },
+ {
+ "epoch": 22.781849912739965,
+ "grad_norm": 0.05145815759897232,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 4.129471778869629,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79581151832461,
+ "grad_norm": 0.05093684792518616,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 4.091841220855713,
+ "step": 1641
+ },
+ {
+ "epoch": 22.80977312390925,
+ "grad_norm": 0.05100737139582634,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 4.076615810394287,
+ "step": 1642
+ },
+ {
+ "epoch": 22.823734729493893,
+ "grad_norm": 0.05079149082303047,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 4.067737579345703,
+ "step": 1643
+ },
+ {
+ "epoch": 22.837696335078533,
+ "grad_norm": 0.05044509842991829,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 4.118824481964111,
+ "step": 1644
+ },
+ {
+ "epoch": 22.851657940663177,
+ "grad_norm": 0.05010584369301796,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 4.113405227661133,
+ "step": 1645
+ },
+ {
+ "epoch": 22.865619546247817,
+ "grad_norm": 0.052232302725315094,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 4.101553440093994,
+ "step": 1646
+ },
+ {
+ "epoch": 22.87958115183246,
+ "grad_norm": 0.05080661550164223,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 4.116809844970703,
+ "step": 1647
+ },
+ {
+ "epoch": 22.8935427574171,
+ "grad_norm": 0.05051033943891525,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 4.128484725952148,
+ "step": 1648
+ },
+ {
+ "epoch": 22.907504363001745,
+ "grad_norm": 0.051927294582128525,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 4.092285633087158,
+ "step": 1649
+ },
+ {
+ "epoch": 22.921465968586386,
+ "grad_norm": 0.049063827842473984,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 4.088473796844482,
+ "step": 1650
+ },
+ {
+ "epoch": 22.93542757417103,
+ "grad_norm": 0.05048854276537895,
+ "learning_rate": 0.000356031397755195,
+ "loss": 4.081521987915039,
+ "step": 1651
+ },
+ {
+ "epoch": 22.949389179755673,
+ "grad_norm": 0.049307651817798615,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 4.122158050537109,
+ "step": 1652
+ },
+ {
+ "epoch": 22.963350785340314,
+ "grad_norm": 0.048311345279216766,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 4.100488662719727,
+ "step": 1653
+ },
+ {
+ "epoch": 22.977312390924958,
+ "grad_norm": 0.05016734451055527,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 4.088587760925293,
+ "step": 1654
+ },
+ {
+ "epoch": 22.991273996509598,
+ "grad_norm": 0.04944450408220291,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 4.114686012268066,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.03592754527926445,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 2.545376777648926,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.5981184840202332,
+ "eval_runtime": 59.9558,
+ "eval_samples_per_second": 40.73,
+ "eval_steps_per_second": 0.65,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013961605584644,
+ "grad_norm": 0.05112785845994949,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 4.00495719909668,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027923211169284,
+ "grad_norm": 0.05807499587535858,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 4.015654563903809,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041884816753928,
+ "grad_norm": 0.05383148789405823,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.979349136352539,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05584642233857,
+ "grad_norm": 0.05221369490027428,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 4.009493827819824,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069808027923212,
+ "grad_norm": 0.05144324526190758,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 4.045680999755859,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083769633507853,
+ "grad_norm": 0.051422782242298126,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.994011878967285,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097731239092496,
+ "grad_norm": 0.05142863467335701,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 4.03209114074707,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111692844677137,
+ "grad_norm": 0.050954073667526245,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 4.00391960144043,
+ "step": 1664
+ },
+ {
+ "epoch": 23.12565445026178,
+ "grad_norm": 0.05365024879574776,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.9839367866516113,
+ "step": 1665
+ },
+ {
+ "epoch": 23.13961605584642,
+ "grad_norm": 0.051181286573410034,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 4.017707824707031,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153577661431065,
+ "grad_norm": 0.054279860109090805,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.989722490310669,
+ "step": 1667
+ },
+ {
+ "epoch": 23.167539267015705,
+ "grad_norm": 0.052196286618709564,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 4.014932155609131,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18150087260035,
+ "grad_norm": 0.05098043382167816,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 4.026510238647461,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195462478184993,
+ "grad_norm": 0.05241888761520386,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.979708433151245,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209424083769633,
+ "grad_norm": 0.051801104098558426,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 4.029015064239502,
+ "step": 1671
+ },
+ {
+ "epoch": 23.223385689354277,
+ "grad_norm": 0.052666619420051575,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.9932432174682617,
+ "step": 1672
+ },
+ {
+ "epoch": 23.237347294938917,
+ "grad_norm": 0.054418303072452545,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 4.05676794052124,
+ "step": 1673
+ },
+ {
+ "epoch": 23.25130890052356,
+ "grad_norm": 0.055074021220207214,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.999929666519165,
+ "step": 1674
+ },
+ {
+ "epoch": 23.2652705061082,
+ "grad_norm": 0.05626136437058449,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 4.013473987579346,
+ "step": 1675
+ },
+ {
+ "epoch": 23.279232111692846,
+ "grad_norm": 0.054028257727622986,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 4.015802383422852,
+ "step": 1676
+ },
+ {
+ "epoch": 23.293193717277486,
+ "grad_norm": 0.054044198244810104,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 4.020464897155762,
+ "step": 1677
+ },
+ {
+ "epoch": 23.30715532286213,
+ "grad_norm": 0.05465259402990341,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 4.010867118835449,
+ "step": 1678
+ },
+ {
+ "epoch": 23.32111692844677,
+ "grad_norm": 0.05545585975050926,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 4.054412841796875,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335078534031414,
+ "grad_norm": 0.055265430361032486,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 4.020756244659424,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349040139616054,
+ "grad_norm": 0.053363267332315445,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 4.033337116241455,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363001745200698,
+ "grad_norm": 0.05347830802202225,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 4.024794578552246,
+ "step": 1682
+ },
+ {
+ "epoch": 23.376963350785342,
+ "grad_norm": 0.05496951565146446,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 4.013302803039551,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390924956369982,
+ "grad_norm": 0.05179804563522339,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 4.030914306640625,
+ "step": 1684
+ },
+ {
+ "epoch": 23.404886561954626,
+ "grad_norm": 0.054958634078502655,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 4.012727737426758,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418848167539267,
+ "grad_norm": 0.05533262714743614,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 4.0543365478515625,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43280977312391,
+ "grad_norm": 0.052329424768686295,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 4.031486511230469,
+ "step": 1687
+ },
+ {
+ "epoch": 23.44677137870855,
+ "grad_norm": 0.05273434892296791,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 4.040292739868164,
+ "step": 1688
+ },
+ {
+ "epoch": 23.460732984293195,
+ "grad_norm": 0.05445481091737747,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 4.025630950927734,
+ "step": 1689
+ },
+ {
+ "epoch": 23.474694589877835,
+ "grad_norm": 0.052102912217378616,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 4.049562454223633,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48865619546248,
+ "grad_norm": 0.051272716373205185,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 4.0431365966796875,
+ "step": 1691
+ },
+ {
+ "epoch": 23.50261780104712,
+ "grad_norm": 0.051311176270246506,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 4.042119026184082,
+ "step": 1692
+ },
+ {
+ "epoch": 23.516579406631763,
+ "grad_norm": 0.05042742192745209,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 4.044881820678711,
+ "step": 1693
+ },
+ {
+ "epoch": 23.530541012216403,
+ "grad_norm": 0.05103316903114319,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 4.033669471740723,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544502617801047,
+ "grad_norm": 0.05441150814294815,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 4.076536178588867,
+ "step": 1695
+ },
+ {
+ "epoch": 23.55846422338569,
+ "grad_norm": 0.05529651418328285,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 4.073678016662598,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57242582897033,
+ "grad_norm": 0.0526033453643322,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 4.067898750305176,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586387434554975,
+ "grad_norm": 0.052817948162555695,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 4.017403602600098,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600349040139616,
+ "grad_norm": 0.052461300045251846,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 4.026311874389648,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61431064572426,
+ "grad_norm": 0.05097689852118492,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 4.076074600219727,
+ "step": 1700
+ },
+ {
+ "epoch": 23.6282722513089,
+ "grad_norm": 0.05387786403298378,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 4.022763729095459,
+ "step": 1701
+ },
+ {
+ "epoch": 23.642233856893544,
+ "grad_norm": 0.05132855102419853,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 4.077712535858154,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656195462478184,
+ "grad_norm": 0.05380886048078537,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 4.038937091827393,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670157068062828,
+ "grad_norm": 0.053299430757761,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 4.063364028930664,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68411867364747,
+ "grad_norm": 0.051913514733314514,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 4.05495548248291,
+ "step": 1705
+ },
+ {
+ "epoch": 23.698080279232112,
+ "grad_norm": 0.05660073831677437,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 4.057949066162109,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712041884816752,
+ "grad_norm": 0.054757390171289444,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 4.05377721786499,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726003490401396,
+ "grad_norm": 0.0517287515103817,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 4.035928726196289,
+ "step": 1708
+ },
+ {
+ "epoch": 23.739965095986037,
+ "grad_norm": 0.05187627300620079,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 4.043362617492676,
+ "step": 1709
+ },
+ {
+ "epoch": 23.75392670157068,
+ "grad_norm": 0.053502220660448074,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 4.007593154907227,
+ "step": 1710
+ },
+ {
+ "epoch": 23.767888307155324,
+ "grad_norm": 0.05755629390478134,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 4.076575756072998,
+ "step": 1711
+ },
+ {
+ "epoch": 23.781849912739965,
+ "grad_norm": 0.05619359388947487,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 4.072240352630615,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79581151832461,
+ "grad_norm": 0.05604318529367447,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 4.0486650466918945,
+ "step": 1713
+ },
+ {
+ "epoch": 23.80977312390925,
+ "grad_norm": 0.05630200728774071,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 4.086282730102539,
+ "step": 1714
+ },
+ {
+ "epoch": 23.823734729493893,
+ "grad_norm": 0.053262822329998016,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 4.089612007141113,
+ "step": 1715
+ },
+ {
+ "epoch": 23.837696335078533,
+ "grad_norm": 0.05699621140956879,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 4.053062438964844,
+ "step": 1716
+ },
+ {
+ "epoch": 23.851657940663177,
+ "grad_norm": 0.05963658541440964,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 4.072884559631348,
+ "step": 1717
+ },
+ {
+ "epoch": 23.865619546247817,
+ "grad_norm": 0.0547492615878582,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 4.066218852996826,
+ "step": 1718
+ },
+ {
+ "epoch": 23.87958115183246,
+ "grad_norm": 0.054908450692892075,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 4.035344123840332,
+ "step": 1719
+ },
+ {
+ "epoch": 23.8935427574171,
+ "grad_norm": 0.055191367864608765,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 4.08471155166626,
+ "step": 1720
+ },
+ {
+ "epoch": 23.907504363001745,
+ "grad_norm": 0.05334226414561272,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 4.076174736022949,
+ "step": 1721
+ },
+ {
+ "epoch": 23.921465968586386,
+ "grad_norm": 0.05579005926847458,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 4.05291223526001,
+ "step": 1722
+ },
+ {
+ "epoch": 23.93542757417103,
+ "grad_norm": 0.05464969575405121,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 4.07335090637207,
+ "step": 1723
+ },
+ {
+ "epoch": 23.949389179755673,
+ "grad_norm": 0.05240754038095474,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 4.0995073318481445,
+ "step": 1724
+ },
+ {
+ "epoch": 23.963350785340314,
+ "grad_norm": 0.05416056886315346,
+ "learning_rate": 0.000336518346307424,
+ "loss": 4.072793960571289,
+ "step": 1725
+ },
+ {
+ "epoch": 23.977312390924958,
+ "grad_norm": 0.05496834218502045,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 4.059562683105469,
+ "step": 1726
+ },
+ {
+ "epoch": 23.991273996509598,
+ "grad_norm": 0.05537887290120125,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 4.050772666931152,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.036677468568086624,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 2.5407257080078125,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.5995977520942688,
+ "eval_runtime": 60.2704,
+ "eval_samples_per_second": 40.517,
+ "eval_steps_per_second": 0.647,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013961605584644,
+ "grad_norm": 0.05583682656288147,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.9873995780944824,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027923211169284,
+ "grad_norm": 0.0605769157409668,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.999206066131592,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041884816753928,
+ "grad_norm": 0.057975225150585175,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.9923095703125,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05584642233857,
+ "grad_norm": 0.056104276329278946,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.946317672729492,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069808027923212,
+ "grad_norm": 0.05481491982936859,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.9973483085632324,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083769633507853,
+ "grad_norm": 0.05460227280855179,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.984617233276367,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097731239092496,
+ "grad_norm": 0.058481909334659576,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 4.0138773918151855,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111692844677137,
+ "grad_norm": 0.05672750994563103,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.9683098793029785,
+ "step": 1736
+ },
+ {
+ "epoch": 24.12565445026178,
+ "grad_norm": 0.05315157398581505,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.9909560680389404,
+ "step": 1737
+ },
+ {
+ "epoch": 24.13961605584642,
+ "grad_norm": 0.05565650761127472,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.954531192779541,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153577661431065,
+ "grad_norm": 0.05550679191946983,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.97672700881958,
+ "step": 1739
+ },
+ {
+ "epoch": 24.167539267015705,
+ "grad_norm": 0.05608576908707619,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.987941265106201,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18150087260035,
+ "grad_norm": 0.05266846343874931,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.97526216506958,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195462478184993,
+ "grad_norm": 0.05439139902591705,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 4.012054920196533,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209424083769633,
+ "grad_norm": 0.05473192036151886,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 4.022945404052734,
+ "step": 1743
+ },
+ {
+ "epoch": 24.223385689354277,
+ "grad_norm": 0.054737839847803116,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.981072425842285,
+ "step": 1744
+ },
+ {
+ "epoch": 24.237347294938917,
+ "grad_norm": 0.05696691572666168,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.9708170890808105,
+ "step": 1745
+ },
+ {
+ "epoch": 24.25130890052356,
+ "grad_norm": 0.05677401274442673,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.985995292663574,
+ "step": 1746
+ },
+ {
+ "epoch": 24.2652705061082,
+ "grad_norm": 0.05895088613033295,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.991887092590332,
+ "step": 1747
+ },
+ {
+ "epoch": 24.279232111692846,
+ "grad_norm": 0.05758961662650108,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.9695687294006348,
+ "step": 1748
+ },
+ {
+ "epoch": 24.293193717277486,
+ "grad_norm": 0.05174405500292778,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.959138870239258,
+ "step": 1749
+ },
+ {
+ "epoch": 24.30715532286213,
+ "grad_norm": 0.055284611880779266,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 4.012698173522949,
+ "step": 1750
+ },
+ {
+ "epoch": 24.32111692844677,
+ "grad_norm": 0.06042446941137314,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.9892425537109375,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335078534031414,
+ "grad_norm": 0.05489135906100273,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.9962222576141357,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349040139616054,
+ "grad_norm": 0.0542326495051384,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.9787094593048096,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363001745200698,
+ "grad_norm": 0.05612313374876976,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.9881339073181152,
+ "step": 1754
+ },
+ {
+ "epoch": 24.376963350785342,
+ "grad_norm": 0.05661709979176521,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 4.009440898895264,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390924956369982,
+ "grad_norm": 0.055537596344947815,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 4.00221061706543,
+ "step": 1756
+ },
+ {
+ "epoch": 24.404886561954626,
+ "grad_norm": 0.054210033267736435,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.9807987213134766,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418848167539267,
+ "grad_norm": 0.05847020074725151,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.960838794708252,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43280977312391,
+ "grad_norm": 0.05386386811733246,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 4.017860412597656,
+ "step": 1759
+ },
+ {
+ "epoch": 24.44677137870855,
+ "grad_norm": 0.056131865829229355,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.980543613433838,
+ "step": 1760
+ },
+ {
+ "epoch": 24.460732984293195,
+ "grad_norm": 0.05648239329457283,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.9948508739471436,
+ "step": 1761
+ },
+ {
+ "epoch": 24.474694589877835,
+ "grad_norm": 0.054138120263814926,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 4.002998352050781,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48865619546248,
+ "grad_norm": 0.05725586414337158,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 4.01864767074585,
+ "step": 1763
+ },
+ {
+ "epoch": 24.50261780104712,
+ "grad_norm": 0.05680665746331215,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 4.023143768310547,
+ "step": 1764
+ },
+ {
+ "epoch": 24.516579406631763,
+ "grad_norm": 0.052627697587013245,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 4.006389617919922,
+ "step": 1765
+ },
+ {
+ "epoch": 24.530541012216403,
+ "grad_norm": 0.053234782069921494,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 4.007419586181641,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544502617801047,
+ "grad_norm": 0.05756475403904915,
+ "learning_rate": 0.000325372061241796,
+ "loss": 4.0226945877075195,
+ "step": 1767
+ },
+ {
+ "epoch": 24.55846422338569,
+ "grad_norm": 0.055328067392110825,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 4.022767066955566,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57242582897033,
+ "grad_norm": 0.05378621071577072,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 4.005022048950195,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586387434554975,
+ "grad_norm": 0.056049644947052,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 4.025008201599121,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600349040139616,
+ "grad_norm": 0.057860117405653,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 4.003593444824219,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61431064572426,
+ "grad_norm": 0.056685276329517365,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.997868299484253,
+ "step": 1772
+ },
+ {
+ "epoch": 24.6282722513089,
+ "grad_norm": 0.053294967859983444,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 4.0211992263793945,
+ "step": 1773
+ },
+ {
+ "epoch": 24.642233856893544,
+ "grad_norm": 0.053967639803886414,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 4.022367477416992,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656195462478184,
+ "grad_norm": 0.052357546985149384,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 4.021792411804199,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670157068062828,
+ "grad_norm": 0.0547296479344368,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 4.008403778076172,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68411867364747,
+ "grad_norm": 0.056104280054569244,
+ "learning_rate": 0.000322712903929477,
+ "loss": 4.033827781677246,
+ "step": 1777
+ },
+ {
+ "epoch": 24.698080279232112,
+ "grad_norm": 0.055415406823158264,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 4.015371322631836,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712041884816752,
+ "grad_norm": 0.054506298154592514,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 4.0000152587890625,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726003490401396,
+ "grad_norm": 0.05468682199716568,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.9905481338500977,
+ "step": 1780
+ },
+ {
+ "epoch": 24.739965095986037,
+ "grad_norm": 0.05323269963264465,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 4.040729999542236,
+ "step": 1781
+ },
+ {
+ "epoch": 24.75392670157068,
+ "grad_norm": 0.052569612860679626,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 4.028299331665039,
+ "step": 1782
+ },
+ {
+ "epoch": 24.767888307155324,
+ "grad_norm": 0.052494753152132034,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 4.028860569000244,
+ "step": 1783
+ },
+ {
+ "epoch": 24.781849912739965,
+ "grad_norm": 0.05346379801630974,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 4.000349044799805,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79581151832461,
+ "grad_norm": 0.051543332636356354,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 4.016782760620117,
+ "step": 1785
+ },
+ {
+ "epoch": 24.80977312390925,
+ "grad_norm": 0.05330873280763626,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 4.024371147155762,
+ "step": 1786
+ },
+ {
+ "epoch": 24.823734729493893,
+ "grad_norm": 0.05306831747293472,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 4.0101213455200195,
+ "step": 1787
+ },
+ {
+ "epoch": 24.837696335078533,
+ "grad_norm": 0.05148417130112648,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 4.020768642425537,
+ "step": 1788
+ },
+ {
+ "epoch": 24.851657940663177,
+ "grad_norm": 0.050848908722400665,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 4.030246734619141,
+ "step": 1789
+ },
+ {
+ "epoch": 24.865619546247817,
+ "grad_norm": 0.05279014632105827,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 4.04910135269165,
+ "step": 1790
+ },
+ {
+ "epoch": 24.87958115183246,
+ "grad_norm": 0.0514594204723835,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 4.025290489196777,
+ "step": 1791
+ },
+ {
+ "epoch": 24.8935427574171,
+ "grad_norm": 0.05200032889842987,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 4.019560813903809,
+ "step": 1792
+ },
+ {
+ "epoch": 24.907504363001745,
+ "grad_norm": 0.052474118769168854,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 4.035177230834961,
+ "step": 1793
+ },
+ {
+ "epoch": 24.921465968586386,
+ "grad_norm": 0.05411219224333763,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.9920859336853027,
+ "step": 1794
+ },
+ {
+ "epoch": 24.93542757417103,
+ "grad_norm": 0.0543992817401886,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 4.017391204833984,
+ "step": 1795
+ },
+ {
+ "epoch": 24.949389179755673,
+ "grad_norm": 0.05187678337097168,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 4.042266368865967,
+ "step": 1796
+ },
+ {
+ "epoch": 24.963350785340314,
+ "grad_norm": 0.053762417286634445,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 4.029013633728027,
+ "step": 1797
+ },
+ {
+ "epoch": 24.977312390924958,
+ "grad_norm": 0.051723796874284744,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 4.038886070251465,
+ "step": 1798
+ },
+ {
+ "epoch": 24.991273996509598,
+ "grad_norm": 0.05242295190691948,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 4.056063652038574,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.03998562693595886,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 2.534770965576172,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 0.6009227633476257,
+ "eval_runtime": 58.9421,
+ "eval_samples_per_second": 41.43,
+ "eval_steps_per_second": 0.662,
+ "step": 1800
+ },
+ {
+ "epoch": 25.013961605584644,
+ "grad_norm": 0.05450186878442764,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 3.936816692352295,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027923211169284,
+ "grad_norm": 0.06389589607715607,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 3.9368391036987305,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041884816753928,
+ "grad_norm": 0.06028348580002785,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 3.953296661376953,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05584642233857,
+ "grad_norm": 0.05359262600541115,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 3.924558639526367,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069808027923212,
+ "grad_norm": 0.059837762266397476,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 3.9608402252197266,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083769633507853,
+ "grad_norm": 0.05523517727851868,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 3.943776845932007,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097731239092496,
+ "grad_norm": 0.05617433041334152,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 3.9635634422302246,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111692844677137,
+ "grad_norm": 0.06066269427537918,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 3.9647364616394043,
+ "step": 1808
+ },
+ {
+ "epoch": 25.12565445026178,
+ "grad_norm": 0.05491397902369499,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 3.95222806930542,
+ "step": 1809
+ },
+ {
+ "epoch": 25.13961605584642,
+ "grad_norm": 0.05731203034520149,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 3.9440197944641113,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153577661431065,
+ "grad_norm": 0.05428403243422508,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 3.9428441524505615,
+ "step": 1811
+ },
+ {
+ "epoch": 25.167539267015705,
+ "grad_norm": 0.05487453565001488,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 3.9489784240722656,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18150087260035,
+ "grad_norm": 0.05725884810090065,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 3.9643478393554688,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195462478184993,
+ "grad_norm": 0.05752462521195412,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 3.963301181793213,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209424083769633,
+ "grad_norm": 0.057387594133615494,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 3.9427294731140137,
+ "step": 1815
+ },
+ {
+ "epoch": 25.223385689354277,
+ "grad_norm": 0.057492151856422424,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 3.9229259490966797,
+ "step": 1816
+ },
+ {
+ "epoch": 25.237347294938917,
+ "grad_norm": 0.05202317237854004,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 3.938037395477295,
+ "step": 1817
+ },
+ {
+ "epoch": 25.25130890052356,
+ "grad_norm": 0.05574658885598183,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 3.9532949924468994,
+ "step": 1818
+ },
+ {
+ "epoch": 25.2652705061082,
+ "grad_norm": 0.05714752897620201,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 3.962207794189453,
+ "step": 1819
+ },
+ {
+ "epoch": 25.279232111692846,
+ "grad_norm": 0.05540093779563904,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 3.96329402923584,
+ "step": 1820
+ },
+ {
+ "epoch": 25.293193717277486,
+ "grad_norm": 0.056597232818603516,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 3.948793411254883,
+ "step": 1821
+ },
+ {
+ "epoch": 25.30715532286213,
+ "grad_norm": 0.055902402848005295,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 3.9111557006835938,
+ "step": 1822
+ },
+ {
+ "epoch": 25.32111692844677,
+ "grad_norm": 0.05509798601269722,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 3.929591655731201,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335078534031414,
+ "grad_norm": 0.05763949453830719,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 3.9758448600769043,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349040139616054,
+ "grad_norm": 0.053838349878787994,
+ "learning_rate": 0.000309930621834911,
+ "loss": 3.92457914352417,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363001745200698,
+ "grad_norm": 0.05467558279633522,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 3.9501285552978516,
+ "step": 1826
+ },
+ {
+ "epoch": 25.376963350785342,
+ "grad_norm": 0.05715392529964447,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 3.9734368324279785,
+ "step": 1827
+ },
+ {
+ "epoch": 25.390924956369982,
+ "grad_norm": 0.0566536970436573,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 3.9933152198791504,
+ "step": 1828
+ },
+ {
+ "epoch": 25.404886561954626,
+ "grad_norm": 0.053221236914396286,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 3.967787265777588,
+ "step": 1829
+ },
+ {
+ "epoch": 25.418848167539267,
+ "grad_norm": 0.05779031664133072,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 3.9415433406829834,
+ "step": 1830
+ },
+ {
+ "epoch": 25.43280977312391,
+ "grad_norm": 0.062013521790504456,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 3.9801456928253174,
+ "step": 1831
+ },
+ {
+ "epoch": 25.44677137870855,
+ "grad_norm": 0.05819004401564598,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 3.966928005218506,
+ "step": 1832
+ },
+ {
+ "epoch": 25.460732984293195,
+ "grad_norm": 0.05754564702510834,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 3.9612770080566406,
+ "step": 1833
+ },
+ {
+ "epoch": 25.474694589877835,
+ "grad_norm": 0.05707791447639465,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 3.935521125793457,
+ "step": 1834
+ },
+ {
+ "epoch": 25.48865619546248,
+ "grad_norm": 0.056380245834589005,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 4.006120681762695,
+ "step": 1835
+ },
+ {
+ "epoch": 25.50261780104712,
+ "grad_norm": 0.0551077239215374,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 3.9544754028320312,
+ "step": 1836
+ },
+ {
+ "epoch": 25.516579406631763,
+ "grad_norm": 0.05720360204577446,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 3.9564664363861084,
+ "step": 1837
+ },
+ {
+ "epoch": 25.530541012216403,
+ "grad_norm": 0.05522238463163376,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 4.0157880783081055,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544502617801047,
+ "grad_norm": 0.05304659530520439,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 3.957207202911377,
+ "step": 1839
+ },
+ {
+ "epoch": 25.55846422338569,
+ "grad_norm": 0.05569292604923248,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 3.9909608364105225,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57242582897033,
+ "grad_norm": 0.05347870662808418,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 3.9537010192871094,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586387434554975,
+ "grad_norm": 0.05572560802102089,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 3.9818642139434814,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600349040139616,
+ "grad_norm": 0.05624900385737419,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 3.995117425918579,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61431064572426,
+ "grad_norm": 0.05600332096219063,
+ "learning_rate": 0.000304866093757771,
+ "loss": 3.9480154514312744,
+ "step": 1844
+ },
+ {
+ "epoch": 25.6282722513089,
+ "grad_norm": 0.05488704890012741,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 3.993014335632324,
+ "step": 1845
+ },
+ {
+ "epoch": 25.642233856893544,
+ "grad_norm": 0.05647030100226402,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 3.994615077972412,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656195462478184,
+ "grad_norm": 0.05460594967007637,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 3.9714808464050293,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670157068062828,
+ "grad_norm": 0.05629213526844978,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 3.962033271789551,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68411867364747,
+ "grad_norm": 0.05666861683130264,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 3.995985269546509,
+ "step": 1849
+ },
+ {
+ "epoch": 25.698080279232112,
+ "grad_norm": 0.05580414459109306,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 4.001766681671143,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712041884816752,
+ "grad_norm": 0.05638023093342781,
+ "learning_rate": 0.000303,
+ "loss": 4.005331993103027,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726003490401396,
+ "grad_norm": 0.05684564262628555,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 4.017581939697266,
+ "step": 1852
+ },
+ {
+ "epoch": 25.739965095986037,
+ "grad_norm": 0.057602159678936005,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 3.9668312072753906,
+ "step": 1853
+ },
+ {
+ "epoch": 25.75392670157068,
+ "grad_norm": 0.055540211498737335,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 3.9885573387145996,
+ "step": 1854
+ },
+ {
+ "epoch": 25.767888307155324,
+ "grad_norm": 0.05655810981988907,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 3.990997314453125,
+ "step": 1855
+ },
+ {
+ "epoch": 25.781849912739965,
+ "grad_norm": 0.05443689972162247,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 3.9554123878479004,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79581151832461,
+ "grad_norm": 0.05570529028773308,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 3.983283519744873,
+ "step": 1857
+ },
+ {
+ "epoch": 25.80977312390925,
+ "grad_norm": 0.05399419739842415,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 4.004507064819336,
+ "step": 1858
+ },
+ {
+ "epoch": 25.823734729493893,
+ "grad_norm": 0.05587748438119888,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 4.00900936126709,
+ "step": 1859
+ },
+ {
+ "epoch": 25.837696335078533,
+ "grad_norm": 0.05465575307607651,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 3.9826579093933105,
+ "step": 1860
+ },
+ {
+ "epoch": 25.851657940663177,
+ "grad_norm": 0.05286499112844467,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 3.973453998565674,
+ "step": 1861
+ },
+ {
+ "epoch": 25.865619546247817,
+ "grad_norm": 0.05495590344071388,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 4.003095626831055,
+ "step": 1862
+ },
+ {
+ "epoch": 25.87958115183246,
+ "grad_norm": 0.053388018161058426,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 4.00042200088501,
+ "step": 1863
+ },
+ {
+ "epoch": 25.8935427574171,
+ "grad_norm": 0.05323237553238869,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 3.992929458618164,
+ "step": 1864
+ },
+ {
+ "epoch": 25.907504363001745,
+ "grad_norm": 0.053946711122989655,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 3.996519088745117,
+ "step": 1865
+ },
+ {
+ "epoch": 25.921465968586386,
+ "grad_norm": 0.05455862358212471,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 4.007123947143555,
+ "step": 1866
+ },
+ {
+ "epoch": 25.93542757417103,
+ "grad_norm": 0.05290355160832405,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 4.001724720001221,
+ "step": 1867
+ },
+ {
+ "epoch": 25.949389179755673,
+ "grad_norm": 0.052808016538619995,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 3.9964821338653564,
+ "step": 1868
+ },
+ {
+ "epoch": 25.963350785340314,
+ "grad_norm": 0.05416771396994591,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 3.98549222946167,
+ "step": 1869
+ },
+ {
+ "epoch": 25.977312390924958,
+ "grad_norm": 0.05264521390199661,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 3.9729766845703125,
+ "step": 1870
+ },
+ {
+ "epoch": 25.991273996509598,
+ "grad_norm": 0.05483962222933769,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 4.018877029418945,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.03927824646234512,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 2.456951856613159,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 0.6026532053947449,
+ "eval_runtime": 60.4079,
+ "eval_samples_per_second": 40.425,
+ "eval_steps_per_second": 0.646,
+ "step": 1872
+ },
+ {
+ "epoch": 26.013961605584644,
+ "grad_norm": 0.05756894499063492,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 3.93898606300354,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027923211169284,
+ "grad_norm": 0.06288876384496689,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 3.9142136573791504,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041884816753928,
+ "grad_norm": 0.0600322000682354,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 3.918700695037842,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05584642233857,
+ "grad_norm": 0.05729348957538605,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 3.9040162563323975,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069808027923212,
+ "grad_norm": 0.05658194422721863,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 3.8809690475463867,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083769633507853,
+ "grad_norm": 0.06274262815713882,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 3.928222179412842,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097731239092496,
+ "grad_norm": 0.06313030421733856,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 3.9395527839660645,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111692844677137,
+ "grad_norm": 0.0554068423807621,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 3.9314563274383545,
+ "step": 1880
+ },
+ {
+ "epoch": 26.12565445026178,
+ "grad_norm": 0.05697336047887802,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 3.9071288108825684,
+ "step": 1881
+ },
+ {
+ "epoch": 26.13961605584642,
+ "grad_norm": 0.05603424087166786,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 3.932054042816162,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153577661431065,
+ "grad_norm": 0.055603109300136566,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 3.9169063568115234,
+ "step": 1883
+ },
+ {
+ "epoch": 26.167539267015705,
+ "grad_norm": 0.055022306740283966,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 3.8985838890075684,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18150087260035,
+ "grad_norm": 0.05383468046784401,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 3.896359443664551,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195462478184993,
+ "grad_norm": 0.055076900869607925,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 3.905672073364258,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209424083769633,
+ "grad_norm": 0.05629882961511612,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 3.9222521781921387,
+ "step": 1887
+ },
+ {
+ "epoch": 26.223385689354277,
+ "grad_norm": 0.056736111640930176,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 3.8918867111206055,
+ "step": 1888
+ },
+ {
+ "epoch": 26.237347294938917,
+ "grad_norm": 0.05625338479876518,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 3.9249935150146484,
+ "step": 1889
+ },
+ {
+ "epoch": 26.25130890052356,
+ "grad_norm": 0.06145375967025757,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 3.904536724090576,
+ "step": 1890
+ },
+ {
+ "epoch": 26.2652705061082,
+ "grad_norm": 0.05601578950881958,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 3.8845863342285156,
+ "step": 1891
+ },
+ {
+ "epoch": 26.279232111692846,
+ "grad_norm": 0.056899815797805786,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 3.917966842651367,
+ "step": 1892
+ },
+ {
+ "epoch": 26.293193717277486,
+ "grad_norm": 0.060314927250146866,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 3.9230940341949463,
+ "step": 1893
+ },
+ {
+ "epoch": 26.30715532286213,
+ "grad_norm": 0.05449546128511429,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 3.9082441329956055,
+ "step": 1894
+ },
+ {
+ "epoch": 26.32111692844677,
+ "grad_norm": 0.05601051449775696,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 3.9174978733062744,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335078534031414,
+ "grad_norm": 0.05770742893218994,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 3.9183058738708496,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349040139616054,
+ "grad_norm": 0.05718272924423218,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 3.9438772201538086,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363001745200698,
+ "grad_norm": 0.052873238921165466,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 3.9554827213287354,
+ "step": 1898
+ },
+ {
+ "epoch": 26.376963350785342,
+ "grad_norm": 0.056853555142879486,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 3.886812210083008,
+ "step": 1899
+ },
+ {
+ "epoch": 26.390924956369982,
+ "grad_norm": 0.05662025138735771,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 3.9384284019470215,
+ "step": 1900
+ },
+ {
+ "epoch": 26.404886561954626,
+ "grad_norm": 0.05693329498171806,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 3.9123764038085938,
+ "step": 1901
+ },
+ {
+ "epoch": 26.418848167539267,
+ "grad_norm": 0.05425408110022545,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 3.9441678524017334,
+ "step": 1902
+ },
+ {
+ "epoch": 26.43280977312391,
+ "grad_norm": 0.05533420667052269,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 3.9086289405822754,
+ "step": 1903
+ },
+ {
+ "epoch": 26.44677137870855,
+ "grad_norm": 0.05527901649475098,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 3.9469594955444336,
+ "step": 1904
+ },
+ {
+ "epoch": 26.460732984293195,
+ "grad_norm": 0.052982039749622345,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 3.9182326793670654,
+ "step": 1905
+ },
+ {
+ "epoch": 26.474694589877835,
+ "grad_norm": 0.05520346015691757,
+ "learning_rate": 0.000288343693342466,
+ "loss": 3.954651117324829,
+ "step": 1906
+ },
+ {
+ "epoch": 26.48865619546248,
+ "grad_norm": 0.05564865842461586,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 3.951840400695801,
+ "step": 1907
+ },
+ {
+ "epoch": 26.50261780104712,
+ "grad_norm": 0.05588578060269356,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 3.933762550354004,
+ "step": 1908
+ },
+ {
+ "epoch": 26.516579406631763,
+ "grad_norm": 0.05727043375372887,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 3.9299421310424805,
+ "step": 1909
+ },
+ {
+ "epoch": 26.530541012216403,
+ "grad_norm": 0.05533654987812042,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 3.941525936126709,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544502617801047,
+ "grad_norm": 0.05524713918566704,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 3.935666084289551,
+ "step": 1911
+ },
+ {
+ "epoch": 26.55846422338569,
+ "grad_norm": 0.055435653775930405,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 3.951631546020508,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57242582897033,
+ "grad_norm": 0.05678093433380127,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 3.960881233215332,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586387434554975,
+ "grad_norm": 0.054896436631679535,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 3.9262163639068604,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600349040139616,
+ "grad_norm": 0.055213335901498795,
+ "learning_rate": 0.000285947841605349,
+ "loss": 3.9284324645996094,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61431064572426,
+ "grad_norm": 0.05644496902823448,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 3.9727559089660645,
+ "step": 1916
+ },
+ {
+ "epoch": 26.6282722513089,
+ "grad_norm": 0.05483568087220192,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 3.9410910606384277,
+ "step": 1917
+ },
+ {
+ "epoch": 26.642233856893544,
+ "grad_norm": 0.054636143147945404,
+ "learning_rate": 0.000285149463934261,
+ "loss": 3.945554733276367,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656195462478184,
+ "grad_norm": 0.056062210351228714,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 3.950249671936035,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670157068062828,
+ "grad_norm": 0.05734739080071449,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 3.9527649879455566,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68411867364747,
+ "grad_norm": 0.05562867224216461,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 3.9563350677490234,
+ "step": 1921
+ },
+ {
+ "epoch": 26.698080279232112,
+ "grad_norm": 0.05565498396754265,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 3.9478225708007812,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712041884816752,
+ "grad_norm": 0.057103481143713,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 3.9302189350128174,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726003490401396,
+ "grad_norm": 0.058241210877895355,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 3.9773902893066406,
+ "step": 1924
+ },
+ {
+ "epoch": 26.739965095986037,
+ "grad_norm": 0.055981557816267014,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 3.923896074295044,
+ "step": 1925
+ },
+ {
+ "epoch": 26.75392670157068,
+ "grad_norm": 0.05711016431450844,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 3.972334861755371,
+ "step": 1926
+ },
+ {
+ "epoch": 26.767888307155324,
+ "grad_norm": 0.05707632377743721,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 3.958393096923828,
+ "step": 1927
+ },
+ {
+ "epoch": 26.781849912739965,
+ "grad_norm": 0.055073633790016174,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 3.9449703693389893,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79581151832461,
+ "grad_norm": 0.05765518546104431,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 3.9442358016967773,
+ "step": 1929
+ },
+ {
+ "epoch": 26.80977312390925,
+ "grad_norm": 0.05692309886217117,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 3.9665679931640625,
+ "step": 1930
+ },
+ {
+ "epoch": 26.823734729493893,
+ "grad_norm": 0.05776236951351166,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 3.9153542518615723,
+ "step": 1931
+ },
+ {
+ "epoch": 26.837696335078533,
+ "grad_norm": 0.05465872213244438,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 3.919874429702759,
+ "step": 1932
+ },
+ {
+ "epoch": 26.851657940663177,
+ "grad_norm": 0.057827629148960114,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 3.9578394889831543,
+ "step": 1933
+ },
+ {
+ "epoch": 26.865619546247817,
+ "grad_norm": 0.055902253836393356,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 3.995711326599121,
+ "step": 1934
+ },
+ {
+ "epoch": 26.87958115183246,
+ "grad_norm": 0.05868638679385185,
+ "learning_rate": 0.000280627938758204,
+ "loss": 3.96519136428833,
+ "step": 1935
+ },
+ {
+ "epoch": 26.8935427574171,
+ "grad_norm": 0.059722673147916794,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 3.9566941261291504,
+ "step": 1936
+ },
+ {
+ "epoch": 26.907504363001745,
+ "grad_norm": 0.055209770798683167,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 3.9929540157318115,
+ "step": 1937
+ },
+ {
+ "epoch": 26.921465968586386,
+ "grad_norm": 0.05694293603301048,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 3.9529929161071777,
+ "step": 1938
+ },
+ {
+ "epoch": 26.93542757417103,
+ "grad_norm": 0.058492012321949005,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 3.9465792179107666,
+ "step": 1939
+ },
+ {
+ "epoch": 26.949389179755673,
+ "grad_norm": 0.056036073714494705,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 3.9590697288513184,
+ "step": 1940
+ },
+ {
+ "epoch": 26.963350785340314,
+ "grad_norm": 0.0547766387462616,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 3.969907522201538,
+ "step": 1941
+ },
+ {
+ "epoch": 26.977312390924958,
+ "grad_norm": 0.05592350661754608,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 3.9509897232055664,
+ "step": 1942
+ },
+ {
+ "epoch": 26.991273996509598,
+ "grad_norm": 0.05589321628212929,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 3.9648609161376953,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.03934016078710556,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 2.4784152507781982,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 0.6035652160644531,
+ "eval_runtime": 60.0656,
+ "eval_samples_per_second": 40.656,
+ "eval_steps_per_second": 0.649,
+ "step": 1944
+ },
+ {
+ "epoch": 27.013961605584644,
+ "grad_norm": 0.05773136392235756,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 3.8625826835632324,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027923211169284,
+ "grad_norm": 0.05844702199101448,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 3.8796658515930176,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041884816753928,
+ "grad_norm": 0.05718810483813286,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 3.8808112144470215,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05584642233857,
+ "grad_norm": 0.05264506861567497,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 3.8776121139526367,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069808027923212,
+ "grad_norm": 0.055468373000621796,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 3.866316318511963,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083769633507853,
+ "grad_norm": 0.053565360605716705,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 3.8870849609375,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097731239092496,
+ "grad_norm": 0.05468900501728058,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 3.8581550121307373,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111692844677137,
+ "grad_norm": 0.05454741418361664,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 3.8752541542053223,
+ "step": 1952
+ },
+ {
+ "epoch": 27.12565445026178,
+ "grad_norm": 0.05647914111614227,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 3.8818912506103516,
+ "step": 1953
+ },
+ {
+ "epoch": 27.13961605584642,
+ "grad_norm": 0.05325528234243393,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 3.8562560081481934,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153577661431065,
+ "grad_norm": 0.05581580847501755,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 3.879669189453125,
+ "step": 1955
+ },
+ {
+ "epoch": 27.167539267015705,
+ "grad_norm": 0.058219097554683685,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 3.8530635833740234,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18150087260035,
+ "grad_norm": 0.05613216012716293,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 3.8905630111694336,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195462478184993,
+ "grad_norm": 0.05486270785331726,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 3.8945157527923584,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209424083769633,
+ "grad_norm": 0.05338508263230324,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 3.909775733947754,
+ "step": 1959
+ },
+ {
+ "epoch": 27.223385689354277,
+ "grad_norm": 0.05602312833070755,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 3.8643651008605957,
+ "step": 1960
+ },
+ {
+ "epoch": 27.237347294938917,
+ "grad_norm": 0.05373037979006767,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 3.8915491104125977,
+ "step": 1961
+ },
+ {
+ "epoch": 27.25130890052356,
+ "grad_norm": 0.05805065110325813,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 3.8836984634399414,
+ "step": 1962
+ },
+ {
+ "epoch": 27.2652705061082,
+ "grad_norm": 0.05469847097992897,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 3.8812215328216553,
+ "step": 1963
+ },
+ {
+ "epoch": 27.279232111692846,
+ "grad_norm": 0.05575987696647644,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 3.8887271881103516,
+ "step": 1964
+ },
+ {
+ "epoch": 27.293193717277486,
+ "grad_norm": 0.05577889457345009,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 3.8917694091796875,
+ "step": 1965
+ },
+ {
+ "epoch": 27.30715532286213,
+ "grad_norm": 0.05604076758027077,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 3.881073474884033,
+ "step": 1966
+ },
+ {
+ "epoch": 27.32111692844677,
+ "grad_norm": 0.055674705654382706,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 3.9303488731384277,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335078534031414,
+ "grad_norm": 0.05300940200686455,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 3.871678113937378,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349040139616054,
+ "grad_norm": 0.05647176504135132,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 3.8848423957824707,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363001745200698,
+ "grad_norm": 0.05647560954093933,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 3.931574583053589,
+ "step": 1970
+ },
+ {
+ "epoch": 27.376963350785342,
+ "grad_norm": 0.055338893085718155,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 3.908456563949585,
+ "step": 1971
+ },
+ {
+ "epoch": 27.390924956369982,
+ "grad_norm": 0.05508219450712204,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 3.8933262825012207,
+ "step": 1972
+ },
+ {
+ "epoch": 27.404886561954626,
+ "grad_norm": 0.05787097290158272,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 3.8908863067626953,
+ "step": 1973
+ },
+ {
+ "epoch": 27.418848167539267,
+ "grad_norm": 0.05587272346019745,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 3.9261322021484375,
+ "step": 1974
+ },
+ {
+ "epoch": 27.43280977312391,
+ "grad_norm": 0.05573277547955513,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 3.9129767417907715,
+ "step": 1975
+ },
+ {
+ "epoch": 27.44677137870855,
+ "grad_norm": 0.05628194287419319,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 3.906001091003418,
+ "step": 1976
+ },
+ {
+ "epoch": 27.460732984293195,
+ "grad_norm": 0.055410996079444885,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 3.9163825511932373,
+ "step": 1977
+ },
+ {
+ "epoch": 27.474694589877835,
+ "grad_norm": 0.05579163134098053,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 3.9286627769470215,
+ "step": 1978
+ },
+ {
+ "epoch": 27.48865619546248,
+ "grad_norm": 0.05793057754635811,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 3.8955917358398438,
+ "step": 1979
+ },
+ {
+ "epoch": 27.50261780104712,
+ "grad_norm": 0.05592824146151543,
+ "learning_rate": 0.00026868712586269,
+ "loss": 3.922128677368164,
+ "step": 1980
+ },
+ {
+ "epoch": 27.516579406631763,
+ "grad_norm": 0.05768568441271782,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 3.897780418395996,
+ "step": 1981
+ },
+ {
+ "epoch": 27.530541012216403,
+ "grad_norm": 0.055304404348134995,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 3.886094570159912,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544502617801047,
+ "grad_norm": 0.054745838046073914,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 3.9272961616516113,
+ "step": 1983
+ },
+ {
+ "epoch": 27.55846422338569,
+ "grad_norm": 0.05616277828812599,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 3.8794174194335938,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57242582897033,
+ "grad_norm": 0.05605581775307655,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 3.9165992736816406,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586387434554975,
+ "grad_norm": 0.057485803961753845,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 3.902085542678833,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600349040139616,
+ "grad_norm": 0.0575849711894989,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 3.92405366897583,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61431064572426,
+ "grad_norm": 0.05563105270266533,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 3.8700900077819824,
+ "step": 1988
+ },
+ {
+ "epoch": 27.6282722513089,
+ "grad_norm": 0.05615481734275818,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 3.9526686668395996,
+ "step": 1989
+ },
+ {
+ "epoch": 27.642233856893544,
+ "grad_norm": 0.056746870279312134,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 3.9000678062438965,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656195462478184,
+ "grad_norm": 0.057345934212207794,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 3.910407543182373,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670157068062828,
+ "grad_norm": 0.05434483289718628,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 3.9132578372955322,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68411867364747,
+ "grad_norm": 0.0582718588411808,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 3.9188318252563477,
+ "step": 1993
+ },
+ {
+ "epoch": 27.698080279232112,
+ "grad_norm": 0.05566178634762764,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 3.8843226432800293,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712041884816752,
+ "grad_norm": 0.05550086498260498,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 3.904834032058716,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726003490401396,
+ "grad_norm": 0.05517689883708954,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 3.9082438945770264,
+ "step": 1996
+ },
+ {
+ "epoch": 27.739965095986037,
+ "grad_norm": 0.05598316714167595,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 3.909560203552246,
+ "step": 1997
+ },
+ {
+ "epoch": 27.75392670157068,
+ "grad_norm": 0.054991044104099274,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 3.903012275695801,
+ "step": 1998
+ },
+ {
+ "epoch": 27.767888307155324,
+ "grad_norm": 0.05612454190850258,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 3.9183692932128906,
+ "step": 1999
+ },
+ {
+ "epoch": 27.781849912739965,
+ "grad_norm": 0.055587366223335266,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 3.918454647064209,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79581151832461,
+ "grad_norm": 0.05595824122428894,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 3.895993232727051,
+ "step": 2001
+ },
+ {
+ "epoch": 27.80977312390925,
+ "grad_norm": 0.054730597883462906,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 3.9305458068847656,
+ "step": 2002
+ },
+ {
+ "epoch": 27.823734729493893,
+ "grad_norm": 0.05543222650885582,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 3.9375224113464355,
+ "step": 2003
+ },
+ {
+ "epoch": 27.837696335078533,
+ "grad_norm": 0.055638376623392105,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 3.9242658615112305,
+ "step": 2004
+ },
+ {
+ "epoch": 27.851657940663177,
+ "grad_norm": 0.05635111406445503,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 3.889681339263916,
+ "step": 2005
+ },
+ {
+ "epoch": 27.865619546247817,
+ "grad_norm": 0.056509099900722504,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 3.9191479682922363,
+ "step": 2006
+ },
+ {
+ "epoch": 27.87958115183246,
+ "grad_norm": 0.055455707013607025,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 3.936049222946167,
+ "step": 2007
+ },
+ {
+ "epoch": 27.8935427574171,
+ "grad_norm": 0.05508789047598839,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 3.901210069656372,
+ "step": 2008
+ },
+ {
+ "epoch": 27.907504363001745,
+ "grad_norm": 0.05542607232928276,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 3.9348654747009277,
+ "step": 2009
+ },
+ {
+ "epoch": 27.921465968586386,
+ "grad_norm": 0.05693232640624046,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 3.940847158432007,
+ "step": 2010
+ },
+ {
+ "epoch": 27.93542757417103,
+ "grad_norm": 0.05707051232457161,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 3.9295883178710938,
+ "step": 2011
+ },
+ {
+ "epoch": 27.949389179755673,
+ "grad_norm": 0.05778150260448456,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 3.9317305088043213,
+ "step": 2012
+ },
+ {
+ "epoch": 27.963350785340314,
+ "grad_norm": 0.05821850523352623,
+ "learning_rate": 0.00025996500713765,
+ "loss": 3.9197943210601807,
+ "step": 2013
+ },
+ {
+ "epoch": 27.977312390924958,
+ "grad_norm": 0.05592802166938782,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 3.9471435546875,
+ "step": 2014
+ },
+ {
+ "epoch": 27.991273996509598,
+ "grad_norm": 0.059447627514600754,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 3.9220168590545654,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.04082372784614563,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 2.433501958847046,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 0.6050393581390381,
+ "eval_runtime": 57.4222,
+ "eval_samples_per_second": 42.527,
+ "eval_steps_per_second": 0.679,
+ "step": 2016
+ },
+ {
+ "epoch": 28.013961605584644,
+ "grad_norm": 0.059585411101579666,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 3.8270139694213867,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027923211169284,
+ "grad_norm": 0.0652594342827797,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 3.842228412628174,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041884816753928,
+ "grad_norm": 0.05607042834162712,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 3.8498799800872803,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05584642233857,
+ "grad_norm": 0.05797959491610527,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 3.8537564277648926,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069808027923212,
+ "grad_norm": 0.06011393293738365,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 3.865638256072998,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083769633507853,
+ "grad_norm": 0.05946647375822067,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 3.8406717777252197,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097731239092496,
+ "grad_norm": 0.05657580494880676,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 3.8516461849212646,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111692844677137,
+ "grad_norm": 0.06000044196844101,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 3.859069347381592,
+ "step": 2024
+ },
+ {
+ "epoch": 28.12565445026178,
+ "grad_norm": 0.05713335797190666,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 3.8558874130249023,
+ "step": 2025
+ },
+ {
+ "epoch": 28.13961605584642,
+ "grad_norm": 0.05555511265993118,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 3.882650852203369,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153577661431065,
+ "grad_norm": 0.05772312358021736,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 3.847323179244995,
+ "step": 2027
+ },
+ {
+ "epoch": 28.167539267015705,
+ "grad_norm": 0.056661207228899,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 3.8922581672668457,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18150087260035,
+ "grad_norm": 0.05714283883571625,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 3.849281072616577,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195462478184993,
+ "grad_norm": 0.05582018196582794,
+ "learning_rate": 0.000255486047794226,
+ "loss": 3.8659892082214355,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209424083769633,
+ "grad_norm": 0.05674750730395317,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 3.8591721057891846,
+ "step": 2031
+ },
+ {
+ "epoch": 28.223385689354277,
+ "grad_norm": 0.060421206057071686,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 3.8599629402160645,
+ "step": 2032
+ },
+ {
+ "epoch": 28.237347294938917,
+ "grad_norm": 0.05727452412247658,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 3.8559670448303223,
+ "step": 2033
+ },
+ {
+ "epoch": 28.25130890052356,
+ "grad_norm": 0.05708890035748482,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 3.8643712997436523,
+ "step": 2034
+ },
+ {
+ "epoch": 28.2652705061082,
+ "grad_norm": 0.05768970027565956,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 3.842012405395508,
+ "step": 2035
+ },
+ {
+ "epoch": 28.279232111692846,
+ "grad_norm": 0.05973518267273903,
+ "learning_rate": 0.000253907826333243,
+ "loss": 3.841464042663574,
+ "step": 2036
+ },
+ {
+ "epoch": 28.293193717277486,
+ "grad_norm": 0.057536788284778595,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 3.8655552864074707,
+ "step": 2037
+ },
+ {
+ "epoch": 28.30715532286213,
+ "grad_norm": 0.05864274501800537,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 3.852839469909668,
+ "step": 2038
+ },
+ {
+ "epoch": 28.32111692844677,
+ "grad_norm": 0.056395065039396286,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 3.859243392944336,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335078534031414,
+ "grad_norm": 0.05933082476258278,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 3.843926191329956,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349040139616054,
+ "grad_norm": 0.05862351134419441,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 3.8284354209899902,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363001745200698,
+ "grad_norm": 0.05564470961689949,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 3.8802285194396973,
+ "step": 2042
+ },
+ {
+ "epoch": 28.376963350785342,
+ "grad_norm": 0.06139146164059639,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 3.867100954055786,
+ "step": 2043
+ },
+ {
+ "epoch": 28.390924956369982,
+ "grad_norm": 0.058729205280542374,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 3.8322975635528564,
+ "step": 2044
+ },
+ {
+ "epoch": 28.404886561954626,
+ "grad_norm": 0.05771554261445999,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 3.855440616607666,
+ "step": 2045
+ },
+ {
+ "epoch": 28.418848167539267,
+ "grad_norm": 0.06000557541847229,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 3.8510828018188477,
+ "step": 2046
+ },
+ {
+ "epoch": 28.43280977312391,
+ "grad_norm": 0.05935550108551979,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 3.8572049140930176,
+ "step": 2047
+ },
+ {
+ "epoch": 28.44677137870855,
+ "grad_norm": 0.056745223701000214,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 3.8892972469329834,
+ "step": 2048
+ },
+ {
+ "epoch": 28.460732984293195,
+ "grad_norm": 0.0588800348341465,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 3.865957260131836,
+ "step": 2049
+ },
+ {
+ "epoch": 28.474694589877835,
+ "grad_norm": 0.05756748467683792,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 3.8530519008636475,
+ "step": 2050
+ },
+ {
+ "epoch": 28.48865619546248,
+ "grad_norm": 0.057398825883865356,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 3.9022018909454346,
+ "step": 2051
+ },
+ {
+ "epoch": 28.50261780104712,
+ "grad_norm": 0.05809267610311508,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 3.853647232055664,
+ "step": 2052
+ },
+ {
+ "epoch": 28.516579406631763,
+ "grad_norm": 0.05625335872173309,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 3.890363931655884,
+ "step": 2053
+ },
+ {
+ "epoch": 28.530541012216403,
+ "grad_norm": 0.05852413922548294,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 3.8836288452148438,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544502617801047,
+ "grad_norm": 0.05545232817530632,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 3.902006149291992,
+ "step": 2055
+ },
+ {
+ "epoch": 28.55846422338569,
+ "grad_norm": 0.05776599794626236,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 3.8910727500915527,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57242582897033,
+ "grad_norm": 0.05847984179854393,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 3.890106201171875,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586387434554975,
+ "grad_norm": 0.056724466383457184,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 3.836899995803833,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600349040139616,
+ "grad_norm": 0.05690842866897583,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 3.834292411804199,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61431064572426,
+ "grad_norm": 0.058501504361629486,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 3.891890048980713,
+ "step": 2060
+ },
+ {
+ "epoch": 28.6282722513089,
+ "grad_norm": 0.059110380709171295,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 3.883753538131714,
+ "step": 2061
+ },
+ {
+ "epoch": 28.642233856893544,
+ "grad_norm": 0.0574255995452404,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 3.8829569816589355,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656195462478184,
+ "grad_norm": 0.05842957645654678,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 3.845698118209839,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670157068062828,
+ "grad_norm": 0.05536387115716934,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 3.9008209705352783,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68411867364747,
+ "grad_norm": 0.05769932642579079,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 3.890972137451172,
+ "step": 2065
+ },
+ {
+ "epoch": 28.698080279232112,
+ "grad_norm": 0.058036014437675476,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 3.873745918273926,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712041884816752,
+ "grad_norm": 0.05829216167330742,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 3.879685401916504,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726003490401396,
+ "grad_norm": 0.06169239804148674,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 3.9156579971313477,
+ "step": 2068
+ },
+ {
+ "epoch": 28.739965095986037,
+ "grad_norm": 0.05861479043960571,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 3.876373052597046,
+ "step": 2069
+ },
+ {
+ "epoch": 28.75392670157068,
+ "grad_norm": 0.05878114700317383,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 3.918684959411621,
+ "step": 2070
+ },
+ {
+ "epoch": 28.767888307155324,
+ "grad_norm": 0.059904664754867554,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 3.8893160820007324,
+ "step": 2071
+ },
+ {
+ "epoch": 28.781849912739965,
+ "grad_norm": 0.057385317981243134,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 3.880828619003296,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79581151832461,
+ "grad_norm": 0.05815180763602257,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 3.8638038635253906,
+ "step": 2073
+ },
+ {
+ "epoch": 28.80977312390925,
+ "grad_norm": 0.05871085450053215,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 3.8427436351776123,
+ "step": 2074
+ },
+ {
+ "epoch": 28.823734729493893,
+ "grad_norm": 0.056275319308042526,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 3.858442544937134,
+ "step": 2075
+ },
+ {
+ "epoch": 28.837696335078533,
+ "grad_norm": 0.05850474536418915,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 3.871570110321045,
+ "step": 2076
+ },
+ {
+ "epoch": 28.851657940663177,
+ "grad_norm": 0.0580497644841671,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 3.89687442779541,
+ "step": 2077
+ },
+ {
+ "epoch": 28.865619546247817,
+ "grad_norm": 0.05944157764315605,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 3.8877720832824707,
+ "step": 2078
+ },
+ {
+ "epoch": 28.87958115183246,
+ "grad_norm": 0.05749465152621269,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 3.8787431716918945,
+ "step": 2079
+ },
+ {
+ "epoch": 28.8935427574171,
+ "grad_norm": 0.05645301938056946,
+ "learning_rate": 0.000242380656502223,
+ "loss": 3.9244091510772705,
+ "step": 2080
+ },
+ {
+ "epoch": 28.907504363001745,
+ "grad_norm": 0.05700884759426117,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 3.9192817211151123,
+ "step": 2081
+ },
+ {
+ "epoch": 28.921465968586386,
+ "grad_norm": 0.05823809653520584,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 3.905806064605713,
+ "step": 2082
+ },
+ {
+ "epoch": 28.93542757417103,
+ "grad_norm": 0.05736592411994934,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 3.8907763957977295,
+ "step": 2083
+ },
+ {
+ "epoch": 28.949389179755673,
+ "grad_norm": 0.05646584555506706,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 3.892714500427246,
+ "step": 2084
+ },
+ {
+ "epoch": 28.963350785340314,
+ "grad_norm": 0.06081141158938408,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 3.864286184310913,
+ "step": 2085
+ },
+ {
+ "epoch": 28.977312390924958,
+ "grad_norm": 0.05633515864610672,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 3.8791451454162598,
+ "step": 2086
+ },
+ {
+ "epoch": 28.991273996509598,
+ "grad_norm": 0.05825566127896309,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 3.8895483016967773,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.042954832315444946,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 2.408613681793213,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 0.6065697073936462,
+ "eval_runtime": 59.1276,
+ "eval_samples_per_second": 41.301,
+ "eval_steps_per_second": 0.66,
+ "step": 2088
+ },
+ {
+ "epoch": 29.013961605584644,
+ "grad_norm": 0.06000809744000435,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 3.8419973850250244,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027923211169284,
+ "grad_norm": 0.06233612447977066,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 3.83025860786438,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041884816753928,
+ "grad_norm": 0.0551699623465538,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 3.8283262252807617,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05584642233857,
+ "grad_norm": 0.06225701421499252,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 3.822310447692871,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069808027923212,
+ "grad_norm": 0.05747336149215698,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 3.8208513259887695,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083769633507853,
+ "grad_norm": 0.0582275427877903,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 3.8013901710510254,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097731239092496,
+ "grad_norm": 0.059701140969991684,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 3.829238176345825,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111692844677137,
+ "grad_norm": 0.057550523430109024,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 3.7963826656341553,
+ "step": 2096
+ },
+ {
+ "epoch": 29.12565445026178,
+ "grad_norm": 0.05855116620659828,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 3.824028491973877,
+ "step": 2097
+ },
+ {
+ "epoch": 29.13961605584642,
+ "grad_norm": 0.056533779948949814,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 3.8151779174804688,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153577661431065,
+ "grad_norm": 0.0573289655148983,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 3.7990314960479736,
+ "step": 2099
+ },
+ {
+ "epoch": 29.167539267015705,
+ "grad_norm": 0.060061171650886536,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 3.8221707344055176,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18150087260035,
+ "grad_norm": 0.05851743370294571,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 3.793701648712158,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195462478184993,
+ "grad_norm": 0.059191569685935974,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 3.8194010257720947,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209424083769633,
+ "grad_norm": 0.05996483936905861,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 3.8316292762756348,
+ "step": 2103
+ },
+ {
+ "epoch": 29.223385689354277,
+ "grad_norm": 0.0555657260119915,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 3.821362018585205,
+ "step": 2104
+ },
+ {
+ "epoch": 29.237347294938917,
+ "grad_norm": 0.058476634323596954,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 3.8098394870758057,
+ "step": 2105
+ },
+ {
+ "epoch": 29.25130890052356,
+ "grad_norm": 0.05669668689370155,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 3.843484878540039,
+ "step": 2106
+ },
+ {
+ "epoch": 29.2652705061082,
+ "grad_norm": 0.05648602545261383,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 3.811458110809326,
+ "step": 2107
+ },
+ {
+ "epoch": 29.279232111692846,
+ "grad_norm": 0.05574130266904831,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 3.837676525115967,
+ "step": 2108
+ },
+ {
+ "epoch": 29.293193717277486,
+ "grad_norm": 0.05538785457611084,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 3.810798168182373,
+ "step": 2109
+ },
+ {
+ "epoch": 29.30715532286213,
+ "grad_norm": 0.05477031692862511,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 3.8156371116638184,
+ "step": 2110
+ },
+ {
+ "epoch": 29.32111692844677,
+ "grad_norm": 0.05664678290486336,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 3.8472559452056885,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335078534031414,
+ "grad_norm": 0.05649891495704651,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 3.824376106262207,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349040139616054,
+ "grad_norm": 0.05665554851293564,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 3.7963147163391113,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363001745200698,
+ "grad_norm": 0.054111722856760025,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 3.819735527038574,
+ "step": 2114
+ },
+ {
+ "epoch": 29.376963350785342,
+ "grad_norm": 0.05805648863315582,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 3.790189743041992,
+ "step": 2115
+ },
+ {
+ "epoch": 29.390924956369982,
+ "grad_norm": 0.05634011700749397,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 3.852161169052124,
+ "step": 2116
+ },
+ {
+ "epoch": 29.404886561954626,
+ "grad_norm": 0.0555867962539196,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 3.846884250640869,
+ "step": 2117
+ },
+ {
+ "epoch": 29.418848167539267,
+ "grad_norm": 0.057229746133089066,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 3.8432483673095703,
+ "step": 2118
+ },
+ {
+ "epoch": 29.43280977312391,
+ "grad_norm": 0.05664287880063057,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 3.842709541320801,
+ "step": 2119
+ },
+ {
+ "epoch": 29.44677137870855,
+ "grad_norm": 0.0567023791372776,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 3.844529628753662,
+ "step": 2120
+ },
+ {
+ "epoch": 29.460732984293195,
+ "grad_norm": 0.057870302349328995,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 3.853799343109131,
+ "step": 2121
+ },
+ {
+ "epoch": 29.474694589877835,
+ "grad_norm": 0.05681667849421501,
+ "learning_rate": 0.000231465389734324,
+ "loss": 3.849405288696289,
+ "step": 2122
+ },
+ {
+ "epoch": 29.48865619546248,
+ "grad_norm": 0.05738653615117073,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 3.8083410263061523,
+ "step": 2123
+ },
+ {
+ "epoch": 29.50261780104712,
+ "grad_norm": 0.058900732547044754,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 3.840458869934082,
+ "step": 2124
+ },
+ {
+ "epoch": 29.516579406631763,
+ "grad_norm": 0.05823393166065216,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 3.821415662765503,
+ "step": 2125
+ },
+ {
+ "epoch": 29.530541012216403,
+ "grad_norm": 0.057413019239902496,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 3.8713812828063965,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544502617801047,
+ "grad_norm": 0.05935502052307129,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 3.8488903045654297,
+ "step": 2127
+ },
+ {
+ "epoch": 29.55846422338569,
+ "grad_norm": 0.055285241454839706,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 3.8196067810058594,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57242582897033,
+ "grad_norm": 0.05894080176949501,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 3.8444557189941406,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586387434554975,
+ "grad_norm": 0.05878668278455734,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 3.8731908798217773,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600349040139616,
+ "grad_norm": 0.05855663865804672,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 3.831028938293457,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61431064572426,
+ "grad_norm": 0.057419586926698685,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 3.849435806274414,
+ "step": 2132
+ },
+ {
+ "epoch": 29.6282722513089,
+ "grad_norm": 0.058863408863544464,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 3.8495519161224365,
+ "step": 2133
+ },
+ {
+ "epoch": 29.642233856893544,
+ "grad_norm": 0.06081194058060646,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 3.844421863555908,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656195462478184,
+ "grad_norm": 0.05597590655088425,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 3.8413243293762207,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670157068062828,
+ "grad_norm": 0.058956339955329895,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 3.861703872680664,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68411867364747,
+ "grad_norm": 0.05883771926164627,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 3.8756394386291504,
+ "step": 2137
+ },
+ {
+ "epoch": 29.698080279232112,
+ "grad_norm": 0.06036122143268585,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 3.855403423309326,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712041884816752,
+ "grad_norm": 0.061243657022714615,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 3.861691474914551,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726003490401396,
+ "grad_norm": 0.05917691811919212,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 3.870516777038574,
+ "step": 2140
+ },
+ {
+ "epoch": 29.739965095986037,
+ "grad_norm": 0.06157418712973595,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 3.8292267322540283,
+ "step": 2141
+ },
+ {
+ "epoch": 29.75392670157068,
+ "grad_norm": 0.0605802945792675,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 3.836577892303467,
+ "step": 2142
+ },
+ {
+ "epoch": 29.767888307155324,
+ "grad_norm": 0.05750525742769241,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 3.8472204208374023,
+ "step": 2143
+ },
+ {
+ "epoch": 29.781849912739965,
+ "grad_norm": 0.057982541620731354,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 3.8682217597961426,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79581151832461,
+ "grad_norm": 0.057891640812158585,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 3.8397727012634277,
+ "step": 2145
+ },
+ {
+ "epoch": 29.80977312390925,
+ "grad_norm": 0.058175452053546906,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 3.8578720092773438,
+ "step": 2146
+ },
+ {
+ "epoch": 29.823734729493893,
+ "grad_norm": 0.05686056986451149,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 3.8467812538146973,
+ "step": 2147
+ },
+ {
+ "epoch": 29.837696335078533,
+ "grad_norm": 0.05755440890789032,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 3.844052314758301,
+ "step": 2148
+ },
+ {
+ "epoch": 29.851657940663177,
+ "grad_norm": 0.057706039398908615,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 3.818631172180176,
+ "step": 2149
+ },
+ {
+ "epoch": 29.865619546247817,
+ "grad_norm": 0.058824241161346436,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 3.8732900619506836,
+ "step": 2150
+ },
+ {
+ "epoch": 29.87958115183246,
+ "grad_norm": 0.058583300560712814,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 3.8373632431030273,
+ "step": 2151
+ },
+ {
+ "epoch": 29.8935427574171,
+ "grad_norm": 0.057403020560741425,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 3.8761391639709473,
+ "step": 2152
+ },
+ {
+ "epoch": 29.907504363001745,
+ "grad_norm": 0.0586736761033535,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 3.8518989086151123,
+ "step": 2153
+ },
+ {
+ "epoch": 29.921465968586386,
+ "grad_norm": 0.05727249011397362,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 3.839515209197998,
+ "step": 2154
+ },
+ {
+ "epoch": 29.93542757417103,
+ "grad_norm": 0.0594087690114975,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 3.834868907928467,
+ "step": 2155
+ },
+ {
+ "epoch": 29.949389179755673,
+ "grad_norm": 0.057742297649383545,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 3.861978530883789,
+ "step": 2156
+ },
+ {
+ "epoch": 29.963350785340314,
+ "grad_norm": 0.05707386136054993,
+ "learning_rate": 0.00022244633283095,
+ "loss": 3.886831760406494,
+ "step": 2157
+ },
+ {
+ "epoch": 29.977312390924958,
+ "grad_norm": 0.05631854757666588,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 3.83940052986145,
+ "step": 2158
+ },
+ {
+ "epoch": 29.991273996509598,
+ "grad_norm": 0.057860348373651505,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 3.8477416038513184,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.04080710560083389,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 2.4169907569885254,
+ "step": 2160
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 9.222244509155328e+17,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-2160/training_args.bin b/runs/l2r90-ssl/checkpoint-2160/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..103048067ef8897d0386013ed89ce6764bad5521
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2160/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f320fb44ea3a4a93198f1a6c2405447428148ce142cd94ee283a3c077a327c2d
+size 4792
diff --git a/runs/l2r90-ssl/checkpoint-2520/chat_template.jinja b/runs/l2r90-ssl/checkpoint-2520/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2520/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-ssl/checkpoint-2520/config.json b/runs/l2r90-ssl/checkpoint-2520/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2520/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-ssl/checkpoint-2520/generation_config.json b/runs/l2r90-ssl/checkpoint-2520/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2520/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-ssl/checkpoint-2520/model.safetensors b/runs/l2r90-ssl/checkpoint-2520/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..dec51dee4dc412361f890e1250ea6bae8c4a6d95
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2520/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1ccd1f3ab05635340c11191f0d71d8b8b8fd6fde23e7da9ea22fc120db6a3a9a
+size 583356232
diff --git a/runs/l2r90-ssl/checkpoint-2520/optimizer.pt b/runs/l2r90-ssl/checkpoint-2520/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..b4cde7051c8537d1b9804ede744dee82281a4179
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2520/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:26b0b0c39c9b51476b7699fbaf836e342da0f147f17ef3e47720b294f22b8478
+size 1166827898
diff --git a/runs/l2r90-ssl/checkpoint-2520/rng_state_0.pth b/runs/l2r90-ssl/checkpoint-2520/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..5a1ed108c5e3dd5a0d01545311a2fbfc85ac080c
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2520/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7e38b28450b484bae9066d3591189b4df81a5b8528d06e5a2b200106829c1ea0
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-2520/rng_state_1.pth b/runs/l2r90-ssl/checkpoint-2520/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..bd3984d5a741ff19bdaf4610d6ecd59df0a8dab0
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2520/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5e143c6e2599c38a9fc0815513d0cafda411c6b0eef6f87171d11fee664205d0
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-2520/scheduler.pt b/runs/l2r90-ssl/checkpoint-2520/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..b53f13f877641b522f2c435934f587fdad6161cc
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2520/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6f776ea81adcefd29eff4d7371e55373e8890bd750697e1c8b781d131e0a39a6
+size 1064
diff --git a/runs/l2r90-ssl/checkpoint-2520/tokenizer.json b/runs/l2r90-ssl/checkpoint-2520/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2520/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-ssl/checkpoint-2520/tokenizer_config.json b/runs/l2r90-ssl/checkpoint-2520/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2520/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-2520/trainer_state.json b/runs/l2r90-ssl/checkpoint-2520/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..cd5acbee3863421c50d2d1123969a0e230348614
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2520/trainer_state.json
@@ -0,0 +1,17938 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 35.0,
+ "eval_steps": 500,
+ "global_step": 2520,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.2588357925415039,
+ "learning_rate": 0.0,
+ "loss": 3.9982247352600098,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.25455471873283386,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.998683452606201,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.2555367052555084,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.9821574687957764,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.2383691370487213,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.9531917572021484,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2023179680109024,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.9148876667022705,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.17691479623317719,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.875638961791992,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.14372293651103973,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.836796283721924,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1074419692158699,
+ "learning_rate": 4.2e-05,
+ "loss": 3.8036489486694336,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.08204353600740433,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.774808406829834,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.0632256343960762,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.7503671646118164,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.047985248267650604,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.7345473766326904,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.03675895184278488,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.721525192260742,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.02908993512392044,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.713454246520996,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.02443614974617958,
+ "learning_rate": 7.8e-05,
+ "loss": 3.7078704833984375,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02109951712191105,
+ "learning_rate": 8.4e-05,
+ "loss": 3.7034401893615723,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.018616363406181335,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.7009854316711426,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.016399752348661423,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.697768211364746,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.014256286434829235,
+ "learning_rate": 0.000102,
+ "loss": 3.6974306106567383,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.013020666316151619,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.694406509399414,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.011879238300025463,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.691807270050049,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.010290366597473621,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.6937897205352783,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.009482803754508495,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.6902384757995605,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.008425266481935978,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.686815023422241,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.007676936220377684,
+ "learning_rate": 0.000138,
+ "loss": 3.685192823410034,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.007491032127290964,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.6839442253112793,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.007428477052599192,
+ "learning_rate": 0.00015,
+ "loss": 3.6783409118652344,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.006965892389416695,
+ "learning_rate": 0.000156,
+ "loss": 3.6813433170318604,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.008106694556772709,
+ "learning_rate": 0.000162,
+ "loss": 3.6741726398468018,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.0076780058443546295,
+ "learning_rate": 0.000168,
+ "loss": 3.67673921585083,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.007904314436018467,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.6724467277526855,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.008422375656664371,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.673013925552368,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00811714306473732,
+ "learning_rate": 0.000186,
+ "loss": 3.6698389053344727,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.009063836187124252,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.668990135192871,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.01878332532942295,
+ "learning_rate": 0.000198,
+ "loss": 3.6658849716186523,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.01812782697379589,
+ "learning_rate": 0.000204,
+ "loss": 3.6697864532470703,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.020587686449289322,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.662757158279419,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.01515690516680479,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.6558732986450195,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.02510508894920349,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.6495113372802734,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.020487826317548752,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.638803243637085,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.017446046695113182,
+ "learning_rate": 0.000234,
+ "loss": 3.633307456970215,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.030667584389448166,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.6277506351470947,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.023789944127202034,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.6189804077148438,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.02660694345831871,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.610522747039795,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.023814303800463676,
+ "learning_rate": 0.000258,
+ "loss": 3.615910291671753,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.021863561123609543,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.612008571624756,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03705216199159622,
+ "learning_rate": 0.00027,
+ "loss": 3.6066887378692627,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.024143919348716736,
+ "learning_rate": 0.000276,
+ "loss": 3.5987331867218018,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.03251276910305023,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.595367431640625,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06619632989168167,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.59018611907959,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.042714253067970276,
+ "learning_rate": 0.000294,
+ "loss": 3.58981990814209,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.018893254920840263,
+ "learning_rate": 0.0003,
+ "loss": 3.578472375869751,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0749685987830162,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.5828654766082764,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.12142857164144516,
+ "learning_rate": 0.000312,
+ "loss": 3.5840892791748047,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.06944665312767029,
+ "learning_rate": 0.000318,
+ "loss": 3.580961227416992,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.04545000568032265,
+ "learning_rate": 0.000324,
+ "loss": 3.5674569606781006,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.061073049902915955,
+ "learning_rate": 0.00033,
+ "loss": 3.569991111755371,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.04133498668670654,
+ "learning_rate": 0.000336,
+ "loss": 3.5657706260681152,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.055982090532779694,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.559979200363159,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.05259208381175995,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.558039426803589,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.029976001009345055,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.5491275787353516,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.04998327046632767,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.5406670570373535,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.03230690956115723,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.5391530990600586,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.06102178618311882,
+ "learning_rate": 0.000372,
+ "loss": 3.5311684608459473,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.031715717166662216,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.518899440765381,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.06841139495372772,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.514200210571289,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.10323675721883774,
+ "learning_rate": 0.00039,
+ "loss": 3.513646125793457,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.1448228508234024,
+ "learning_rate": 0.000396,
+ "loss": 3.5110087394714355,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.12129563093185425,
+ "learning_rate": 0.000402,
+ "loss": 3.501497507095337,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.030224842950701714,
+ "learning_rate": 0.000408,
+ "loss": 3.4870352745056152,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.07492101192474365,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.482234001159668,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.03942760080099106,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.4746525287628174,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.04967200756072998,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.604069232940674,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8671882152557373,
+ "eval_runtime": 42.6602,
+ "eval_samples_per_second": 57.243,
+ "eval_steps_per_second": 0.469,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.05385325849056244,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.462052822113037,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.04870929941534996,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.459104537963867,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.08701689541339874,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.4533209800720215,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.12219952791929245,
+ "learning_rate": 0.00045,
+ "loss": 3.4558796882629395,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.0823654904961586,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.443049192428589,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.12765446305274963,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.4373486042022705,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.14215047657489777,
+ "learning_rate": 0.000468,
+ "loss": 3.449429988861084,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.03649887070059776,
+ "learning_rate": 0.000474,
+ "loss": 3.42877197265625,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09551949054002762,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.4352874755859375,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.030483486130833626,
+ "learning_rate": 0.000486,
+ "loss": 3.4125874042510986,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.0745147094130516,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.4153835773468018,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.06620422005653381,
+ "learning_rate": 0.000498,
+ "loss": 3.4053854942321777,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.06098725646734238,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.404346466064453,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.047781601548194885,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.3991761207580566,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.04409855231642723,
+ "learning_rate": 0.000516,
+ "loss": 3.396144390106201,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.04355994611978531,
+ "learning_rate": 0.000522,
+ "loss": 3.3928911685943604,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.0670008510351181,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.3831443786621094,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.0812622532248497,
+ "learning_rate": 0.000534,
+ "loss": 3.382413148880005,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.0632157102227211,
+ "learning_rate": 0.00054,
+ "loss": 3.3784892559051514,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.028686825186014175,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.3673207759857178,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.034818850457668304,
+ "learning_rate": 0.000552,
+ "loss": 3.3648972511291504,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.04509204626083374,
+ "learning_rate": 0.000558,
+ "loss": 3.35561203956604,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.054211195558309555,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.3612167835235596,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.03752049803733826,
+ "learning_rate": 0.00057,
+ "loss": 3.357103109359741,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.03620241954922676,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.3469700813293457,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.02454194985330105,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.3429036140441895,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.041030462831258774,
+ "learning_rate": 0.000588,
+ "loss": 3.338247299194336,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.03266424685716629,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.332688331604004,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.04730091616511345,
+ "learning_rate": 0.0006,
+ "loss": 3.335902452468872,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.065738245844841,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.3289852142333984,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.09062394499778748,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.3278567790985107,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.1482805460691452,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.332685947418213,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.20066532492637634,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.363180637359619,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.11500677466392517,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.3308753967285156,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.11754138022661209,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.3383703231811523,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.09672239422798157,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.3267624378204346,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.07190799713134766,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.317737579345703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.0815449133515358,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.3198161125183105,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09612327069044113,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.315814971923828,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.14795289933681488,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.317605495452881,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06667589396238327,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.3127357959747314,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.058537885546684265,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.319807767868042,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.06428155303001404,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.297541618347168,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.03881775587797165,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.297539710998535,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.061801470816135406,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.290297508239746,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.07642851769924164,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.284428596496582,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08123870939016342,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.2820234298706055,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07174684852361679,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.2697296142578125,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.04030010476708412,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.264328956604004,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.05662153288722038,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.2699337005615234,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.03761111572384834,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.255704879760742,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.043027132749557495,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.2579030990600586,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.044409338384866714,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.248547077178955,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.0239730142056942,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.249307632446289,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03628705441951752,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.2451741695404053,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.029564054682850838,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.238445997238159,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.032036975026130676,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.2320892810821533,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.030085794627666473,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.231468915939331,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.025232303887605667,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.232199192047119,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.027053389698266983,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.210745334625244,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.028464114293456078,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.210076332092285,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04616783559322357,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.2103490829467773,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.039081938564777374,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.202662467956543,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.03428936004638672,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.1997475624084473,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.04351917654275894,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.19844913482666,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.0840945914387703,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.2013347148895264,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.36558300256729126,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.26310396194458,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.09432770311832428,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.2881736755371094,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.061970267444849014,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.2840280532836914,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.05112256482243538,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.2632040977478027,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06797203421592712,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.254606246948242,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.051026541739702225,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.429918050765991,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.8117228746414185,
+ "eval_runtime": 42.8572,
+ "eval_samples_per_second": 56.98,
+ "eval_steps_per_second": 0.467,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.0608622208237648,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.2348179817199707,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.05581629276275635,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.22896146774292,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.09162136912345886,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.2343664169311523,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.13750384747982025,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.2507717609405518,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.0872517079114914,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.2440056800842285,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.07161762565374374,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.229494571685791,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04082738608121872,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.2183239459991455,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.04972497373819351,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.218075752258301,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.049254074692726135,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.213379383087158,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.056671254336833954,
+ "learning_rate": 0.000599663984012462,
+ "loss": 3.209568977355957,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.05541909113526344,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.207760810852051,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.03819117695093155,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 3.199732780456543,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.0529443584382534,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.1945419311523438,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.043010979890823364,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.185936212539673,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.07311078161001205,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.189894914627075,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06110258400440216,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 3.1911773681640625,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.06880149245262146,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 3.187437057495117,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.051326874643564224,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 3.1826558113098145,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.03363822400569916,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 3.175527572631836,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.10268989205360413,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 3.172367572784424,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.09723010659217834,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 3.1873984336853027,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.04129317030310631,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 3.165799617767334,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.05292229354381561,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 3.171780824661255,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.03873143345117569,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 3.1680874824523926,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.047018859535455704,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 3.161402940750122,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.03942955285310745,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 3.1530532836914062,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.039801858365535736,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 3.1465344429016113,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.03795155510306358,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 3.1364030838012695,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.0342850424349308,
+ "learning_rate": 0.000599379982511273,
+ "loss": 3.1391525268554688,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.03243280574679375,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 3.1331253051757812,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.029679352417588234,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 3.1346120834350586,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.030230293050408363,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 3.132615804672241,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.02845083735883236,
+ "learning_rate": 0.000599309205504528,
+ "loss": 3.1242547035217285,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.03208369016647339,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 3.119110584259033,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.040268607437610626,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 3.113833427429199,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.06951889395713806,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 3.1186914443969727,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.08879823237657547,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 3.1140987873077393,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07374681532382965,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 3.1147828102111816,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.038190409541130066,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 3.1161253452301025,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.03297388181090355,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 3.1006083488464355,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.05905177444219589,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 3.106564521789551,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.037567321211099625,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 3.0884456634521484,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.02725658379495144,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 3.09000825881958,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.04272424429655075,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 3.090528964996338,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.027587950229644775,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 3.085131883621216,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.03282274678349495,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 3.0880112648010254,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.04064026474952698,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 3.079259157180786,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.023374097421765327,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 3.0747299194335938,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.025003870949149132,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 3.0721259117126465,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.030445197597146034,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 3.0733323097229004,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.025935636833310127,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 3.070709228515625,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.021337421610951424,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 3.0627551078796387,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.040481697767972946,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 3.0529394149780273,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.047608256340026855,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 3.041020631790161,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.03356955572962761,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 3.0532307624816895,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.02101682312786579,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 3.0492985248565674,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.030098581686615944,
+ "learning_rate": 0.000598804365316586,
+ "loss": 3.0527801513671875,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.040166422724723816,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 3.0505762100219727,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.051579318940639496,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 3.0422110557556152,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.06485334783792496,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 3.0435101985931396,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.055590901523828506,
+ "learning_rate": 0.000598706872384039,
+ "loss": 3.0448739528656006,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.03183475881814957,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 3.0416858196258545,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.03995629772543907,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 3.0255980491638184,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.03907827287912369,
+ "learning_rate": 0.000598631251036868,
+ "loss": 3.032339096069336,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.027681387960910797,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 3.025439739227295,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.03470184653997421,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 3.02883243560791,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.031407181173563004,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 3.0289883613586426,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.0371522456407547,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 3.013608932495117,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.03651345893740654,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 3.0120253562927246,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.06004595756530762,
+ "learning_rate": 0.000598473577921366,
+ "loss": 3.017958402633667,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.09979518502950668,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 3.023041248321533,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.10138462483882904,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.275729179382324,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7571707367897034,
+ "eval_runtime": 42.8283,
+ "eval_samples_per_second": 57.018,
+ "eval_steps_per_second": 0.467,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.04982660710811615,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 3.016209602355957,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.06532539427280426,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 3.015043020248413,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.04369623214006424,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 3.0096778869628906,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.05180002748966217,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 3.0121543407440186,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.0440233014523983,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.9937305450439453,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.05645671859383583,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 3.0023865699768066,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.03990716487169266,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 3.0003232955932617,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.03818608075380325,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.994137763977051,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.035388972610235214,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.993170738220215,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.03162284567952156,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.9922614097595215,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.029497336596250534,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.9852986335754395,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.02586926706135273,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.9798672199249268,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.027355050668120384,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.985158920288086,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.0341978520154953,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.982581615447998,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.030727429315447807,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.9701411724090576,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.04429870471358299,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.9789254665374756,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.05775132030248642,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.9758143424987793,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.05635962635278702,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.981480121612549,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.03639192506670952,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.968752861022949,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.028748342767357826,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.9584972858428955,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.035653006285429,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.967545747756958,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.03697885945439339,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.9578099250793457,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.03518654406070709,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.9680328369140625,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.034762222319841385,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.9670262336730957,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03813016042113304,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.957522392272949,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.04372904822230339,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.9540514945983887,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.06646446138620377,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.951028823852539,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.07831092923879623,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.95898699760437,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.05779588967561722,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.965884208679199,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.04414296895265579,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.9510560035705566,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.04797815904021263,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.955803394317627,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.04482145980000496,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.954059362411499,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.037104446440935135,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.94792103767395,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.03616701066493988,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.9500932693481445,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.03882797807455063,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.9558305740356445,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.04356204345822334,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.9237029552459717,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.04535257816314697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.925853729248047,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.04519587755203247,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.924984931945801,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.055922407656908035,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.932561159133911,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.06330762803554535,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.9388322830200195,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.0517115443944931,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.9272942543029785,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.057369694113731384,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.9201431274414062,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.05687880516052246,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.922886371612549,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.055525410920381546,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.9211201667785645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04503652825951576,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.9164583683013916,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.03385302051901817,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.9145395755767822,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03564384952187538,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.916003704071045,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.04188712686300278,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.915839433670044,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03657370060682297,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.903120279312134,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.03126626834273338,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.895824432373047,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.04196290299296379,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.9093422889709473,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.1706087589263916,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.9183316230773926,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.05810421332716942,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.911245107650757,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.061330314725637436,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.9159231185913086,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.06062281131744385,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.9211535453796387,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.05590800568461418,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.9140496253967285,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.04496239125728607,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.9040920734405518,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.03835335746407509,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.9026548862457275,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.05137286335229874,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.889634132385254,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.04091101884841919,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.886838674545288,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.04076292738318443,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.892918109893799,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.045437734574079514,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.893305778503418,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.041114673018455505,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.88517165184021,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.03600773215293884,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.8779795169830322,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.04311220720410347,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.87247896194458,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.04818756878376007,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.889103412628174,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.053460877388715744,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.8899683952331543,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.06763825565576553,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.8823537826538086,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.07304801046848297,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.877340316772461,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.061337269842624664,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.8820111751556396,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.04660911113023758,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.868886709213257,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.053929202258586884,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.159337043762207,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.721762478351593,
+ "eval_runtime": 42.7977,
+ "eval_samples_per_second": 57.059,
+ "eval_steps_per_second": 0.467,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.04259369149804115,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.859579086303711,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.044425517320632935,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.8681139945983887,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.04301869496703148,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.853911876678467,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.04574296995997429,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.8724427223205566,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.050247322767972946,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.8528172969818115,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.046683840453624725,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.853405714035034,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.05966169387102127,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.8605127334594727,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.08935574442148209,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.862339973449707,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.08019179105758667,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.8655967712402344,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.04753892868757248,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.8657939434051514,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06868117302656174,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.861976146697998,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.07040470093488693,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.8562135696411133,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.05683179199695587,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.8576629161834717,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.04918001592159271,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.861943244934082,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.04921915754675865,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.8495378494262695,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.037651967257261276,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.8428618907928467,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.03522057458758354,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.8372769355773926,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.03111186809837818,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.8377151489257812,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.03906773030757904,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.846196174621582,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.02546556107699871,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.8288016319274902,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.03119821846485138,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.8334617614746094,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.030583377927541733,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.8209362030029297,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.029693076387047768,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.8267486095428467,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.028978517279028893,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.819244861602783,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.027871523052453995,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.8289647102355957,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.029287064447999,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.8235747814178467,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.024990517646074295,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.819016218185425,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.030980689451098442,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.828155040740967,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03721801936626434,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.813418388366699,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03738148882985115,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.808328866958618,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.03580411896109581,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.812070846557617,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.039036594331264496,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.807159423828125,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.044666968286037445,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.799992561340332,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.055589817464351654,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.804142951965332,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.06139020994305611,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.8182730674743652,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.07769318670034409,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.8166518211364746,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.06521332263946533,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.8142249584198,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.09083619713783264,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.8166933059692383,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.09737467020750046,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.8111555576324463,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.07148188352584839,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.8168413639068604,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.055516090244054794,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.8183553218841553,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.05338391661643982,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.8127028942108154,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.040785472840070724,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.8124775886535645,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.037135496735572815,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.8065357208251953,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.032455578446388245,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.798459053039551,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.03465662896633148,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.798649549484253,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.032026033848524094,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.7906997203826904,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.026019956916570663,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.791969060897827,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.030298635363578796,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.784593105316162,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.028147678822278976,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.782454013824463,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.023933131247758865,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.787851333618164,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.026789527386426926,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.788067579269409,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.024554884061217308,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.777033567428589,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.028679484501481056,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.786342144012451,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.026410218328237534,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.7790610790252686,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.027413051575422287,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.778944969177246,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.02908242493867874,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.763850212097168,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.029632994905114174,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.7672982215881348,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.027458716183900833,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.7702178955078125,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.03263958916068077,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.7581803798675537,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04427146911621094,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.768141508102417,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.07142394036054611,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.75819730758667,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.08833584934473038,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.782716751098633,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.08168833702802658,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.7763240337371826,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07253874838352203,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.7717690467834473,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.08186902105808258,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.767303466796875,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.05854064226150513,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.761544704437256,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.04602917283773422,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.7669177055358887,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.03920970857143402,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.766085624694824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.04551195353269577,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.7623450756073,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.040861159563064575,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.7471840381622314,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.026907291263341904,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 2.0710244178771973,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6934371590614319,
+ "eval_runtime": 42.6015,
+ "eval_samples_per_second": 57.322,
+ "eval_steps_per_second": 0.469,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.03801899403333664,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.739086627960205,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.03707238286733627,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.741638422012329,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.0387551411986351,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.740777015686035,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.028010359033942223,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.7353010177612305,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.026474518701434135,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.73842453956604,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.02920856513082981,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.736295461654663,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.026074763387441635,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.7295751571655273,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.026862047612667084,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.723040819168091,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.026300370693206787,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.7199485301971436,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.026030950248241425,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.7276217937469482,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.028342435136437416,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.7193045616149902,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.032135091722011566,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.7143118381500244,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.03068603202700615,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.7191050052642822,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.035828713327646255,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.718085765838623,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.03343002125620842,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.707131862640381,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.035051796585321426,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.7091879844665527,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.03395913913846016,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6980862617492676,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.034198079258203506,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.699704647064209,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.0322839580476284,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.702791213989258,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.041531339287757874,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.7018051147460938,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.040806882083415985,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.7096598148345947,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.04156940430402756,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6956775188446045,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.05608797073364258,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.70127272605896,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0729481428861618,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.708817720413208,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.0905836969614029,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.718085289001465,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.08345075696706772,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.7235910892486572,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.08599896728992462,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.7178783416748047,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.07301691174507141,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.714029312133789,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.08231015503406525,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.7196826934814453,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.08286412060260773,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.7176082134246826,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.08590883016586304,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.7242164611816406,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.07531397044658661,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.7225024700164795,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.04711190238595009,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.695887327194214,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.05076327919960022,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.7029781341552734,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.04495669901371002,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.7057900428771973,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.04312257468700409,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.697537899017334,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.043580975383520126,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.67315673828125,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.03360886126756668,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.6883413791656494,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.030360296368598938,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.700927257537842,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.0310810673981905,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.6917803287506104,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.02703232690691948,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.6667580604553223,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.02756045199930668,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.6702821254730225,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.02611083723604679,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.669790744781494,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.02615324780344963,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.6672821044921875,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.026000119745731354,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.6671881675720215,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.02882794849574566,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.67586088180542,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03327213600277901,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.6618690490722656,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.032920945435762405,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.665444850921631,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.02812962979078293,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.6680707931518555,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.025886602699756622,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.652898073196411,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.025222789496183395,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.6560399532318115,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.023131893947720528,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.66127347946167,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.02513662725687027,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.6463732719421387,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.027243830263614655,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.6507630348205566,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.025437965989112854,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.6485800743103027,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.02770340070128441,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.657241106033325,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.03594067320227623,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.6281137466430664,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04486451297998428,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.6340909004211426,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.05102410912513733,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.6463394165039062,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04514491558074951,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.636554002761841,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04469453915953636,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.6453680992126465,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.05421523004770279,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.646010398864746,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04779462888836861,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.6315622329711914,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.06697724014520645,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.6416730880737305,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.07520416378974915,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.641632318496704,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.055692676454782486,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.6289877891540527,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.04539895057678223,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.6363251209259033,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.04829065501689911,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.641608476638794,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.041314542293548584,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.6269643306732178,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.04100198298692703,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.6243174076080322,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.04105962812900543,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.626585006713867,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.03172122687101364,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9550899267196655,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6614077091217041,
+ "eval_runtime": 42.9009,
+ "eval_samples_per_second": 56.922,
+ "eval_steps_per_second": 0.466,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.03258531913161278,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.612126588821411,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.040250636637210846,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.6025328636169434,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.034234337508678436,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.6085972785949707,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.03635786101222038,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.6126222610473633,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.039404451847076416,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.6035327911376953,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.038504838943481445,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5985798835754395,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.03744236007332802,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5873115062713623,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.034184787422418594,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.5899314880371094,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.03766429424285889,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.602734327316284,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.03741588443517685,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.592207193374634,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.03187853842973709,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5850372314453125,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.038447409868240356,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.592654228210449,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.04046263173222542,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.5881710052490234,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.03872602432966232,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5881471633911133,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.03960775211453438,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5826945304870605,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.033132150769233704,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.584038734436035,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.029504230245947838,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.584357261657715,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.03665253147482872,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5677168369293213,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.037208545953035355,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.560795783996582,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.038244519382715225,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.591458559036255,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.043307431042194366,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.5642588138580322,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.055119726806879044,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.57377290725708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.053323689848184586,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5767312049865723,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.0502864345908165,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.562190055847168,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.05568842589855194,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.565556049346924,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.061149828135967255,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5883288383483887,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.059933699667453766,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.5710184574127197,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.06041933596134186,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.572603225708008,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.05482695251703262,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.5667271614074707,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.04712968319654465,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.554041624069214,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.04842432215809822,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.5660672187805176,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.04273047298192978,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.554839849472046,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.05356749892234802,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.5521483421325684,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.06062676012516022,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.5551886558532715,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.05431462824344635,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5654542446136475,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.0505850687623024,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5648036003112793,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.04696856811642647,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.5559377670288086,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.042585279792547226,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.5660319328308105,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.04026484116911888,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.551206588745117,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03382421284914017,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.556361198425293,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.031177924945950508,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.5413894653320312,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.028217196464538574,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.5600497722625732,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.027062954381108284,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5447568893432617,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.03001212701201439,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.540614604949951,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.028869854286313057,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.544583797454834,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.02869918756186962,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.5471596717834473,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03359716013073921,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5416982173919678,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.04000182822346687,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.534921407699585,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.043030042201280594,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5368564128875732,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.04329848289489746,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.539764881134033,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.03740967437624931,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.5246665477752686,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.03733935207128525,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.529362916946411,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.039609868079423904,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.542544364929199,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.03937261551618576,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5366649627685547,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.03261927142739296,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.5250535011291504,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.03904294595122337,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.520721435546875,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.03913239762187004,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.5122013092041016,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.03610851988196373,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.5212767124176025,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.03944548964500427,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.537186622619629,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04762452840805054,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.5204620361328125,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04427091404795647,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5286149978637695,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04366641491651535,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.5398361682891846,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.04644155502319336,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.519331932067871,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.04372382536530495,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.531022787094116,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.04404324293136597,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.521758794784546,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.042115990072488785,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.5146074295043945,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.03796013817191124,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.5040054321289062,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.04327977076172829,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.5209312438964844,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04150385782122612,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.515871286392212,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.04465346783399582,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.5247764587402344,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.05690346285700798,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.501911163330078,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04907930642366409,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8968371152877808,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6358620524406433,
+ "eval_runtime": 42.9991,
+ "eval_samples_per_second": 56.792,
+ "eval_steps_per_second": 0.465,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.03755255416035652,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.486576795578003,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.040234558284282684,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.501087188720703,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.041416723281145096,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4942233562469482,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.039655014872550964,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.503453254699707,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.041707683354616165,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.490978717803955,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.04545537009835243,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4757118225097656,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.04328389838337898,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.504516839981079,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.043312299996614456,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.4920215606689453,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.046466976404190063,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.494760513305664,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.041762981563806534,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4875552654266357,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.03713781014084816,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.495652675628662,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.040603458881378174,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4821407794952393,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.037969645112752914,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.484981060028076,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.03652074560523033,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4736409187316895,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.03783264383673668,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.483346700668335,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.03404324874281883,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.48856782913208,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.03294725716114044,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.483706474304199,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.0347001850605011,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.474888324737549,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.0359107106924057,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.478990077972412,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.03738103806972504,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.4709420204162598,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.03722188249230385,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.465726852416992,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.034907322376966476,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.482231855392456,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.035695310682058334,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4785828590393066,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.04009179025888443,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.4795875549316406,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.038996778428554535,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.478696823120117,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04082407429814339,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.465200901031494,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04856136813759804,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4605681896209717,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.05671004578471184,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.47277569770813,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.0481606163084507,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.476151466369629,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.037689510732889175,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.466142177581787,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.05005146935582161,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.4798035621643066,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.04745829105377197,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.46163272857666,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.04638158157467842,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4728071689605713,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.04530330374836922,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.4621877670288086,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.03971488028764725,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.474224090576172,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.04521876201033592,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.46122670173645,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03923020884394646,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.4603679180145264,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03986494988203049,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.4589664936065674,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.04280020296573639,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.4624319076538086,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.03930043429136276,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.47080135345459,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.0314876064658165,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.458728790283203,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.03189660236239433,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4467148780822754,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03264275938272476,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4613699913024902,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03021484799683094,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.4535279273986816,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.030740290880203247,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.4621875286102295,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03096810355782509,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.458500385284424,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03105205111205578,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.461958646774292,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.03330068290233612,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.449777364730835,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.032831475138664246,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.4616191387176514,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.0387776717543602,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.445256471633911,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.041803423315286636,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.4545960426330566,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.04148292914032936,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.433408737182617,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.04648579657077789,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.449134349822998,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.04186409339308739,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.4449565410614014,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.038569554686546326,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.443675994873047,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03908300772309303,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4626526832580566,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.038897737860679626,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.438591957092285,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.0360480472445488,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.436077117919922,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0429830439388752,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4384970664978027,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.04465422406792641,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.458029270172119,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.037096526473760605,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.436816930770874,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.047823466360569,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.44439959526062,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.05929310992360115,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.4599215984344482,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.053512319922447205,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4387903213500977,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.042625587433576584,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4359588623046875,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.056042976677417755,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.4441447257995605,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05813189223408699,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4399144649505615,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.06313478201627731,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.453447103500366,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.05281798169016838,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.4428963661193848,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.03991910442709923,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4359219074249268,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.047981832176446915,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4506607055664062,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.03646523132920265,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8225739002227783,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6196661591529846,
+ "eval_runtime": 43.8106,
+ "eval_samples_per_second": 55.74,
+ "eval_steps_per_second": 0.457,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.03673036769032478,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.4227209091186523,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.044547371566295624,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.4073169231414795,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.04089285805821419,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4279446601867676,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.03538106009364128,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.4234392642974854,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.03152978792786598,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4214606285095215,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.03324202075600624,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.4085447788238525,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.03352442383766174,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.417776107788086,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.03582802042365074,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.411144256591797,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.0351114459335804,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.405872344970703,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.038898173719644547,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4029338359832764,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.03926656022667885,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4100089073181152,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.04285634309053421,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.410360336303711,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.04179581627249718,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4093637466430664,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.03356701880693436,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4118056297302246,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.038140635937452316,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.4154677391052246,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.04468780383467674,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416846513748169,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.062421441078186035,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4157638549804688,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06317313015460968,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.418593406677246,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.049498945474624634,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4204845428466797,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.046322405338287354,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.403489589691162,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.053081002086400986,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.4148740768432617,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05222945287823677,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.416757583618164,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04893574118614197,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4241795539855957,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04951956123113632,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.401386260986328,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04019852355122566,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.4168622493743896,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.03625848889350891,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4026846885681152,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.03464345261454582,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4021215438842773,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.03750818222761154,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4113571643829346,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.0360100194811821,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.3977622985839844,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.03464623540639877,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.413273334503174,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.03193575143814087,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.423659324645996,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.02710454724729061,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.396085262298584,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.02830432914197445,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.420255422592163,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.029717663303017616,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.3958120346069336,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.030005933716893196,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3908543586730957,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.029054716229438782,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.3873138427734375,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.031556472182273865,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3892760276794434,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03518872708082199,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4113478660583496,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.03471249341964722,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.384857654571533,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03223155066370964,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3840765953063965,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.030012644827365875,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.388247489929199,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.03162285313010216,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.3987512588500977,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.03700442612171173,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.3993430137634277,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.03710822016000748,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.387359380722046,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.028411181643605232,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3925249576568604,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031246855854988098,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.38845157623291,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.034664880484342575,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.392436981201172,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.03354701027274132,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.390890121459961,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.03527764603495598,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.3954997062683105,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03325049206614494,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3847177028656006,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.028021875768899918,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.385910987854004,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.02964954823255539,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3787484169006348,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.03068520873785019,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.400784969329834,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03416873514652252,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.391033411026001,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.03534875437617302,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.3796474933624268,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.03732474148273468,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3831934928894043,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.03648678958415985,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3785600662231445,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03801294043660164,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.387587070465088,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.04186025261878967,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.3893089294433594,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.04416101053357124,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3852195739746094,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.04137178510427475,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.391308307647705,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.05010323226451874,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.3906192779541016,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.049587685614824295,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.3846192359924316,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.04125971347093582,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.384507179260254,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.039676666259765625,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.38167142868042,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.038775283843278885,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.3789167404174805,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.03634221479296684,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.383060932159424,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03464275971055031,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.388552665710449,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.031778786331415176,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.3903298377990723,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.027904309332370758,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.377493381500244,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.028964119032025337,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.381959915161133,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.0233962032943964,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7810559272766113,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6070130467414856,
+ "eval_runtime": 50.871,
+ "eval_samples_per_second": 48.004,
+ "eval_steps_per_second": 0.393,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.030209315940737724,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.3569841384887695,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.04112662002444267,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.376680374145508,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.055858466774225235,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.3621644973754883,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.06746986508369446,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3607563972473145,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.06736797094345093,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3548707962036133,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.0562627948820591,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.364654541015625,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.051833927631378174,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.3748884201049805,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.04824110120534897,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.3502087593078613,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.05518793687224388,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3687515258789062,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.06564056873321533,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.367131233215332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.0605909489095211,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.3589868545532227,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.05744732916355133,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.370117664337158,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.06430432945489883,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.3703484535217285,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.06186915934085846,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3720293045043945,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.05977998673915863,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.3512699604034424,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.06413097679615021,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.377257823944092,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.053935080766677856,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.350717544555664,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.049110863357782364,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.3688902854919434,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.04313219338655472,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3374459743499756,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.039556123316287994,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3482556343078613,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.03796547278761864,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3547937870025635,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.03646104037761688,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.344686269760132,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.03355357423424721,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.3588008880615234,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.03004707582294941,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.352043628692627,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.030186176300048828,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3500709533691406,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.029748762026429176,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.3523058891296387,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.03238696977496147,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3635129928588867,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.032599564641714096,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.346269130706787,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.028626957908272743,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.357151746749878,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.02909751422703266,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3557677268981934,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.029951099306344986,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.34753680229187,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.029768310487270355,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3408985137939453,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.03005973994731903,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.362542152404785,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03360462561249733,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.352529287338257,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.034002047032117844,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.3334460258483887,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.03292316198348999,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3446991443634033,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.0337117575109005,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.350605010986328,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03428212180733681,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.357408046722412,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.03169601410627365,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.349088191986084,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.03389411047101021,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.361576795578003,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.036497049033641815,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.35420560836792,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03574785590171814,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3678526878356934,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.030757993459701538,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3520517349243164,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.0289121363312006,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.337864637374878,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.028012607246637344,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.353999614715576,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.027558492496609688,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.342132568359375,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.02871510572731495,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3425655364990234,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03070886805653572,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3401665687561035,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.03056657873094082,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.345277786254883,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.030907969921827316,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.346494674682617,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0318334735929966,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.3390543460845947,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.029700733721256256,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.3398728370666504,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.029687359929084778,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.3474364280700684,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.027234600856900215,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3457484245300293,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.02823043428361416,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.337700366973877,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03002646192908287,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.335954189300537,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.027246717363595963,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.330207347869873,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.029703574255108833,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.328280448913574,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.028796236962080002,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3448386192321777,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.027944380417466164,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.320234537124634,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.030626650899648666,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.334698438644409,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.029164662584662437,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3450326919555664,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03034045360982418,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.3420588970184326,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.02856968529522419,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3396072387695312,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.02892562374472618,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.340027332305908,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.02938242256641388,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.33550763130188,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.029261089861392975,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3246421813964844,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.030340535566210747,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.3465514183044434,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.033335570245981216,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.3467206954956055,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.03452218323945999,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.343702793121338,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.03326255828142166,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.337631940841675,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025433633476495743,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7402067184448242,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03175730258226395,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.62623405456543,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.04780422896146774,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.623239517211914,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.0504758283495903,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.6369194984436035,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.04481920227408409,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.57640266418457,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0440552793443203,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.630992889404297,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0503336526453495,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.627910137176514,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.053885314613580704,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.637897491455078,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.05481956899166107,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.610715866088867,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.050510745495557785,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.639884948730469,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.04579266160726547,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.590188980102539,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.04817410930991173,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.636295318603516,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.058862220495939255,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.614317417144775,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.06328638643026352,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.655928611755371,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.058752477169036865,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.652004241943359,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.06049015372991562,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.6513800621032715,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.058941151946783066,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.657688140869141,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.05542639642953873,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.637376308441162,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06024057790637016,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.63592529296875,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.06040216609835625,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.6275787353515625,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.059600673615932465,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.652995586395264,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.05379730463027954,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.597358226776123,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05355260521173477,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.6106367111206055,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.06649646162986755,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.654443740844727,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.058309994637966156,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.6869659423828125,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.05403570458292961,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.654750823974609,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04678157716989517,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.653955459594727,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04297948628664017,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.644915580749512,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.04235142841935158,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.662282943725586,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.0405130498111248,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.64569091796875,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.03351007401943207,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.653165817260742,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.03182631731033325,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.606291770935059,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.031552448868751526,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.642643928527832,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027941036969423294,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.623664379119873,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.027744563296437263,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.605886459350586,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02962736040353775,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.606762886047363,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.027251560240983963,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.620492935180664,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.027894044294953346,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.613903999328613,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.02778882533311844,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.634457588195801,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.028082096949219704,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.610980033874512,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.026210976764559746,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.612090110778809,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.026537755504250526,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.581171035766602,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.028709523379802704,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.628799915313721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.028233813121914864,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.586182117462158,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.02966046892106533,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.62320613861084,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.03198636695742607,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.6018524169921875,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.033404890447854996,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.610121726989746,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.029217196628451347,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.601312160491943,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.029335757717490196,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.620820999145508,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.031007781624794006,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.626503944396973,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.028453458100557327,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.589323997497559,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02526312880218029,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.598350524902344,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.027317512780427933,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.62388277053833,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02710605598986149,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.606523513793945,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.027491388842463493,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.586644172668457,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.027158567681908607,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.606039047241211,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.0298276599496603,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.593776702880859,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.026622479781508446,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.59510612487793,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.028238043189048767,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.5899505615234375,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.029481008648872375,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.6092071533203125,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.030138323083519936,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.600069046020508,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.02951825223863125,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.600125789642334,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.029242202639579773,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.590723037719727,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.02941613830626011,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.623342514038086,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.029764151200652122,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.631185531616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.031503792852163315,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.605094909667969,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.03184918314218521,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.592767715454102,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.03223618119955063,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.5930585861206055,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.0340263694524765,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.609347343444824,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.03245621174573898,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.598440170288086,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.03204867243766785,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.619702339172363,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.030866602435708046,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.609165191650391,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.01960863545536995,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8655829429626465,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.593852162361145,
+ "eval_runtime": 60.0971,
+ "eval_samples_per_second": 40.634,
+ "eval_steps_per_second": 0.649,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03552858531475067,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.539240837097168,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.0536109134554863,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.52609395980835,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.061411239206790924,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.536109447479248,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.06352879852056503,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.546850204467773,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.05683520436286926,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.563828468322754,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.05522637069225311,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.550929069519043,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.058990392833948135,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.554411888122559,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.057443976402282715,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.545921325683594,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.05383576825261116,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.5481414794921875,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04797440022230148,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.55556583404541,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04923664405941963,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.545513153076172,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.056191347539424896,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.536705017089844,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.06088989973068237,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.570624351501465,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.04802761599421501,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.526930809020996,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.04018320515751839,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.562432289123535,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.04280981421470642,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.559737205505371,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.04258878529071808,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.561749458312988,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.04411563277244568,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.562010765075684,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.041077401489019394,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.573984622955322,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.04125386103987694,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.550564289093018,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.041974026709795,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.552739143371582,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03689907118678093,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.540107727050781,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03586242347955704,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.554999351501465,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.036602675914764404,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.546337127685547,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.03707926720380783,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.543377876281738,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.03322125971317291,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.574325084686279,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03359243646264076,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.5613603591918945,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.0314965583384037,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.530994415283203,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.031806789338588715,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.56138801574707,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.034988485276699066,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.553149700164795,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.037627559155225754,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.555401802062988,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03723530471324921,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.52452278137207,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.036220453679561615,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.54799747467041,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03228041157126427,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.574814319610596,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.030872676521539688,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.539666652679443,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.033935241401195526,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.579252243041992,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.034338828176259995,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.538324356079102,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.03441951423883438,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.540358543395996,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03328141197562218,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.520340919494629,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03224223107099533,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.5646209716796875,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.02999890223145485,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.504915714263916,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.03014362044632435,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.516615867614746,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.029469173401594162,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.560475826263428,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.027887821197509766,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.586333751678467,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.029970254749059677,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.530396461486816,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.031648170202970505,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.554404258728027,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.0325581431388855,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.544102668762207,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.03555625304579735,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.543701171875,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.033053912222385406,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.508375644683838,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.03059076890349388,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.545750617980957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.029640182852745056,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.54715633392334,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030139951035380363,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.533708572387695,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029753828421235085,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.543603897094727,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.029481034725904465,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.534200668334961,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.028991524130105972,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.5683674812316895,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.028145765885710716,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.543053150177002,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.02999844029545784,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.5482354164123535,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03134117275476456,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.5360107421875,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.029277140274643898,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.5729827880859375,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028516866266727448,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.543548583984375,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028387080878019333,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.552616596221924,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.028977863490581512,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.5456695556640625,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.028010815382003784,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.534921646118164,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.030222605913877487,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.561526298522949,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.03140007704496384,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.500124931335449,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.029269184917211533,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.536653995513916,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.02935507707297802,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.521398544311523,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02801917865872383,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.527442455291748,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.027995768934488297,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.556139945983887,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.028441084548830986,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.555442810058594,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.029100807383656502,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.532083511352539,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020385215058922768,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.8488006591796875,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5902590751647949,
+ "eval_runtime": 60.1424,
+ "eval_samples_per_second": 40.604,
+ "eval_steps_per_second": 0.648,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03513244539499283,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.486500263214111,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.052679311484098434,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.489691734313965,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.06171489134430885,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.471555709838867,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.06164781004190445,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.486891746520996,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.06410388648509979,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.494929313659668,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06399523466825485,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.5334320068359375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.0636502280831337,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.5039381980896,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.06139931082725525,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.512765884399414,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.07460282742977142,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.477229118347168,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.075187548995018,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.484212398529053,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.07584068924188614,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.495598793029785,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.07192158699035645,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.507598400115967,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.05366869643330574,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.514532089233398,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.06397061794996262,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.512920379638672,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.06391509622335434,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.465702056884766,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.059490568935871124,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.5351881980896,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.057501111179590225,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.510270595550537,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.056947533041238785,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.509241104125977,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.050393324345350266,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.50428581237793,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.057608287781476974,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.474919319152832,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.054470717906951904,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.521528244018555,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.054405540227890015,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.510211944580078,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.051328495144844055,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.480720520019531,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.045930180698633194,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.503024101257324,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.04298274591565132,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.49046516418457,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.0406254380941391,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.544064521789551,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.03557496517896652,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.496966361999512,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.0364493802189827,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.5047502517700195,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.034521300345659256,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.485365867614746,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.03305799141526222,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.484884738922119,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.033053625375032425,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.478784561157227,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03426017984747887,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.482428073883057,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.03638783469796181,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.433219909667969,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.03597250208258629,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.522584915161133,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.03086097538471222,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.500803470611572,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.029642771929502487,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.444099426269531,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.03190716728568077,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.493034362792969,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.0312873050570488,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.487887382507324,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03201943263411522,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.479177474975586,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.033088769763708115,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.5020246505737305,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.02975946106016636,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.455426216125488,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03213953226804733,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.501615047454834,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.03262908384203911,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.485121250152588,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.031102091073989868,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.488893508911133,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.02945449948310852,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.490121841430664,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.033680472522974014,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.51833438873291,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.03277525305747986,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.480165481567383,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.03303041681647301,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.501155376434326,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.034927863627672195,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.513175010681152,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.03098093718290329,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.503798484802246,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.029513906687498093,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.49006986618042,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029630785807967186,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.522514343261719,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.030888330191373825,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.4895734786987305,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.032392196357250214,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.471953868865967,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.03215661272406578,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.48297643661499,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.03346327692270279,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.518160820007324,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.031311362981796265,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.484991550445557,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.03056488186120987,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.525470733642578,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.03043230250477791,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.510650634765625,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.02916652522981167,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.496840953826904,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.030673664063215256,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.494868278503418,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030033906921744347,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.479182720184326,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.028828797861933708,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.471983909606934,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028316056355834007,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.5049147605896,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.03158092871308327,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.449270248413086,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.03041706420481205,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.508289813995361,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.02967553399503231,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.465579032897949,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030721619725227356,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.471003532409668,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03066832385957241,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.488987445831299,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.032297998666763306,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.491018295288086,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03221398591995239,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.471237659454346,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.021120507270097733,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.8080053329467773,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5888046026229858,
+ "eval_runtime": 59.1638,
+ "eval_samples_per_second": 41.275,
+ "eval_steps_per_second": 0.659,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.03799329698085785,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.401787757873535,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.05785121023654938,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.408820629119873,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.05964238941669464,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.427263259887695,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05466582626104355,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.403415679931641,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.05613672360777855,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.413750648498535,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.05396881327033043,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.41916561126709,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.05689029023051262,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.395590782165527,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.056090328842401505,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.421253204345703,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.0473211295902729,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.385158061981201,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.04464765638113022,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.431746959686279,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.046421997249126434,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.428816795349121,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.05039866268634796,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.452417850494385,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.05491664633154869,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.447483062744141,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05866196006536484,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.434992790222168,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.05532094091176987,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.450716018676758,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.051083534955978394,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.402632713317871,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.03815016150474548,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.433640956878662,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04203395918011665,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.415888786315918,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.043729111552238464,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.423260688781738,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.039426498115062714,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.428949356079102,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.03837858885526657,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.434070587158203,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.038315922021865845,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.471246719360352,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.03963236138224602,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.4434990882873535,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04027022793889046,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.436992645263672,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.0409243181347847,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.426936149597168,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03404631093144417,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.43462610244751,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.034393735229969025,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.437409400939941,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03635180741548538,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.451286315917969,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03349733352661133,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.419209003448486,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.03238677605986595,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.456454277038574,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.034473925828933716,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.3988728523254395,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.03455805033445358,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.434884071350098,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.037486057728528976,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.466741561889648,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.03617125004529953,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.429305076599121,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.033646389842033386,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.407780170440674,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03401118144392967,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.42977237701416,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.036534134298563004,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.451094627380371,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.0384049266576767,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.414777755737305,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03460627421736717,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.420644760131836,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.033746346831321716,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.461175918579102,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.037153810262680054,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.444699287414551,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03283539041876793,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.431022644042969,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03248969465494156,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.447571277618408,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03327852487564087,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.42236328125,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.031114818528294563,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.418244361877441,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.03231207653880119,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.455004692077637,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.030702251940965652,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.483610153198242,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031027289107441902,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.4470109939575195,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.03129073604941368,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.452799320220947,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.03634301945567131,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4661664962768555,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03718913346529007,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.412006855010986,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.034018341451883316,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.436755180358887,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032502319663763046,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.452866554260254,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.032902952283620834,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.449224472045898,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.0340626984834671,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.445425510406494,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.033442508429288864,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.432218551635742,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.03489428386092186,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.4205827713012695,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.031868621706962585,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.414196968078613,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.032986633479595184,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.46116304397583,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.03348355367779732,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.424936771392822,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03174544870853424,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.434438705444336,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.03293656185269356,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.440502166748047,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.035124607384204865,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.440943717956543,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03438971936702728,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.431408882141113,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.03537093102931976,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.43867301940918,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.03267325833439827,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.462226867675781,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.030995232984423637,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.433740615844727,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.03384929150342941,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.508179664611816,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.032025814056396484,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.407378196716309,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03288578242063522,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.455531120300293,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.032061848789453506,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.452212333679199,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022320276126265526,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7567386627197266,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5878658294677734,
+ "eval_runtime": 59.5054,
+ "eval_samples_per_second": 41.038,
+ "eval_steps_per_second": 0.655,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.039114587008953094,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.347412586212158,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.05164389684796333,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.391279220581055,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055882278829813004,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.355602264404297,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.05625041201710701,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3934855461120605,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05861108750104904,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.364226341247559,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.052431970834732056,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.363334655761719,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.06147193908691406,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.384820938110352,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.06701470166444778,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.373311996459961,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.06714873015880585,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.32745361328125,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.06684907525777817,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.373660087585449,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.05738082900643349,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.362325668334961,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05306725949048996,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.351685047149658,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.05831029266119003,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.374813079833984,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05547972396016121,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.361842632293701,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.048057056963443756,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.383205413818359,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.04682513326406479,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.365846157073975,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.05001377314329147,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.362151145935059,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.05140263959765434,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.376064300537109,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.04640384018421173,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.363801002502441,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04175020009279251,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.3477983474731445,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.04134776070713997,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.337469100952148,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.042178574949502945,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.351490020751953,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.03989115729928017,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.40725040435791,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04086652770638466,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.341119766235352,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.03931201249361038,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.3786468505859375,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.03855905309319496,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.3701605796813965,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.03803952783346176,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.401841163635254,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.036908265203237534,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.410647392272949,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.0398186556994915,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.380431175231934,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03899035602807999,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.374293804168701,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.03946344554424286,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.396918773651123,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.03738604858517647,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.404016017913818,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.03579043596982956,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.396914958953857,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.037139154970645905,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.407739639282227,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.03581099584698677,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.411419868469238,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.03548453748226166,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.332690238952637,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.037561237812042236,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.381314277648926,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.035827480256557465,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.371536731719971,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03538314625620842,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.395467758178711,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.036672890186309814,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.420590400695801,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.038981080055236816,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.395296096801758,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.038234200328588486,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.415799140930176,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.038830097764730453,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.413394927978516,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03990919142961502,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.39515495300293,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.036424681544303894,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.39730978012085,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03488154709339142,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.403696060180664,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.0363851860165596,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.368661403656006,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.035461440682411194,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.396518707275391,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03647598251700401,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.440667152404785,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.03667188808321953,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.3781352043151855,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034516457468271255,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.414007186889648,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.03536032512784004,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.42198371887207,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.037454646080732346,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.409174919128418,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03632275015115738,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3547682762146,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.038922134786844254,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.39993143081665,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.03562692925333977,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.409564971923828,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.033203717321157455,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.421782493591309,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03830202296376228,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.430092811584473,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.03835884481668472,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.379505634307861,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.037440430372953415,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.403937339782715,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03732272610068321,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.440683364868164,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03325957432389259,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.3949294090271,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.035377148538827896,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.400777816772461,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.03725374862551689,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.422494411468506,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03601110726594925,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.4008941650390625,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03193027526140213,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.400361061096191,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.034382354468107224,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.408535003662109,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.03747732564806938,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.412189483642578,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03641800582408905,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.381453037261963,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.0315779373049736,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.4169769287109375,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03583509847521782,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.385746955871582,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02574268914759159,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7456769943237305,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.588273823261261,
+ "eval_runtime": 59.9356,
+ "eval_samples_per_second": 40.744,
+ "eval_steps_per_second": 0.651,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013961605584642,
+ "grad_norm": 0.037516895681619644,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 4.33476448059082,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027923211169284,
+ "grad_norm": 0.048450272530317307,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 4.296454429626465,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041884816753926,
+ "grad_norm": 0.05368480458855629,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 4.314815998077393,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055846422338568,
+ "grad_norm": 0.05520009621977806,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 4.330798625946045,
+ "step": 1084
+ },
+ {
+ "epoch": 15.06980802792321,
+ "grad_norm": 0.05928800627589226,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 4.317663192749023,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083769633507853,
+ "grad_norm": 0.06315027177333832,
+ "learning_rate": 0.000491285979372924,
+ "loss": 4.329221725463867,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097731239092496,
+ "grad_norm": 0.06342213600873947,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 4.32546329498291,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111692844677139,
+ "grad_norm": 0.0667002946138382,
+ "learning_rate": 0.000490873338088198,
+ "loss": 4.309828758239746,
+ "step": 1088
+ },
+ {
+ "epoch": 15.12565445026178,
+ "grad_norm": 0.05264135077595711,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 4.307073593139648,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139616055846423,
+ "grad_norm": 0.051895879209041595,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 4.333526611328125,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153577661431065,
+ "grad_norm": 0.05508396774530411,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 4.330171585083008,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167539267015707,
+ "grad_norm": 0.057272691279649734,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 4.329268455505371,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181500872600349,
+ "grad_norm": 0.058956146240234375,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 4.341983795166016,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195462478184991,
+ "grad_norm": 0.0557059645652771,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 4.348274230957031,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209424083769633,
+ "grad_norm": 0.05661974847316742,
+ "learning_rate": 0.000489424334303338,
+ "loss": 4.347631454467773,
+ "step": 1095
+ },
+ {
+ "epoch": 15.223385689354275,
+ "grad_norm": 0.05338657647371292,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 4.334639072418213,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237347294938917,
+ "grad_norm": 0.05091148987412453,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 4.355465412139893,
+ "step": 1097
+ },
+ {
+ "epoch": 15.25130890052356,
+ "grad_norm": 0.04996965080499649,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 4.346457004547119,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265270506108202,
+ "grad_norm": 0.05573750659823418,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 4.346520900726318,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279232111692846,
+ "grad_norm": 0.05776917189359665,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 4.359007835388184,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293193717277488,
+ "grad_norm": 0.050820041447877884,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 4.331710338592529,
+ "step": 1101
+ },
+ {
+ "epoch": 15.30715532286213,
+ "grad_norm": 0.04441820830106735,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 4.3126420974731445,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321116928446772,
+ "grad_norm": 0.04582618921995163,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 4.312925338745117,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335078534031414,
+ "grad_norm": 0.043592359870672226,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 4.345674514770508,
+ "step": 1104
+ },
+ {
+ "epoch": 15.349040139616056,
+ "grad_norm": 0.04069037735462189,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 4.334713935852051,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363001745200698,
+ "grad_norm": 0.04083041846752167,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 4.358171463012695,
+ "step": 1106
+ },
+ {
+ "epoch": 15.37696335078534,
+ "grad_norm": 0.04119594022631645,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 4.317891597747803,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390924956369982,
+ "grad_norm": 0.042060643434524536,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 4.355810165405273,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404886561954624,
+ "grad_norm": 0.04721749201416969,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 4.353546619415283,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418848167539267,
+ "grad_norm": 0.04657730832695961,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 4.341683387756348,
+ "step": 1110
+ },
+ {
+ "epoch": 15.432809773123909,
+ "grad_norm": 0.04096715524792671,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 4.357192039489746,
+ "step": 1111
+ },
+ {
+ "epoch": 15.44677137870855,
+ "grad_norm": 0.03904155269265175,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 4.354775428771973,
+ "step": 1112
+ },
+ {
+ "epoch": 15.460732984293193,
+ "grad_norm": 0.03828521445393562,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 4.356499671936035,
+ "step": 1113
+ },
+ {
+ "epoch": 15.474694589877837,
+ "grad_norm": 0.03688685595989227,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 4.339485168457031,
+ "step": 1114
+ },
+ {
+ "epoch": 15.488656195462479,
+ "grad_norm": 0.03945618495345116,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 4.363500118255615,
+ "step": 1115
+ },
+ {
+ "epoch": 15.502617801047121,
+ "grad_norm": 0.03821462020277977,
+ "learning_rate": 0.000485033394934934,
+ "loss": 4.353529930114746,
+ "step": 1116
+ },
+ {
+ "epoch": 15.516579406631763,
+ "grad_norm": 0.036265499889850616,
+ "learning_rate": 0.000484822676912974,
+ "loss": 4.337663173675537,
+ "step": 1117
+ },
+ {
+ "epoch": 15.530541012216405,
+ "grad_norm": 0.03803930804133415,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 4.34660530090332,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544502617801047,
+ "grad_norm": 0.038425102829933167,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 4.373246192932129,
+ "step": 1119
+ },
+ {
+ "epoch": 15.55846422338569,
+ "grad_norm": 0.036852262914180756,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 4.331226825714111,
+ "step": 1120
+ },
+ {
+ "epoch": 15.572425828970331,
+ "grad_norm": 0.03700809180736542,
+ "learning_rate": 0.000483978341612154,
+ "loss": 4.375818729400635,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586387434554974,
+ "grad_norm": 0.03628664091229439,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 4.32705020904541,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600349040139616,
+ "grad_norm": 0.036831192672252655,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 4.337127208709717,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614310645724258,
+ "grad_norm": 0.03736541047692299,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 4.36502742767334,
+ "step": 1124
+ },
+ {
+ "epoch": 15.6282722513089,
+ "grad_norm": 0.03658579662442207,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 4.359844207763672,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642233856893542,
+ "grad_norm": 0.03688633441925049,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 4.3355793952941895,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656195462478184,
+ "grad_norm": 0.03477008268237114,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 4.333405494689941,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670157068062828,
+ "grad_norm": 0.03655015677213669,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 4.338570594787598,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68411867364747,
+ "grad_norm": 0.03664765506982803,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 4.352496147155762,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698080279232112,
+ "grad_norm": 0.037019114941358566,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 4.335735321044922,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712041884816754,
+ "grad_norm": 0.03949085995554924,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 4.342585563659668,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726003490401396,
+ "grad_norm": 0.04084350913763046,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 4.329622268676758,
+ "step": 1132
+ },
+ {
+ "epoch": 15.739965095986038,
+ "grad_norm": 0.038439832627773285,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 4.360318660736084,
+ "step": 1133
+ },
+ {
+ "epoch": 15.75392670157068,
+ "grad_norm": 0.03932987526059151,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 4.343446254730225,
+ "step": 1134
+ },
+ {
+ "epoch": 15.767888307155323,
+ "grad_norm": 0.041595373302698135,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 4.351495742797852,
+ "step": 1135
+ },
+ {
+ "epoch": 15.781849912739965,
+ "grad_norm": 0.03917490318417549,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 4.380537033081055,
+ "step": 1136
+ },
+ {
+ "epoch": 15.795811518324607,
+ "grad_norm": 0.038597021251916885,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 4.3527116775512695,
+ "step": 1137
+ },
+ {
+ "epoch": 15.809773123909249,
+ "grad_norm": 0.03651711344718933,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 4.369256496429443,
+ "step": 1138
+ },
+ {
+ "epoch": 15.823734729493891,
+ "grad_norm": 0.035118963569402695,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 4.350616455078125,
+ "step": 1139
+ },
+ {
+ "epoch": 15.837696335078533,
+ "grad_norm": 0.034945692867040634,
+ "learning_rate": 0.000479936075927473,
+ "loss": 4.365446090698242,
+ "step": 1140
+ },
+ {
+ "epoch": 15.851657940663177,
+ "grad_norm": 0.03519820421934128,
+ "learning_rate": 0.000479721889242305,
+ "loss": 4.357390403747559,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86561954624782,
+ "grad_norm": 0.03407098725438118,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 4.350573539733887,
+ "step": 1142
+ },
+ {
+ "epoch": 15.879581151832461,
+ "grad_norm": 0.036511123180389404,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 4.328500270843506,
+ "step": 1143
+ },
+ {
+ "epoch": 15.893542757417103,
+ "grad_norm": 0.039194293320178986,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 4.346471786499023,
+ "step": 1144
+ },
+ {
+ "epoch": 15.907504363001745,
+ "grad_norm": 0.03777911514043808,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 4.358243942260742,
+ "step": 1145
+ },
+ {
+ "epoch": 15.921465968586388,
+ "grad_norm": 0.0354655385017395,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 4.368768692016602,
+ "step": 1146
+ },
+ {
+ "epoch": 15.93542757417103,
+ "grad_norm": 0.03710418567061424,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 4.3482184410095215,
+ "step": 1147
+ },
+ {
+ "epoch": 15.949389179755672,
+ "grad_norm": 0.03545419126749039,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 4.360692501068115,
+ "step": 1148
+ },
+ {
+ "epoch": 15.963350785340314,
+ "grad_norm": 0.035963475704193115,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 4.3592119216918945,
+ "step": 1149
+ },
+ {
+ "epoch": 15.977312390924956,
+ "grad_norm": 0.03425297513604164,
+ "learning_rate": 0.000477787822648007,
+ "loss": 4.329504013061523,
+ "step": 1150
+ },
+ {
+ "epoch": 15.991273996509598,
+ "grad_norm": 0.03453255817294121,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 4.298142433166504,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.025766054168343544,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 2.7026405334472656,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.5888971090316772,
+ "eval_runtime": 58.9721,
+ "eval_samples_per_second": 41.409,
+ "eval_steps_per_second": 0.661,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013961605584644,
+ "grad_norm": 0.04088004305958748,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 4.268601417541504,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027923211169284,
+ "grad_norm": 0.05194641277194023,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 4.267483711242676,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041884816753928,
+ "grad_norm": 0.05488700792193413,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 4.2527570724487305,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05584642233857,
+ "grad_norm": 0.059292279183864594,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 4.283864974975586,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069808027923212,
+ "grad_norm": 0.0567227303981781,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 4.291148662567139,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083769633507853,
+ "grad_norm": 0.05582835525274277,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 4.250524520874023,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097731239092496,
+ "grad_norm": 0.055709999054670334,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 4.259716033935547,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111692844677137,
+ "grad_norm": 0.053324732929468155,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 4.247231483459473,
+ "step": 1160
+ },
+ {
+ "epoch": 16.12565445026178,
+ "grad_norm": 0.05367910861968994,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 4.263905048370361,
+ "step": 1161
+ },
+ {
+ "epoch": 16.13961605584642,
+ "grad_norm": 0.057094886898994446,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 4.266510009765625,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153577661431065,
+ "grad_norm": 0.05921962857246399,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 4.278281211853027,
+ "step": 1163
+ },
+ {
+ "epoch": 16.167539267015705,
+ "grad_norm": 0.060246821492910385,
+ "learning_rate": 0.000474756648870413,
+ "loss": 4.2848052978515625,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18150087260035,
+ "grad_norm": 0.05686050280928612,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 4.286434173583984,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195462478184993,
+ "grad_norm": 0.05207265168428421,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 4.246121406555176,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209424083769633,
+ "grad_norm": 0.0477963425219059,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 4.287411689758301,
+ "step": 1167
+ },
+ {
+ "epoch": 16.223385689354277,
+ "grad_norm": 0.05586806684732437,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 4.2721333503723145,
+ "step": 1168
+ },
+ {
+ "epoch": 16.237347294938917,
+ "grad_norm": 0.05427060276269913,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 4.2751641273498535,
+ "step": 1169
+ },
+ {
+ "epoch": 16.25130890052356,
+ "grad_norm": 0.0518609918653965,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 4.287444114685059,
+ "step": 1170
+ },
+ {
+ "epoch": 16.2652705061082,
+ "grad_norm": 0.053606316447257996,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 4.287125587463379,
+ "step": 1171
+ },
+ {
+ "epoch": 16.279232111692846,
+ "grad_norm": 0.05118813365697861,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 4.291893005371094,
+ "step": 1172
+ },
+ {
+ "epoch": 16.293193717277486,
+ "grad_norm": 0.04774491861462593,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 4.325250148773193,
+ "step": 1173
+ },
+ {
+ "epoch": 16.30715532286213,
+ "grad_norm": 0.04634870961308479,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 4.311764717102051,
+ "step": 1174
+ },
+ {
+ "epoch": 16.32111692844677,
+ "grad_norm": 0.04411927983164787,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 4.277068138122559,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335078534031414,
+ "grad_norm": 0.043819356709718704,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 4.309638977050781,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349040139616054,
+ "grad_norm": 0.0418904684484005,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 4.306234836578369,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363001745200698,
+ "grad_norm": 0.04327298328280449,
+ "learning_rate": 0.000471698352726896,
+ "loss": 4.296874046325684,
+ "step": 1178
+ },
+ {
+ "epoch": 16.376963350785342,
+ "grad_norm": 0.045296184718608856,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 4.288957595825195,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390924956369982,
+ "grad_norm": 0.04447923228144646,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 4.313834190368652,
+ "step": 1180
+ },
+ {
+ "epoch": 16.404886561954626,
+ "grad_norm": 0.04548390209674835,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 4.284562587738037,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418848167539267,
+ "grad_norm": 0.04567817226052284,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 4.295994758605957,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43280977312391,
+ "grad_norm": 0.04277439042925835,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 4.262106895446777,
+ "step": 1183
+ },
+ {
+ "epoch": 16.44677137870855,
+ "grad_norm": 0.043597545474767685,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 4.294536590576172,
+ "step": 1184
+ },
+ {
+ "epoch": 16.460732984293195,
+ "grad_norm": 0.04424033313989639,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 4.297497749328613,
+ "step": 1185
+ },
+ {
+ "epoch": 16.474694589877835,
+ "grad_norm": 0.04354134574532509,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 4.321508884429932,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48865619546248,
+ "grad_norm": 0.04241178184747696,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 4.322726249694824,
+ "step": 1187
+ },
+ {
+ "epoch": 16.50261780104712,
+ "grad_norm": 0.039317622780799866,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 4.311155319213867,
+ "step": 1188
+ },
+ {
+ "epoch": 16.516579406631763,
+ "grad_norm": 0.04126536473631859,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 4.306636810302734,
+ "step": 1189
+ },
+ {
+ "epoch": 16.530541012216403,
+ "grad_norm": 0.040116794407367706,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 4.314671516418457,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544502617801047,
+ "grad_norm": 0.04045999422669411,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 4.301012992858887,
+ "step": 1191
+ },
+ {
+ "epoch": 16.55846422338569,
+ "grad_norm": 0.04013422504067421,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 4.29913330078125,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57242582897033,
+ "grad_norm": 0.03878502920269966,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 4.351600646972656,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586387434554975,
+ "grad_norm": 0.038935888558626175,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 4.27667236328125,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600349040139616,
+ "grad_norm": 0.04055892676115036,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 4.29019021987915,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61431064572426,
+ "grad_norm": 0.03959144279360771,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 4.327725887298584,
+ "step": 1196
+ },
+ {
+ "epoch": 16.6282722513089,
+ "grad_norm": 0.039878539741039276,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 4.297952651977539,
+ "step": 1197
+ },
+ {
+ "epoch": 16.642233856893544,
+ "grad_norm": 0.041209928691387177,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 4.313030242919922,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656195462478184,
+ "grad_norm": 0.04346393793821335,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 4.274901390075684,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670157068062828,
+ "grad_norm": 0.04090269282460213,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 4.330657958984375,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68411867364747,
+ "grad_norm": 0.04004506766796112,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 4.323106288909912,
+ "step": 1201
+ },
+ {
+ "epoch": 16.698080279232112,
+ "grad_norm": 0.041541460901498795,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 4.280947685241699,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712041884816752,
+ "grad_norm": 0.04033653065562248,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 4.341434955596924,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726003490401396,
+ "grad_norm": 0.04104280099272728,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 4.2931671142578125,
+ "step": 1204
+ },
+ {
+ "epoch": 16.739965095986037,
+ "grad_norm": 0.039398156106472015,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 4.316840648651123,
+ "step": 1205
+ },
+ {
+ "epoch": 16.75392670157068,
+ "grad_norm": 0.03980560600757599,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 4.325345039367676,
+ "step": 1206
+ },
+ {
+ "epoch": 16.767888307155324,
+ "grad_norm": 0.040656186640262604,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 4.32675838470459,
+ "step": 1207
+ },
+ {
+ "epoch": 16.781849912739965,
+ "grad_norm": 0.03775986284017563,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 4.314423561096191,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79581151832461,
+ "grad_norm": 0.04043284431099892,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 4.344202041625977,
+ "step": 1209
+ },
+ {
+ "epoch": 16.80977312390925,
+ "grad_norm": 0.04074503108859062,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 4.350123405456543,
+ "step": 1210
+ },
+ {
+ "epoch": 16.823734729493893,
+ "grad_norm": 0.04117420315742493,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 4.350917816162109,
+ "step": 1211
+ },
+ {
+ "epoch": 16.837696335078533,
+ "grad_norm": 0.03799610957503319,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 4.289987564086914,
+ "step": 1212
+ },
+ {
+ "epoch": 16.851657940663177,
+ "grad_norm": 0.04071265831589699,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 4.306744575500488,
+ "step": 1213
+ },
+ {
+ "epoch": 16.865619546247817,
+ "grad_norm": 0.04109715297818184,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 4.318753242492676,
+ "step": 1214
+ },
+ {
+ "epoch": 16.87958115183246,
+ "grad_norm": 0.03910145163536072,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 4.281325340270996,
+ "step": 1215
+ },
+ {
+ "epoch": 16.8935427574171,
+ "grad_norm": 0.03752733767032623,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 4.3231048583984375,
+ "step": 1216
+ },
+ {
+ "epoch": 16.907504363001745,
+ "grad_norm": 0.03752103075385094,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 4.351336479187012,
+ "step": 1217
+ },
+ {
+ "epoch": 16.921465968586386,
+ "grad_norm": 0.03675422817468643,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 4.282256126403809,
+ "step": 1218
+ },
+ {
+ "epoch": 16.93542757417103,
+ "grad_norm": 0.03850258141756058,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 4.322597503662109,
+ "step": 1219
+ },
+ {
+ "epoch": 16.949389179755673,
+ "grad_norm": 0.03692392632365227,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 4.320333957672119,
+ "step": 1220
+ },
+ {
+ "epoch": 16.963350785340314,
+ "grad_norm": 0.035366240888834,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 4.31341552734375,
+ "step": 1221
+ },
+ {
+ "epoch": 16.977312390924958,
+ "grad_norm": 0.037120968103408813,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 4.321141242980957,
+ "step": 1222
+ },
+ {
+ "epoch": 16.991273996509598,
+ "grad_norm": 0.03721754252910614,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 4.305757522583008,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.026698043569922447,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 2.716372489929199,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.5894472599029541,
+ "eval_runtime": 58.927,
+ "eval_samples_per_second": 41.441,
+ "eval_steps_per_second": 0.662,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013961605584644,
+ "grad_norm": 0.044532276690006256,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 4.243244171142578,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027923211169284,
+ "grad_norm": 0.050529688596725464,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 4.221382141113281,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041884816753928,
+ "grad_norm": 0.05690870061516762,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 4.240285873413086,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05584642233857,
+ "grad_norm": 0.054803814738988876,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 4.213047027587891,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069808027923212,
+ "grad_norm": 0.051825862377882004,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 4.226880073547363,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083769633507853,
+ "grad_norm": 0.049802035093307495,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 4.25303840637207,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097731239092496,
+ "grad_norm": 0.05290328338742256,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 4.199422836303711,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111692844677137,
+ "grad_norm": 0.0531814880669117,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 4.201648712158203,
+ "step": 1232
+ },
+ {
+ "epoch": 17.12565445026178,
+ "grad_norm": 0.048972588032484055,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 4.207578182220459,
+ "step": 1233
+ },
+ {
+ "epoch": 17.13961605584642,
+ "grad_norm": 0.051345713436603546,
+ "learning_rate": 0.000459203665939781,
+ "loss": 4.206205368041992,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153577661431065,
+ "grad_norm": 0.059012677520513535,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 4.228360176086426,
+ "step": 1235
+ },
+ {
+ "epoch": 17.167539267015705,
+ "grad_norm": 0.05984492972493172,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 4.234362602233887,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18150087260035,
+ "grad_norm": 0.060886185616254807,
+ "learning_rate": 0.000458522886595554,
+ "loss": 4.253525257110596,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195462478184993,
+ "grad_norm": 0.06476911902427673,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 4.249889373779297,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209424083769633,
+ "grad_norm": 0.059773124754428864,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 4.225249290466309,
+ "step": 1239
+ },
+ {
+ "epoch": 17.223385689354277,
+ "grad_norm": 0.06063077598810196,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 4.25465202331543,
+ "step": 1240
+ },
+ {
+ "epoch": 17.237347294938917,
+ "grad_norm": 0.059922896325588226,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 4.252101898193359,
+ "step": 1241
+ },
+ {
+ "epoch": 17.25130890052356,
+ "grad_norm": 0.05849695950746536,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 4.268218040466309,
+ "step": 1242
+ },
+ {
+ "epoch": 17.2652705061082,
+ "grad_norm": 0.05769657716155052,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 4.2395782470703125,
+ "step": 1243
+ },
+ {
+ "epoch": 17.279232111692846,
+ "grad_norm": 0.058189719915390015,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 4.257606506347656,
+ "step": 1244
+ },
+ {
+ "epoch": 17.293193717277486,
+ "grad_norm": 0.059679049998521805,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 4.278588771820068,
+ "step": 1245
+ },
+ {
+ "epoch": 17.30715532286213,
+ "grad_norm": 0.0579286590218544,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 4.25511360168457,
+ "step": 1246
+ },
+ {
+ "epoch": 17.32111692844677,
+ "grad_norm": 0.05269965901970863,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 4.241326332092285,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335078534031414,
+ "grad_norm": 0.04828358441591263,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 4.243182182312012,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349040139616054,
+ "grad_norm": 0.050172045826911926,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 4.295144081115723,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363001745200698,
+ "grad_norm": 0.047297943383455276,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 4.250682830810547,
+ "step": 1250
+ },
+ {
+ "epoch": 17.376963350785342,
+ "grad_norm": 0.04574039950966835,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 4.269444465637207,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390924956369982,
+ "grad_norm": 0.042310427874326706,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 4.255361557006836,
+ "step": 1252
+ },
+ {
+ "epoch": 17.404886561954626,
+ "grad_norm": 0.04268861189484596,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 4.252292156219482,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418848167539267,
+ "grad_norm": 0.0424286350607872,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 4.24439811706543,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43280977312391,
+ "grad_norm": 0.042049869894981384,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 4.236266136169434,
+ "step": 1255
+ },
+ {
+ "epoch": 17.44677137870855,
+ "grad_norm": 0.04185241088271141,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 4.268826484680176,
+ "step": 1256
+ },
+ {
+ "epoch": 17.460732984293195,
+ "grad_norm": 0.04423609375953674,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 4.240240573883057,
+ "step": 1257
+ },
+ {
+ "epoch": 17.474694589877835,
+ "grad_norm": 0.041972566395998,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 4.269453048706055,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48865619546248,
+ "grad_norm": 0.04493343085050583,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 4.270544052124023,
+ "step": 1259
+ },
+ {
+ "epoch": 17.50261780104712,
+ "grad_norm": 0.04502921551465988,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 4.271880149841309,
+ "step": 1260
+ },
+ {
+ "epoch": 17.516579406631763,
+ "grad_norm": 0.044518254697322845,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 4.249786376953125,
+ "step": 1261
+ },
+ {
+ "epoch": 17.530541012216403,
+ "grad_norm": 0.044482357800006866,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 4.300779819488525,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544502617801047,
+ "grad_norm": 0.0424671471118927,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 4.248597621917725,
+ "step": 1263
+ },
+ {
+ "epoch": 17.55846422338569,
+ "grad_norm": 0.04136480763554573,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 4.276763916015625,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57242582897033,
+ "grad_norm": 0.04094843938946724,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 4.263628005981445,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586387434554975,
+ "grad_norm": 0.04415388032793999,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 4.257946014404297,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600349040139616,
+ "grad_norm": 0.04177384823560715,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 4.289669036865234,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61431064572426,
+ "grad_norm": 0.040863484144210815,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 4.320903778076172,
+ "step": 1268
+ },
+ {
+ "epoch": 17.6282722513089,
+ "grad_norm": 0.04025879129767418,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 4.268717288970947,
+ "step": 1269
+ },
+ {
+ "epoch": 17.642233856893544,
+ "grad_norm": 0.04161325842142105,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 4.275268077850342,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656195462478184,
+ "grad_norm": 0.04260355979204178,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 4.263361930847168,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670157068062828,
+ "grad_norm": 0.043366871774196625,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 4.254581928253174,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68411867364747,
+ "grad_norm": 0.04299113526940346,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 4.268922328948975,
+ "step": 1273
+ },
+ {
+ "epoch": 17.698080279232112,
+ "grad_norm": 0.041865330189466476,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 4.290097236633301,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712041884816752,
+ "grad_norm": 0.04515853524208069,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 4.261376857757568,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726003490401396,
+ "grad_norm": 0.046030085533857346,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 4.290409564971924,
+ "step": 1276
+ },
+ {
+ "epoch": 17.739965095986037,
+ "grad_norm": 0.045034267008304596,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 4.28201961517334,
+ "step": 1277
+ },
+ {
+ "epoch": 17.75392670157068,
+ "grad_norm": 0.04293077066540718,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 4.282614707946777,
+ "step": 1278
+ },
+ {
+ "epoch": 17.767888307155324,
+ "grad_norm": 0.043866172432899475,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 4.2777557373046875,
+ "step": 1279
+ },
+ {
+ "epoch": 17.781849912739965,
+ "grad_norm": 0.04693951830267906,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 4.272597312927246,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79581151832461,
+ "grad_norm": 0.04072774201631546,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 4.284282684326172,
+ "step": 1281
+ },
+ {
+ "epoch": 17.80977312390925,
+ "grad_norm": 0.04241872951388359,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 4.254406452178955,
+ "step": 1282
+ },
+ {
+ "epoch": 17.823734729493893,
+ "grad_norm": 0.04614942893385887,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 4.297187805175781,
+ "step": 1283
+ },
+ {
+ "epoch": 17.837696335078533,
+ "grad_norm": 0.04326155036687851,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 4.292263984680176,
+ "step": 1284
+ },
+ {
+ "epoch": 17.851657940663177,
+ "grad_norm": 0.04053817689418793,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 4.2925920486450195,
+ "step": 1285
+ },
+ {
+ "epoch": 17.865619546247817,
+ "grad_norm": 0.040867388248443604,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 4.276965141296387,
+ "step": 1286
+ },
+ {
+ "epoch": 17.87958115183246,
+ "grad_norm": 0.041000183671712875,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 4.257638454437256,
+ "step": 1287
+ },
+ {
+ "epoch": 17.8935427574171,
+ "grad_norm": 0.04014866054058075,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 4.2734479904174805,
+ "step": 1288
+ },
+ {
+ "epoch": 17.907504363001745,
+ "grad_norm": 0.041237372905015945,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 4.263461112976074,
+ "step": 1289
+ },
+ {
+ "epoch": 17.921465968586386,
+ "grad_norm": 0.039109110832214355,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 4.295286655426025,
+ "step": 1290
+ },
+ {
+ "epoch": 17.93542757417103,
+ "grad_norm": 0.03911379352211952,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 4.254448413848877,
+ "step": 1291
+ },
+ {
+ "epoch": 17.949389179755673,
+ "grad_norm": 0.03997721150517464,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 4.284714698791504,
+ "step": 1292
+ },
+ {
+ "epoch": 17.963350785340314,
+ "grad_norm": 0.04011402651667595,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 4.278304100036621,
+ "step": 1293
+ },
+ {
+ "epoch": 17.977312390924958,
+ "grad_norm": 0.04040595516562462,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 4.288183212280273,
+ "step": 1294
+ },
+ {
+ "epoch": 17.991273996509598,
+ "grad_norm": 0.038857635110616684,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 4.274664402008057,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.028110472485423088,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 2.6625723838806152,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.5912551283836365,
+ "eval_runtime": 60.273,
+ "eval_samples_per_second": 40.516,
+ "eval_steps_per_second": 0.647,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013961605584644,
+ "grad_norm": 0.041634928435087204,
+ "learning_rate": 0.000444677103403854,
+ "loss": 4.181324005126953,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027923211169284,
+ "grad_norm": 0.051596369594335556,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 4.204722881317139,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041884816753928,
+ "grad_norm": 0.058580417186021805,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 4.207996845245361,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05584642233857,
+ "grad_norm": 0.05986522138118744,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 4.184011459350586,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069808027923212,
+ "grad_norm": 0.051045969128608704,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 4.148118019104004,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083769633507853,
+ "grad_norm": 0.045477766543626785,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 4.207469940185547,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097731239092496,
+ "grad_norm": 0.0488731674849987,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 4.201406478881836,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111692844677137,
+ "grad_norm": 0.05013517662882805,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 4.192239761352539,
+ "step": 1304
+ },
+ {
+ "epoch": 18.12565445026178,
+ "grad_norm": 0.05091651529073715,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 4.188252925872803,
+ "step": 1305
+ },
+ {
+ "epoch": 18.13961605584642,
+ "grad_norm": 0.051825884729623795,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 4.157929420471191,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153577661431065,
+ "grad_norm": 0.049625713378190994,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 4.208776473999023,
+ "step": 1307
+ },
+ {
+ "epoch": 18.167539267015705,
+ "grad_norm": 0.04816461354494095,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 4.180496692657471,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18150087260035,
+ "grad_norm": 0.04720531404018402,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 4.178536415100098,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195462478184993,
+ "grad_norm": 0.0475568026304245,
+ "learning_rate": 0.000441621628895411,
+ "loss": 4.203497886657715,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209424083769633,
+ "grad_norm": 0.04613947868347168,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 4.226914405822754,
+ "step": 1311
+ },
+ {
+ "epoch": 18.223385689354277,
+ "grad_norm": 0.04585929960012436,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 4.175119400024414,
+ "step": 1312
+ },
+ {
+ "epoch": 18.237347294938917,
+ "grad_norm": 0.045777060091495514,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 4.181760311126709,
+ "step": 1313
+ },
+ {
+ "epoch": 18.25130890052356,
+ "grad_norm": 0.04583602771162987,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 4.179868698120117,
+ "step": 1314
+ },
+ {
+ "epoch": 18.2652705061082,
+ "grad_norm": 0.04706503078341484,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 4.203325271606445,
+ "step": 1315
+ },
+ {
+ "epoch": 18.279232111692846,
+ "grad_norm": 0.045122336596250534,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 4.187152862548828,
+ "step": 1316
+ },
+ {
+ "epoch": 18.293193717277486,
+ "grad_norm": 0.04495825991034508,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 4.2080535888671875,
+ "step": 1317
+ },
+ {
+ "epoch": 18.30715532286213,
+ "grad_norm": 0.04767543077468872,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 4.207225322723389,
+ "step": 1318
+ },
+ {
+ "epoch": 18.32111692844677,
+ "grad_norm": 0.053396135568618774,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 4.227478981018066,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335078534031414,
+ "grad_norm": 0.0512089841067791,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 4.1904754638671875,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349040139616054,
+ "grad_norm": 0.04967977851629257,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 4.2079925537109375,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363001745200698,
+ "grad_norm": 0.04800970479846001,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 4.206037998199463,
+ "step": 1322
+ },
+ {
+ "epoch": 18.376963350785342,
+ "grad_norm": 0.04993556812405586,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 4.210054397583008,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390924956369982,
+ "grad_norm": 0.0480492040514946,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 4.229897499084473,
+ "step": 1324
+ },
+ {
+ "epoch": 18.404886561954626,
+ "grad_norm": 0.048699282109737396,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 4.17500114440918,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418848167539267,
+ "grad_norm": 0.04681490734219551,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 4.196526527404785,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43280977312391,
+ "grad_norm": 0.04674449935555458,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 4.246790409088135,
+ "step": 1327
+ },
+ {
+ "epoch": 18.44677137870855,
+ "grad_norm": 0.04650607705116272,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 4.225878715515137,
+ "step": 1328
+ },
+ {
+ "epoch": 18.460732984293195,
+ "grad_norm": 0.047200269997119904,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 4.20685338973999,
+ "step": 1329
+ },
+ {
+ "epoch": 18.474694589877835,
+ "grad_norm": 0.04322655498981476,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 4.212920665740967,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48865619546248,
+ "grad_norm": 0.04489128664135933,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 4.266653060913086,
+ "step": 1331
+ },
+ {
+ "epoch": 18.50261780104712,
+ "grad_norm": 0.048838935792446136,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 4.2364501953125,
+ "step": 1332
+ },
+ {
+ "epoch": 18.516579406631763,
+ "grad_norm": 0.04893286153674126,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 4.202394485473633,
+ "step": 1333
+ },
+ {
+ "epoch": 18.530541012216403,
+ "grad_norm": 0.046900101006031036,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 4.209832191467285,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544502617801047,
+ "grad_norm": 0.044056762009859085,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 4.229903221130371,
+ "step": 1335
+ },
+ {
+ "epoch": 18.55846422338569,
+ "grad_norm": 0.04681328684091568,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 4.22783899307251,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57242582897033,
+ "grad_norm": 0.046087440103292465,
+ "learning_rate": 0.000435215814386134,
+ "loss": 4.2078986167907715,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586387434554975,
+ "grad_norm": 0.04363090172410011,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 4.218686103820801,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600349040139616,
+ "grad_norm": 0.046255532652139664,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 4.2475128173828125,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61431064572426,
+ "grad_norm": 0.045545510947704315,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 4.233272552490234,
+ "step": 1340
+ },
+ {
+ "epoch": 18.6282722513089,
+ "grad_norm": 0.045259326696395874,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 4.252043724060059,
+ "step": 1341
+ },
+ {
+ "epoch": 18.642233856893544,
+ "grad_norm": 0.042144861072301865,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 4.247762203216553,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656195462478184,
+ "grad_norm": 0.04298558086156845,
+ "learning_rate": 0.000433781621332752,
+ "loss": 4.222329616546631,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670157068062828,
+ "grad_norm": 0.044070884585380554,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 4.222488880157471,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68411867364747,
+ "grad_norm": 0.044758375734090805,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 4.238160133361816,
+ "step": 1345
+ },
+ {
+ "epoch": 18.698080279232112,
+ "grad_norm": 0.04324857518076897,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 4.220003128051758,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712041884816752,
+ "grad_norm": 0.04349793121218681,
+ "learning_rate": 0.000432823382695327,
+ "loss": 4.251608848571777,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726003490401396,
+ "grad_norm": 0.04483242705464363,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 4.243319511413574,
+ "step": 1348
+ },
+ {
+ "epoch": 18.739965095986037,
+ "grad_norm": 0.04391994699835777,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 4.237479209899902,
+ "step": 1349
+ },
+ {
+ "epoch": 18.75392670157068,
+ "grad_norm": 0.042584922164678574,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 4.2309980392456055,
+ "step": 1350
+ },
+ {
+ "epoch": 18.767888307155324,
+ "grad_norm": 0.04363991320133209,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 4.216980934143066,
+ "step": 1351
+ },
+ {
+ "epoch": 18.781849912739965,
+ "grad_norm": 0.04662707448005676,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 4.257450103759766,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79581151832461,
+ "grad_norm": 0.0462074875831604,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 4.222329139709473,
+ "step": 1353
+ },
+ {
+ "epoch": 18.80977312390925,
+ "grad_norm": 0.043543554842472076,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 4.231960773468018,
+ "step": 1354
+ },
+ {
+ "epoch": 18.823734729493893,
+ "grad_norm": 0.04444567486643791,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 4.225714683532715,
+ "step": 1355
+ },
+ {
+ "epoch": 18.837696335078533,
+ "grad_norm": 0.04507977142930031,
+ "learning_rate": 0.000430661245733797,
+ "loss": 4.234722137451172,
+ "step": 1356
+ },
+ {
+ "epoch": 18.851657940663177,
+ "grad_norm": 0.043786536902189255,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 4.274998664855957,
+ "step": 1357
+ },
+ {
+ "epoch": 18.865619546247817,
+ "grad_norm": 0.04188325256109238,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 4.260570526123047,
+ "step": 1358
+ },
+ {
+ "epoch": 18.87958115183246,
+ "grad_norm": 0.0439927875995636,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 4.250492572784424,
+ "step": 1359
+ },
+ {
+ "epoch": 18.8935427574171,
+ "grad_norm": 0.04520147293806076,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 4.250911712646484,
+ "step": 1360
+ },
+ {
+ "epoch": 18.907504363001745,
+ "grad_norm": 0.04442249983549118,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 4.255832672119141,
+ "step": 1361
+ },
+ {
+ "epoch": 18.921465968586386,
+ "grad_norm": 0.044756144285202026,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 4.226641654968262,
+ "step": 1362
+ },
+ {
+ "epoch": 18.93542757417103,
+ "grad_norm": 0.04647127911448479,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 4.2110915184021,
+ "step": 1363
+ },
+ {
+ "epoch": 18.949389179755673,
+ "grad_norm": 0.045986004173755646,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 4.243263244628906,
+ "step": 1364
+ },
+ {
+ "epoch": 18.963350785340314,
+ "grad_norm": 0.04533308371901512,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 4.253750801086426,
+ "step": 1365
+ },
+ {
+ "epoch": 18.977312390924958,
+ "grad_norm": 0.04307471215724945,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 4.233024597167969,
+ "step": 1366
+ },
+ {
+ "epoch": 18.991273996509598,
+ "grad_norm": 0.0416078194975853,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 4.264965534210205,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.030926328152418137,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 2.651331901550293,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.592528223991394,
+ "eval_runtime": 59.67,
+ "eval_samples_per_second": 40.925,
+ "eval_steps_per_second": 0.654,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013961605584644,
+ "grad_norm": 0.04524356126785278,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 4.153796672821045,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027923211169284,
+ "grad_norm": 0.05105067789554596,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 4.154603004455566,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041884816753928,
+ "grad_norm": 0.052823506295681,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 4.1599273681640625,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05584642233857,
+ "grad_norm": 0.05688423290848732,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 4.124622344970703,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069808027923212,
+ "grad_norm": 0.053178343921899796,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 4.138644218444824,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083769633507853,
+ "grad_norm": 0.05473732203245163,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 4.138207912445068,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097731239092496,
+ "grad_norm": 0.055207233875989914,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 4.155735015869141,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111692844677137,
+ "grad_norm": 0.05484927073121071,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 4.140205383300781,
+ "step": 1376
+ },
+ {
+ "epoch": 19.12565445026178,
+ "grad_norm": 0.05589596927165985,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 4.160840034484863,
+ "step": 1377
+ },
+ {
+ "epoch": 19.13961605584642,
+ "grad_norm": 0.05222902446985245,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 4.154781341552734,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153577661431065,
+ "grad_norm": 0.051824700087308884,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 4.134243488311768,
+ "step": 1379
+ },
+ {
+ "epoch": 19.167539267015705,
+ "grad_norm": 0.0517590269446373,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 4.125945568084717,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18150087260035,
+ "grad_norm": 0.05399726703763008,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 4.146993160247803,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195462478184993,
+ "grad_norm": 0.054430052638053894,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 4.1310906410217285,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209424083769633,
+ "grad_norm": 0.0551968477666378,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 4.170640468597412,
+ "step": 1383
+ },
+ {
+ "epoch": 19.223385689354277,
+ "grad_norm": 0.06050962209701538,
+ "learning_rate": 0.000423881957237287,
+ "loss": 4.166876316070557,
+ "step": 1384
+ },
+ {
+ "epoch": 19.237347294938917,
+ "grad_norm": 0.055587392300367355,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 4.150339603424072,
+ "step": 1385
+ },
+ {
+ "epoch": 19.25130890052356,
+ "grad_norm": 0.05193328112363815,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 4.171754360198975,
+ "step": 1386
+ },
+ {
+ "epoch": 19.2652705061082,
+ "grad_norm": 0.05272402986884117,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 4.165149688720703,
+ "step": 1387
+ },
+ {
+ "epoch": 19.279232111692846,
+ "grad_norm": 0.056283943355083466,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 4.193273544311523,
+ "step": 1388
+ },
+ {
+ "epoch": 19.293193717277486,
+ "grad_norm": 0.05825977027416229,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 4.192652225494385,
+ "step": 1389
+ },
+ {
+ "epoch": 19.30715532286213,
+ "grad_norm": 0.05539379641413689,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 4.187044143676758,
+ "step": 1390
+ },
+ {
+ "epoch": 19.32111692844677,
+ "grad_norm": 0.05300498753786087,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 4.168220043182373,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335078534031414,
+ "grad_norm": 0.05425756797194481,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 4.185193061828613,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349040139616054,
+ "grad_norm": 0.05527182295918465,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 4.1618332862854,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363001745200698,
+ "grad_norm": 0.054422613233327866,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 4.195295810699463,
+ "step": 1394
+ },
+ {
+ "epoch": 19.376963350785342,
+ "grad_norm": 0.052717432379722595,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 4.143290996551514,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390924956369982,
+ "grad_norm": 0.05154350399971008,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 4.115989685058594,
+ "step": 1396
+ },
+ {
+ "epoch": 19.404886561954626,
+ "grad_norm": 0.0489451140165329,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 4.169506072998047,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418848167539267,
+ "grad_norm": 0.049093786627054214,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 4.164122104644775,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43280977312391,
+ "grad_norm": 0.04690805822610855,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 4.186679840087891,
+ "step": 1399
+ },
+ {
+ "epoch": 19.44677137870855,
+ "grad_norm": 0.046708665788173676,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 4.184325695037842,
+ "step": 1400
+ },
+ {
+ "epoch": 19.460732984293195,
+ "grad_norm": 0.04617927595973015,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 4.201678276062012,
+ "step": 1401
+ },
+ {
+ "epoch": 19.474694589877835,
+ "grad_norm": 0.05017530545592308,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 4.18928337097168,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48865619546248,
+ "grad_norm": 0.05197977274656296,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 4.188801288604736,
+ "step": 1403
+ },
+ {
+ "epoch": 19.50261780104712,
+ "grad_norm": 0.049691297113895416,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 4.209467887878418,
+ "step": 1404
+ },
+ {
+ "epoch": 19.516579406631763,
+ "grad_norm": 0.049709953367710114,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 4.19733190536499,
+ "step": 1405
+ },
+ {
+ "epoch": 19.530541012216403,
+ "grad_norm": 0.05274137482047081,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 4.221798896789551,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544502617801047,
+ "grad_norm": 0.05015392601490021,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 4.210565567016602,
+ "step": 1407
+ },
+ {
+ "epoch": 19.55846422338569,
+ "grad_norm": 0.05023058131337166,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 4.184835910797119,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57242582897033,
+ "grad_norm": 0.0479239858686924,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 4.199846267700195,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586387434554975,
+ "grad_norm": 0.0481228232383728,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 4.185157775878906,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600349040139616,
+ "grad_norm": 0.04909738525748253,
+ "learning_rate": 0.000417272427439646,
+ "loss": 4.207550525665283,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61431064572426,
+ "grad_norm": 0.0475567989051342,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 4.213225841522217,
+ "step": 1412
+ },
+ {
+ "epoch": 19.6282722513089,
+ "grad_norm": 0.048073310405015945,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 4.212669372558594,
+ "step": 1413
+ },
+ {
+ "epoch": 19.642233856893544,
+ "grad_norm": 0.04523796588182449,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 4.1950860023498535,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656195462478184,
+ "grad_norm": 0.04359745234251022,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 4.2136640548706055,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670157068062828,
+ "grad_norm": 0.047282423824071884,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 4.209680557250977,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68411867364747,
+ "grad_norm": 0.0474061481654644,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 4.188494682312012,
+ "step": 1417
+ },
+ {
+ "epoch": 19.698080279232112,
+ "grad_norm": 0.04402102902531624,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 4.207236289978027,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712041884816752,
+ "grad_norm": 0.04398655146360397,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 4.210899353027344,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726003490401396,
+ "grad_norm": 0.045767925679683685,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 4.171738624572754,
+ "step": 1420
+ },
+ {
+ "epoch": 19.739965095986037,
+ "grad_norm": 0.04498986527323723,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 4.222414970397949,
+ "step": 1421
+ },
+ {
+ "epoch": 19.75392670157068,
+ "grad_norm": 0.04282551258802414,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 4.202138900756836,
+ "step": 1422
+ },
+ {
+ "epoch": 19.767888307155324,
+ "grad_norm": 0.04392262175679207,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 4.184439659118652,
+ "step": 1423
+ },
+ {
+ "epoch": 19.781849912739965,
+ "grad_norm": 0.04343154653906822,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 4.204674243927002,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79581151832461,
+ "grad_norm": 0.04496419429779053,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 4.23316764831543,
+ "step": 1425
+ },
+ {
+ "epoch": 19.80977312390925,
+ "grad_norm": 0.04488370940089226,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 4.201659202575684,
+ "step": 1426
+ },
+ {
+ "epoch": 19.823734729493893,
+ "grad_norm": 0.045598775148391724,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 4.187412261962891,
+ "step": 1427
+ },
+ {
+ "epoch": 19.837696335078533,
+ "grad_norm": 0.043846048414707184,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 4.215444564819336,
+ "step": 1428
+ },
+ {
+ "epoch": 19.851657940663177,
+ "grad_norm": 0.043062254786491394,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 4.2074055671691895,
+ "step": 1429
+ },
+ {
+ "epoch": 19.865619546247817,
+ "grad_norm": 0.04544302448630333,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 4.184919357299805,
+ "step": 1430
+ },
+ {
+ "epoch": 19.87958115183246,
+ "grad_norm": 0.046144865453243256,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 4.2175750732421875,
+ "step": 1431
+ },
+ {
+ "epoch": 19.8935427574171,
+ "grad_norm": 0.04465445503592491,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 4.1851115226745605,
+ "step": 1432
+ },
+ {
+ "epoch": 19.907504363001745,
+ "grad_norm": 0.045913565903902054,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 4.2226152420043945,
+ "step": 1433
+ },
+ {
+ "epoch": 19.921465968586386,
+ "grad_norm": 0.04401300847530365,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 4.216772556304932,
+ "step": 1434
+ },
+ {
+ "epoch": 19.93542757417103,
+ "grad_norm": 0.042624711990356445,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 4.2133684158325195,
+ "step": 1435
+ },
+ {
+ "epoch": 19.949389179755673,
+ "grad_norm": 0.04533256217837334,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 4.218449592590332,
+ "step": 1436
+ },
+ {
+ "epoch": 19.963350785340314,
+ "grad_norm": 0.04764309898018837,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 4.191402912139893,
+ "step": 1437
+ },
+ {
+ "epoch": 19.977312390924958,
+ "grad_norm": 0.04689318686723709,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 4.2113037109375,
+ "step": 1438
+ },
+ {
+ "epoch": 19.991273996509598,
+ "grad_norm": 0.04479913413524628,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 4.194989204406738,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.03151386231184006,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 2.6285438537597656,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 0.5938133597373962,
+ "eval_runtime": 58.9656,
+ "eval_samples_per_second": 41.414,
+ "eval_steps_per_second": 0.661,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013961605584644,
+ "grad_norm": 0.04884875938296318,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 4.1257734298706055,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027923211169284,
+ "grad_norm": 0.05968257039785385,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 4.091641426086426,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041884816753928,
+ "grad_norm": 0.06437838077545166,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 4.122465133666992,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05584642233857,
+ "grad_norm": 0.057831212878227234,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 4.129654884338379,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069808027923212,
+ "grad_norm": 0.049407217651605606,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 4.116935729980469,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083769633507853,
+ "grad_norm": 0.051957421004772186,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 4.125039577484131,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097731239092496,
+ "grad_norm": 0.05425718426704407,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 4.113485336303711,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111692844677137,
+ "grad_norm": 0.055318959057331085,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 4.119422912597656,
+ "step": 1448
+ },
+ {
+ "epoch": 20.12565445026178,
+ "grad_norm": 0.0548059344291687,
+ "learning_rate": 0.000407857329622967,
+ "loss": 4.097405910491943,
+ "step": 1449
+ },
+ {
+ "epoch": 20.13961605584642,
+ "grad_norm": 0.051046356558799744,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 4.115051746368408,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153577661431065,
+ "grad_norm": 0.05079492926597595,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 4.126070499420166,
+ "step": 1451
+ },
+ {
+ "epoch": 20.167539267015705,
+ "grad_norm": 0.04777265712618828,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 4.118952751159668,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18150087260035,
+ "grad_norm": 0.04895703122019768,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 4.095477104187012,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195462478184993,
+ "grad_norm": 0.051196496933698654,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 4.116866111755371,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209424083769633,
+ "grad_norm": 0.05625293031334877,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 4.141980171203613,
+ "step": 1455
+ },
+ {
+ "epoch": 20.223385689354277,
+ "grad_norm": 0.05852660909295082,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 4.126821517944336,
+ "step": 1456
+ },
+ {
+ "epoch": 20.237347294938917,
+ "grad_norm": 0.05419151484966278,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 4.10935640335083,
+ "step": 1457
+ },
+ {
+ "epoch": 20.25130890052356,
+ "grad_norm": 0.05251913517713547,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 4.151289463043213,
+ "step": 1458
+ },
+ {
+ "epoch": 20.2652705061082,
+ "grad_norm": 0.05223092809319496,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 4.149935722351074,
+ "step": 1459
+ },
+ {
+ "epoch": 20.279232111692846,
+ "grad_norm": 0.05314765125513077,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 4.106821060180664,
+ "step": 1460
+ },
+ {
+ "epoch": 20.293193717277486,
+ "grad_norm": 0.05472502112388611,
+ "learning_rate": 0.000404858275823277,
+ "loss": 4.120498180389404,
+ "step": 1461
+ },
+ {
+ "epoch": 20.30715532286213,
+ "grad_norm": 0.05233139917254448,
+ "learning_rate": 0.000404607816444578,
+ "loss": 4.1237335205078125,
+ "step": 1462
+ },
+ {
+ "epoch": 20.32111692844677,
+ "grad_norm": 0.05210232734680176,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 4.1271209716796875,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335078534031414,
+ "grad_norm": 0.049779925495386124,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 4.121427536010742,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349040139616054,
+ "grad_norm": 0.05232279747724533,
+ "learning_rate": 0.000403855947934478,
+ "loss": 4.136429786682129,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363001745200698,
+ "grad_norm": 0.05235981196165085,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 4.1808600425720215,
+ "step": 1466
+ },
+ {
+ "epoch": 20.376963350785342,
+ "grad_norm": 0.04902610555291176,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 4.138566017150879,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390924956369982,
+ "grad_norm": 0.04923798143863678,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 4.135126113891602,
+ "step": 1468
+ },
+ {
+ "epoch": 20.404886561954626,
+ "grad_norm": 0.05059202015399933,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 4.1548285484313965,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418848167539267,
+ "grad_norm": 0.05058307573199272,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 4.149041652679443,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43280977312391,
+ "grad_norm": 0.050172705203294754,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 4.148054122924805,
+ "step": 1471
+ },
+ {
+ "epoch": 20.44677137870855,
+ "grad_norm": 0.05158299580216408,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 4.1425018310546875,
+ "step": 1472
+ },
+ {
+ "epoch": 20.460732984293195,
+ "grad_norm": 0.052176836878061295,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 4.129712104797363,
+ "step": 1473
+ },
+ {
+ "epoch": 20.474694589877835,
+ "grad_norm": 0.052095964550971985,
+ "learning_rate": 0.000401595976318565,
+ "loss": 4.134920597076416,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48865619546248,
+ "grad_norm": 0.04944197088479996,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 4.165334701538086,
+ "step": 1475
+ },
+ {
+ "epoch": 20.50261780104712,
+ "grad_norm": 0.049351636320352554,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 4.144362926483154,
+ "step": 1476
+ },
+ {
+ "epoch": 20.516579406631763,
+ "grad_norm": 0.053657080978155136,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 4.164234161376953,
+ "step": 1477
+ },
+ {
+ "epoch": 20.530541012216403,
+ "grad_norm": 0.052578140050172806,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 4.131514549255371,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544502617801047,
+ "grad_norm": 0.049683794379234314,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 4.129961967468262,
+ "step": 1479
+ },
+ {
+ "epoch": 20.55846422338569,
+ "grad_norm": 0.05115316063165665,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 4.164663314819336,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57242582897033,
+ "grad_norm": 0.051620446145534515,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 4.161067485809326,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586387434554975,
+ "grad_norm": 0.04873501881957054,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 4.161318302154541,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600349040139616,
+ "grad_norm": 0.04968130216002464,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 4.151819229125977,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61431064572426,
+ "grad_norm": 0.05476681888103485,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 4.14247989654541,
+ "step": 1484
+ },
+ {
+ "epoch": 20.6282722513089,
+ "grad_norm": 0.05098642408847809,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 4.178083419799805,
+ "step": 1485
+ },
+ {
+ "epoch": 20.642233856893544,
+ "grad_norm": 0.047264356166124344,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 4.15170955657959,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656195462478184,
+ "grad_norm": 0.050382886081933975,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 4.166364669799805,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670157068062828,
+ "grad_norm": 0.04956239461898804,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 4.164124488830566,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68411867364747,
+ "grad_norm": 0.049608584493398666,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 4.130239486694336,
+ "step": 1489
+ },
+ {
+ "epoch": 20.698080279232112,
+ "grad_norm": 0.0489494763314724,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 4.163503646850586,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712041884816752,
+ "grad_norm": 0.04847714304924011,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 4.1744585037231445,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726003490401396,
+ "grad_norm": 0.050256017595529556,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 4.1502227783203125,
+ "step": 1492
+ },
+ {
+ "epoch": 20.739965095986037,
+ "grad_norm": 0.048107463866472244,
+ "learning_rate": 0.000396803974909638,
+ "loss": 4.1790971755981445,
+ "step": 1493
+ },
+ {
+ "epoch": 20.75392670157068,
+ "grad_norm": 0.04828944429755211,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 4.128883361816406,
+ "step": 1494
+ },
+ {
+ "epoch": 20.767888307155324,
+ "grad_norm": 0.04720750078558922,
+ "learning_rate": 0.000396297942393265,
+ "loss": 4.1661577224731445,
+ "step": 1495
+ },
+ {
+ "epoch": 20.781849912739965,
+ "grad_norm": 0.047148339450359344,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 4.172522068023682,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79581151832461,
+ "grad_norm": 0.04612366110086441,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 4.149024963378906,
+ "step": 1497
+ },
+ {
+ "epoch": 20.80977312390925,
+ "grad_norm": 0.048896901309490204,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 4.179235458374023,
+ "step": 1498
+ },
+ {
+ "epoch": 20.823734729493893,
+ "grad_norm": 0.04779870808124542,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 4.1649932861328125,
+ "step": 1499
+ },
+ {
+ "epoch": 20.837696335078533,
+ "grad_norm": 0.04949707165360451,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 4.17645263671875,
+ "step": 1500
+ },
+ {
+ "epoch": 20.851657940663177,
+ "grad_norm": 0.04645023122429848,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 4.216496467590332,
+ "step": 1501
+ },
+ {
+ "epoch": 20.865619546247817,
+ "grad_norm": 0.046520136296749115,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 4.152667999267578,
+ "step": 1502
+ },
+ {
+ "epoch": 20.87958115183246,
+ "grad_norm": 0.04882653057575226,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 4.187082767486572,
+ "step": 1503
+ },
+ {
+ "epoch": 20.8935427574171,
+ "grad_norm": 0.04669157415628433,
+ "learning_rate": 0.000394017098798633,
+ "loss": 4.172348499298096,
+ "step": 1504
+ },
+ {
+ "epoch": 20.907504363001745,
+ "grad_norm": 0.04595336690545082,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 4.180559158325195,
+ "step": 1505
+ },
+ {
+ "epoch": 20.921465968586386,
+ "grad_norm": 0.048478737473487854,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 4.142880439758301,
+ "step": 1506
+ },
+ {
+ "epoch": 20.93542757417103,
+ "grad_norm": 0.04609905555844307,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 4.1459641456604,
+ "step": 1507
+ },
+ {
+ "epoch": 20.949389179755673,
+ "grad_norm": 0.04524408280849457,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 4.185099124908447,
+ "step": 1508
+ },
+ {
+ "epoch": 20.963350785340314,
+ "grad_norm": 0.0475911870598793,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 4.2085161209106445,
+ "step": 1509
+ },
+ {
+ "epoch": 20.977312390924958,
+ "grad_norm": 0.047796521335840225,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 4.165589332580566,
+ "step": 1510
+ },
+ {
+ "epoch": 20.991273996509598,
+ "grad_norm": 0.047358278185129166,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 4.1634368896484375,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.03328625112771988,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 2.6023788452148438,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.5952173471450806,
+ "eval_runtime": 58.6816,
+ "eval_samples_per_second": 41.614,
+ "eval_steps_per_second": 0.665,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013961605584644,
+ "grad_norm": 0.04970651492476463,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 4.077130317687988,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027923211169284,
+ "grad_norm": 0.060882844030857086,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 4.074376106262207,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041884816753928,
+ "grad_norm": 0.06609825044870377,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 4.107635021209717,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05584642233857,
+ "grad_norm": 0.056066352874040604,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 4.068084716796875,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069808027923212,
+ "grad_norm": 0.053013063967227936,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 4.069967269897461,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083769633507853,
+ "grad_norm": 0.059536613523960114,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 4.0639142990112305,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097731239092496,
+ "grad_norm": 0.06084490567445755,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 4.093890190124512,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111692844677137,
+ "grad_norm": 0.05299340933561325,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 4.100567817687988,
+ "step": 1520
+ },
+ {
+ "epoch": 21.12565445026178,
+ "grad_norm": 0.05136161297559738,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 4.08171272277832,
+ "step": 1521
+ },
+ {
+ "epoch": 21.13961605584642,
+ "grad_norm": 0.05201329290866852,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 4.0869975090026855,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153577661431065,
+ "grad_norm": 0.053958263248205185,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 4.0928850173950195,
+ "step": 1523
+ },
+ {
+ "epoch": 21.167539267015705,
+ "grad_norm": 0.053465038537979126,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 4.097828388214111,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18150087260035,
+ "grad_norm": 0.050980277359485626,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 4.101456165313721,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195462478184993,
+ "grad_norm": 0.05415358766913414,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 4.09096097946167,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209424083769633,
+ "grad_norm": 0.05567048117518425,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 4.061287879943848,
+ "step": 1527
+ },
+ {
+ "epoch": 21.223385689354277,
+ "grad_norm": 0.052690375596284866,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 4.080180644989014,
+ "step": 1528
+ },
+ {
+ "epoch": 21.237347294938917,
+ "grad_norm": 0.05448971316218376,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 4.079794883728027,
+ "step": 1529
+ },
+ {
+ "epoch": 21.25130890052356,
+ "grad_norm": 0.055472876876592636,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 4.072025299072266,
+ "step": 1530
+ },
+ {
+ "epoch": 21.2652705061082,
+ "grad_norm": 0.05418326333165169,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 4.103858470916748,
+ "step": 1531
+ },
+ {
+ "epoch": 21.279232111692846,
+ "grad_norm": 0.0524221733212471,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 4.0808916091918945,
+ "step": 1532
+ },
+ {
+ "epoch": 21.293193717277486,
+ "grad_norm": 0.05224563553929329,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 4.090270042419434,
+ "step": 1533
+ },
+ {
+ "epoch": 21.30715532286213,
+ "grad_norm": 0.052917953580617905,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 4.081887245178223,
+ "step": 1534
+ },
+ {
+ "epoch": 21.32111692844677,
+ "grad_norm": 0.05619358271360397,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 4.089571475982666,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335078534031414,
+ "grad_norm": 0.05455326288938522,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 4.096992015838623,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349040139616054,
+ "grad_norm": 0.049311455339193344,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 4.114621639251709,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363001745200698,
+ "grad_norm": 0.05175276845693588,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 4.107684135437012,
+ "step": 1538
+ },
+ {
+ "epoch": 21.376963350785342,
+ "grad_norm": 0.05236738547682762,
+ "learning_rate": 0.00038509205478744,
+ "loss": 4.075709342956543,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390924956369982,
+ "grad_norm": 0.04899081960320473,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 4.108396530151367,
+ "step": 1540
+ },
+ {
+ "epoch": 21.404886561954626,
+ "grad_norm": 0.05129663646221161,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 4.1135711669921875,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418848167539267,
+ "grad_norm": 0.05478592589497566,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 4.11839485168457,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43280977312391,
+ "grad_norm": 0.055830392986536026,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 4.11065149307251,
+ "step": 1543
+ },
+ {
+ "epoch": 21.44677137870855,
+ "grad_norm": 0.05370553582906723,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 4.113103866577148,
+ "step": 1544
+ },
+ {
+ "epoch": 21.460732984293195,
+ "grad_norm": 0.04968908056616783,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 4.119443893432617,
+ "step": 1545
+ },
+ {
+ "epoch": 21.474694589877835,
+ "grad_norm": 0.056149110198020935,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 4.126479625701904,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48865619546248,
+ "grad_norm": 0.053045690059661865,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 4.138535499572754,
+ "step": 1547
+ },
+ {
+ "epoch": 21.50261780104712,
+ "grad_norm": 0.04864223673939705,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 4.099347114562988,
+ "step": 1548
+ },
+ {
+ "epoch": 21.516579406631763,
+ "grad_norm": 0.05234768986701965,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 4.112222194671631,
+ "step": 1549
+ },
+ {
+ "epoch": 21.530541012216403,
+ "grad_norm": 0.050486158579587936,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 4.107913970947266,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544502617801047,
+ "grad_norm": 0.048641227185726166,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 4.114014148712158,
+ "step": 1551
+ },
+ {
+ "epoch": 21.55846422338569,
+ "grad_norm": 0.047759316861629486,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 4.128541946411133,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57242582897033,
+ "grad_norm": 0.046098388731479645,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 4.12142276763916,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586387434554975,
+ "grad_norm": 0.04934769123792648,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 4.13035249710083,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600349040139616,
+ "grad_norm": 0.048911672085523605,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 4.111344337463379,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61431064572426,
+ "grad_norm": 0.049368683248758316,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 4.1345672607421875,
+ "step": 1556
+ },
+ {
+ "epoch": 21.6282722513089,
+ "grad_norm": 0.04890736937522888,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 4.1286821365356445,
+ "step": 1557
+ },
+ {
+ "epoch": 21.642233856893544,
+ "grad_norm": 0.04812583327293396,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 4.09993314743042,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656195462478184,
+ "grad_norm": 0.05025152117013931,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 4.150769233703613,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670157068062828,
+ "grad_norm": 0.05007397383451462,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 4.074525833129883,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68411867364747,
+ "grad_norm": 0.050168465822935104,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 4.0933732986450195,
+ "step": 1561
+ },
+ {
+ "epoch": 21.698080279232112,
+ "grad_norm": 0.04990607872605324,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 4.126741886138916,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712041884816752,
+ "grad_norm": 0.05023440346121788,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 4.120314598083496,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726003490401396,
+ "grad_norm": 0.048100546002388,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 4.1242356300354,
+ "step": 1564
+ },
+ {
+ "epoch": 21.739965095986037,
+ "grad_norm": 0.047999415546655655,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 4.127745628356934,
+ "step": 1565
+ },
+ {
+ "epoch": 21.75392670157068,
+ "grad_norm": 0.05180094763636589,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 4.112565994262695,
+ "step": 1566
+ },
+ {
+ "epoch": 21.767888307155324,
+ "grad_norm": 0.053367964923381805,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 4.109419822692871,
+ "step": 1567
+ },
+ {
+ "epoch": 21.781849912739965,
+ "grad_norm": 0.05205400660634041,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 4.111886024475098,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79581151832461,
+ "grad_norm": 0.0464242659509182,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 4.104648590087891,
+ "step": 1569
+ },
+ {
+ "epoch": 21.80977312390925,
+ "grad_norm": 0.047767721116542816,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 4.156681060791016,
+ "step": 1570
+ },
+ {
+ "epoch": 21.823734729493893,
+ "grad_norm": 0.05028790235519409,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 4.132200241088867,
+ "step": 1571
+ },
+ {
+ "epoch": 21.837696335078533,
+ "grad_norm": 0.04714380204677582,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 4.132688999176025,
+ "step": 1572
+ },
+ {
+ "epoch": 21.851657940663177,
+ "grad_norm": 0.04726320132613182,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 4.173277378082275,
+ "step": 1573
+ },
+ {
+ "epoch": 21.865619546247817,
+ "grad_norm": 0.049558594822883606,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 4.15473747253418,
+ "step": 1574
+ },
+ {
+ "epoch": 21.87958115183246,
+ "grad_norm": 0.04742267727851868,
+ "learning_rate": 0.000375827577564042,
+ "loss": 4.10634708404541,
+ "step": 1575
+ },
+ {
+ "epoch": 21.8935427574171,
+ "grad_norm": 0.048223357647657394,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 4.133949279785156,
+ "step": 1576
+ },
+ {
+ "epoch": 21.907504363001745,
+ "grad_norm": 0.046766068786382675,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 4.153807640075684,
+ "step": 1577
+ },
+ {
+ "epoch": 21.921465968586386,
+ "grad_norm": 0.04586656019091606,
+ "learning_rate": 0.000375051971546236,
+ "loss": 4.169903755187988,
+ "step": 1578
+ },
+ {
+ "epoch": 21.93542757417103,
+ "grad_norm": 0.047912538051605225,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 4.148102283477783,
+ "step": 1579
+ },
+ {
+ "epoch": 21.949389179755673,
+ "grad_norm": 0.048949215561151505,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 4.149707317352295,
+ "step": 1580
+ },
+ {
+ "epoch": 21.963350785340314,
+ "grad_norm": 0.04765952378511429,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 4.128008842468262,
+ "step": 1581
+ },
+ {
+ "epoch": 21.977312390924958,
+ "grad_norm": 0.04726482927799225,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 4.1268720626831055,
+ "step": 1582
+ },
+ {
+ "epoch": 21.991273996509598,
+ "grad_norm": 0.0468299426138401,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 4.1332926750183105,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.03415689244866371,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 2.5875940322875977,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.5966277122497559,
+ "eval_runtime": 59.7704,
+ "eval_samples_per_second": 40.856,
+ "eval_steps_per_second": 0.652,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013961605584644,
+ "grad_norm": 0.04954535514116287,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 4.054527759552002,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027923211169284,
+ "grad_norm": 0.0498427115380764,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 4.048842430114746,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041884816753928,
+ "grad_norm": 0.051013074815273285,
+ "learning_rate": 0.000372722041257003,
+ "loss": 4.036187171936035,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05584642233857,
+ "grad_norm": 0.04973271116614342,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 4.050318717956543,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069808027923212,
+ "grad_norm": 0.051144856959581375,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 4.036319732666016,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083769633507853,
+ "grad_norm": 0.051874030381441116,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 4.017857551574707,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097731239092496,
+ "grad_norm": 0.05181736871600151,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 4.0519819259643555,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111692844677137,
+ "grad_norm": 0.05467049032449722,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 4.064291954040527,
+ "step": 1592
+ },
+ {
+ "epoch": 22.12565445026178,
+ "grad_norm": 0.05251661688089371,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 4.056951522827148,
+ "step": 1593
+ },
+ {
+ "epoch": 22.13961605584642,
+ "grad_norm": 0.05278680846095085,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 4.037125587463379,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153577661431065,
+ "grad_norm": 0.05276667699217796,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 4.053675174713135,
+ "step": 1595
+ },
+ {
+ "epoch": 22.167539267015705,
+ "grad_norm": 0.05007271468639374,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 4.044122695922852,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18150087260035,
+ "grad_norm": 0.051073770970106125,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 4.0265960693359375,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195462478184993,
+ "grad_norm": 0.05045647919178009,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 4.060938835144043,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209424083769633,
+ "grad_norm": 0.05075158178806305,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 4.035902976989746,
+ "step": 1599
+ },
+ {
+ "epoch": 22.223385689354277,
+ "grad_norm": 0.04867244139313698,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 4.055220603942871,
+ "step": 1600
+ },
+ {
+ "epoch": 22.237347294938917,
+ "grad_norm": 0.049976419657468796,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 4.026124954223633,
+ "step": 1601
+ },
+ {
+ "epoch": 22.25130890052356,
+ "grad_norm": 0.04968957230448723,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 4.081721782684326,
+ "step": 1602
+ },
+ {
+ "epoch": 22.2652705061082,
+ "grad_norm": 0.0503058098256588,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 4.0384721755981445,
+ "step": 1603
+ },
+ {
+ "epoch": 22.279232111692846,
+ "grad_norm": 0.050643209367990494,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 4.049397945404053,
+ "step": 1604
+ },
+ {
+ "epoch": 22.293193717277486,
+ "grad_norm": 0.05297046899795532,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 4.046389579772949,
+ "step": 1605
+ },
+ {
+ "epoch": 22.30715532286213,
+ "grad_norm": 0.053165238350629807,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 4.090221405029297,
+ "step": 1606
+ },
+ {
+ "epoch": 22.32111692844677,
+ "grad_norm": 0.05104832351207733,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 4.023227691650391,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335078534031414,
+ "grad_norm": 0.048736415803432465,
+ "learning_rate": 0.000367268105959126,
+ "loss": 4.07478666305542,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349040139616054,
+ "grad_norm": 0.05190873518586159,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 4.068147659301758,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363001745200698,
+ "grad_norm": 0.05409928783774376,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 4.060364723205566,
+ "step": 1610
+ },
+ {
+ "epoch": 22.376963350785342,
+ "grad_norm": 0.05112519487738609,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 4.072710990905762,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390924956369982,
+ "grad_norm": 0.04818312078714371,
+ "learning_rate": 0.000366226612810024,
+ "loss": 4.042757034301758,
+ "step": 1612
+ },
+ {
+ "epoch": 22.404886561954626,
+ "grad_norm": 0.050733599811792374,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 4.064810752868652,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418848167539267,
+ "grad_norm": 0.054070211946964264,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 4.054926872253418,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43280977312391,
+ "grad_norm": 0.05459332838654518,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 4.0652923583984375,
+ "step": 1615
+ },
+ {
+ "epoch": 22.44677137870855,
+ "grad_norm": 0.05064154416322708,
+ "learning_rate": 0.000365184304613104,
+ "loss": 4.076216220855713,
+ "step": 1616
+ },
+ {
+ "epoch": 22.460732984293195,
+ "grad_norm": 0.05071895569562912,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 4.070466041564941,
+ "step": 1617
+ },
+ {
+ "epoch": 22.474694589877835,
+ "grad_norm": 0.05049987882375717,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 4.062691688537598,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48865619546248,
+ "grad_norm": 0.05174451321363449,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 4.07501220703125,
+ "step": 1619
+ },
+ {
+ "epoch": 22.50261780104712,
+ "grad_norm": 0.05360838770866394,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 4.054711818695068,
+ "step": 1620
+ },
+ {
+ "epoch": 22.516579406631763,
+ "grad_norm": 0.051809027791023254,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 4.075811386108398,
+ "step": 1621
+ },
+ {
+ "epoch": 22.530541012216403,
+ "grad_norm": 0.05586795508861542,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 4.0638203620910645,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544502617801047,
+ "grad_norm": 0.05571979284286499,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 4.089414596557617,
+ "step": 1623
+ },
+ {
+ "epoch": 22.55846422338569,
+ "grad_norm": 0.05098723992705345,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 4.091159820556641,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57242582897033,
+ "grad_norm": 0.05359277129173279,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 4.060613632202148,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586387434554975,
+ "grad_norm": 0.05281100794672966,
+ "learning_rate": 0.000362575056487632,
+ "loss": 4.08416748046875,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600349040139616,
+ "grad_norm": 0.053913842886686325,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 4.088703155517578,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61431064572426,
+ "grad_norm": 0.05476798862218857,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 4.064074516296387,
+ "step": 1628
+ },
+ {
+ "epoch": 22.6282722513089,
+ "grad_norm": 0.051888495683670044,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 4.116634368896484,
+ "step": 1629
+ },
+ {
+ "epoch": 22.642233856893544,
+ "grad_norm": 0.051054876297712326,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 4.0800018310546875,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656195462478184,
+ "grad_norm": 0.054063159972429276,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 4.112773418426514,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670157068062828,
+ "grad_norm": 0.05173034593462944,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 4.062599182128906,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68411867364747,
+ "grad_norm": 0.054618336260318756,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 4.110188961029053,
+ "step": 1633
+ },
+ {
+ "epoch": 22.698080279232112,
+ "grad_norm": 0.055602896958589554,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 4.140998840332031,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712041884816752,
+ "grad_norm": 0.05332222580909729,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 4.119804382324219,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726003490401396,
+ "grad_norm": 0.05162122845649719,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 4.0731282234191895,
+ "step": 1636
+ },
+ {
+ "epoch": 22.739965095986037,
+ "grad_norm": 0.0514204278588295,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 4.068007946014404,
+ "step": 1637
+ },
+ {
+ "epoch": 22.75392670157068,
+ "grad_norm": 0.05027637258172035,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 4.097658157348633,
+ "step": 1638
+ },
+ {
+ "epoch": 22.767888307155324,
+ "grad_norm": 0.051812347024679184,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 4.083312511444092,
+ "step": 1639
+ },
+ {
+ "epoch": 22.781849912739965,
+ "grad_norm": 0.05145815759897232,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 4.129471778869629,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79581151832461,
+ "grad_norm": 0.05093684792518616,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 4.091841220855713,
+ "step": 1641
+ },
+ {
+ "epoch": 22.80977312390925,
+ "grad_norm": 0.05100737139582634,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 4.076615810394287,
+ "step": 1642
+ },
+ {
+ "epoch": 22.823734729493893,
+ "grad_norm": 0.05079149082303047,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 4.067737579345703,
+ "step": 1643
+ },
+ {
+ "epoch": 22.837696335078533,
+ "grad_norm": 0.05044509842991829,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 4.118824481964111,
+ "step": 1644
+ },
+ {
+ "epoch": 22.851657940663177,
+ "grad_norm": 0.05010584369301796,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 4.113405227661133,
+ "step": 1645
+ },
+ {
+ "epoch": 22.865619546247817,
+ "grad_norm": 0.052232302725315094,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 4.101553440093994,
+ "step": 1646
+ },
+ {
+ "epoch": 22.87958115183246,
+ "grad_norm": 0.05080661550164223,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 4.116809844970703,
+ "step": 1647
+ },
+ {
+ "epoch": 22.8935427574171,
+ "grad_norm": 0.05051033943891525,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 4.128484725952148,
+ "step": 1648
+ },
+ {
+ "epoch": 22.907504363001745,
+ "grad_norm": 0.051927294582128525,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 4.092285633087158,
+ "step": 1649
+ },
+ {
+ "epoch": 22.921465968586386,
+ "grad_norm": 0.049063827842473984,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 4.088473796844482,
+ "step": 1650
+ },
+ {
+ "epoch": 22.93542757417103,
+ "grad_norm": 0.05048854276537895,
+ "learning_rate": 0.000356031397755195,
+ "loss": 4.081521987915039,
+ "step": 1651
+ },
+ {
+ "epoch": 22.949389179755673,
+ "grad_norm": 0.049307651817798615,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 4.122158050537109,
+ "step": 1652
+ },
+ {
+ "epoch": 22.963350785340314,
+ "grad_norm": 0.048311345279216766,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 4.100488662719727,
+ "step": 1653
+ },
+ {
+ "epoch": 22.977312390924958,
+ "grad_norm": 0.05016734451055527,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 4.088587760925293,
+ "step": 1654
+ },
+ {
+ "epoch": 22.991273996509598,
+ "grad_norm": 0.04944450408220291,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 4.114686012268066,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.03592754527926445,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 2.545376777648926,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.5981184840202332,
+ "eval_runtime": 59.9558,
+ "eval_samples_per_second": 40.73,
+ "eval_steps_per_second": 0.65,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013961605584644,
+ "grad_norm": 0.05112785845994949,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 4.00495719909668,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027923211169284,
+ "grad_norm": 0.05807499587535858,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 4.015654563903809,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041884816753928,
+ "grad_norm": 0.05383148789405823,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.979349136352539,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05584642233857,
+ "grad_norm": 0.05221369490027428,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 4.009493827819824,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069808027923212,
+ "grad_norm": 0.05144324526190758,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 4.045680999755859,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083769633507853,
+ "grad_norm": 0.051422782242298126,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.994011878967285,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097731239092496,
+ "grad_norm": 0.05142863467335701,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 4.03209114074707,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111692844677137,
+ "grad_norm": 0.050954073667526245,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 4.00391960144043,
+ "step": 1664
+ },
+ {
+ "epoch": 23.12565445026178,
+ "grad_norm": 0.05365024879574776,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.9839367866516113,
+ "step": 1665
+ },
+ {
+ "epoch": 23.13961605584642,
+ "grad_norm": 0.051181286573410034,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 4.017707824707031,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153577661431065,
+ "grad_norm": 0.054279860109090805,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.989722490310669,
+ "step": 1667
+ },
+ {
+ "epoch": 23.167539267015705,
+ "grad_norm": 0.052196286618709564,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 4.014932155609131,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18150087260035,
+ "grad_norm": 0.05098043382167816,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 4.026510238647461,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195462478184993,
+ "grad_norm": 0.05241888761520386,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.979708433151245,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209424083769633,
+ "grad_norm": 0.051801104098558426,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 4.029015064239502,
+ "step": 1671
+ },
+ {
+ "epoch": 23.223385689354277,
+ "grad_norm": 0.052666619420051575,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.9932432174682617,
+ "step": 1672
+ },
+ {
+ "epoch": 23.237347294938917,
+ "grad_norm": 0.054418303072452545,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 4.05676794052124,
+ "step": 1673
+ },
+ {
+ "epoch": 23.25130890052356,
+ "grad_norm": 0.055074021220207214,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.999929666519165,
+ "step": 1674
+ },
+ {
+ "epoch": 23.2652705061082,
+ "grad_norm": 0.05626136437058449,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 4.013473987579346,
+ "step": 1675
+ },
+ {
+ "epoch": 23.279232111692846,
+ "grad_norm": 0.054028257727622986,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 4.015802383422852,
+ "step": 1676
+ },
+ {
+ "epoch": 23.293193717277486,
+ "grad_norm": 0.054044198244810104,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 4.020464897155762,
+ "step": 1677
+ },
+ {
+ "epoch": 23.30715532286213,
+ "grad_norm": 0.05465259402990341,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 4.010867118835449,
+ "step": 1678
+ },
+ {
+ "epoch": 23.32111692844677,
+ "grad_norm": 0.05545585975050926,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 4.054412841796875,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335078534031414,
+ "grad_norm": 0.055265430361032486,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 4.020756244659424,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349040139616054,
+ "grad_norm": 0.053363267332315445,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 4.033337116241455,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363001745200698,
+ "grad_norm": 0.05347830802202225,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 4.024794578552246,
+ "step": 1682
+ },
+ {
+ "epoch": 23.376963350785342,
+ "grad_norm": 0.05496951565146446,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 4.013302803039551,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390924956369982,
+ "grad_norm": 0.05179804563522339,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 4.030914306640625,
+ "step": 1684
+ },
+ {
+ "epoch": 23.404886561954626,
+ "grad_norm": 0.054958634078502655,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 4.012727737426758,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418848167539267,
+ "grad_norm": 0.05533262714743614,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 4.0543365478515625,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43280977312391,
+ "grad_norm": 0.052329424768686295,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 4.031486511230469,
+ "step": 1687
+ },
+ {
+ "epoch": 23.44677137870855,
+ "grad_norm": 0.05273434892296791,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 4.040292739868164,
+ "step": 1688
+ },
+ {
+ "epoch": 23.460732984293195,
+ "grad_norm": 0.05445481091737747,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 4.025630950927734,
+ "step": 1689
+ },
+ {
+ "epoch": 23.474694589877835,
+ "grad_norm": 0.052102912217378616,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 4.049562454223633,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48865619546248,
+ "grad_norm": 0.051272716373205185,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 4.0431365966796875,
+ "step": 1691
+ },
+ {
+ "epoch": 23.50261780104712,
+ "grad_norm": 0.051311176270246506,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 4.042119026184082,
+ "step": 1692
+ },
+ {
+ "epoch": 23.516579406631763,
+ "grad_norm": 0.05042742192745209,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 4.044881820678711,
+ "step": 1693
+ },
+ {
+ "epoch": 23.530541012216403,
+ "grad_norm": 0.05103316903114319,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 4.033669471740723,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544502617801047,
+ "grad_norm": 0.05441150814294815,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 4.076536178588867,
+ "step": 1695
+ },
+ {
+ "epoch": 23.55846422338569,
+ "grad_norm": 0.05529651418328285,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 4.073678016662598,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57242582897033,
+ "grad_norm": 0.0526033453643322,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 4.067898750305176,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586387434554975,
+ "grad_norm": 0.052817948162555695,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 4.017403602600098,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600349040139616,
+ "grad_norm": 0.052461300045251846,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 4.026311874389648,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61431064572426,
+ "grad_norm": 0.05097689852118492,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 4.076074600219727,
+ "step": 1700
+ },
+ {
+ "epoch": 23.6282722513089,
+ "grad_norm": 0.05387786403298378,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 4.022763729095459,
+ "step": 1701
+ },
+ {
+ "epoch": 23.642233856893544,
+ "grad_norm": 0.05132855102419853,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 4.077712535858154,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656195462478184,
+ "grad_norm": 0.05380886048078537,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 4.038937091827393,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670157068062828,
+ "grad_norm": 0.053299430757761,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 4.063364028930664,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68411867364747,
+ "grad_norm": 0.051913514733314514,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 4.05495548248291,
+ "step": 1705
+ },
+ {
+ "epoch": 23.698080279232112,
+ "grad_norm": 0.05660073831677437,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 4.057949066162109,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712041884816752,
+ "grad_norm": 0.054757390171289444,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 4.05377721786499,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726003490401396,
+ "grad_norm": 0.0517287515103817,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 4.035928726196289,
+ "step": 1708
+ },
+ {
+ "epoch": 23.739965095986037,
+ "grad_norm": 0.05187627300620079,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 4.043362617492676,
+ "step": 1709
+ },
+ {
+ "epoch": 23.75392670157068,
+ "grad_norm": 0.053502220660448074,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 4.007593154907227,
+ "step": 1710
+ },
+ {
+ "epoch": 23.767888307155324,
+ "grad_norm": 0.05755629390478134,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 4.076575756072998,
+ "step": 1711
+ },
+ {
+ "epoch": 23.781849912739965,
+ "grad_norm": 0.05619359388947487,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 4.072240352630615,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79581151832461,
+ "grad_norm": 0.05604318529367447,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 4.0486650466918945,
+ "step": 1713
+ },
+ {
+ "epoch": 23.80977312390925,
+ "grad_norm": 0.05630200728774071,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 4.086282730102539,
+ "step": 1714
+ },
+ {
+ "epoch": 23.823734729493893,
+ "grad_norm": 0.053262822329998016,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 4.089612007141113,
+ "step": 1715
+ },
+ {
+ "epoch": 23.837696335078533,
+ "grad_norm": 0.05699621140956879,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 4.053062438964844,
+ "step": 1716
+ },
+ {
+ "epoch": 23.851657940663177,
+ "grad_norm": 0.05963658541440964,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 4.072884559631348,
+ "step": 1717
+ },
+ {
+ "epoch": 23.865619546247817,
+ "grad_norm": 0.0547492615878582,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 4.066218852996826,
+ "step": 1718
+ },
+ {
+ "epoch": 23.87958115183246,
+ "grad_norm": 0.054908450692892075,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 4.035344123840332,
+ "step": 1719
+ },
+ {
+ "epoch": 23.8935427574171,
+ "grad_norm": 0.055191367864608765,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 4.08471155166626,
+ "step": 1720
+ },
+ {
+ "epoch": 23.907504363001745,
+ "grad_norm": 0.05334226414561272,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 4.076174736022949,
+ "step": 1721
+ },
+ {
+ "epoch": 23.921465968586386,
+ "grad_norm": 0.05579005926847458,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 4.05291223526001,
+ "step": 1722
+ },
+ {
+ "epoch": 23.93542757417103,
+ "grad_norm": 0.05464969575405121,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 4.07335090637207,
+ "step": 1723
+ },
+ {
+ "epoch": 23.949389179755673,
+ "grad_norm": 0.05240754038095474,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 4.0995073318481445,
+ "step": 1724
+ },
+ {
+ "epoch": 23.963350785340314,
+ "grad_norm": 0.05416056886315346,
+ "learning_rate": 0.000336518346307424,
+ "loss": 4.072793960571289,
+ "step": 1725
+ },
+ {
+ "epoch": 23.977312390924958,
+ "grad_norm": 0.05496834218502045,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 4.059562683105469,
+ "step": 1726
+ },
+ {
+ "epoch": 23.991273996509598,
+ "grad_norm": 0.05537887290120125,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 4.050772666931152,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.036677468568086624,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 2.5407257080078125,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.5995977520942688,
+ "eval_runtime": 60.2704,
+ "eval_samples_per_second": 40.517,
+ "eval_steps_per_second": 0.647,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013961605584644,
+ "grad_norm": 0.05583682656288147,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.9873995780944824,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027923211169284,
+ "grad_norm": 0.0605769157409668,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.999206066131592,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041884816753928,
+ "grad_norm": 0.057975225150585175,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.9923095703125,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05584642233857,
+ "grad_norm": 0.056104276329278946,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.946317672729492,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069808027923212,
+ "grad_norm": 0.05481491982936859,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.9973483085632324,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083769633507853,
+ "grad_norm": 0.05460227280855179,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.984617233276367,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097731239092496,
+ "grad_norm": 0.058481909334659576,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 4.0138773918151855,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111692844677137,
+ "grad_norm": 0.05672750994563103,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.9683098793029785,
+ "step": 1736
+ },
+ {
+ "epoch": 24.12565445026178,
+ "grad_norm": 0.05315157398581505,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.9909560680389404,
+ "step": 1737
+ },
+ {
+ "epoch": 24.13961605584642,
+ "grad_norm": 0.05565650761127472,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.954531192779541,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153577661431065,
+ "grad_norm": 0.05550679191946983,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.97672700881958,
+ "step": 1739
+ },
+ {
+ "epoch": 24.167539267015705,
+ "grad_norm": 0.05608576908707619,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.987941265106201,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18150087260035,
+ "grad_norm": 0.05266846343874931,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.97526216506958,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195462478184993,
+ "grad_norm": 0.05439139902591705,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 4.012054920196533,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209424083769633,
+ "grad_norm": 0.05473192036151886,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 4.022945404052734,
+ "step": 1743
+ },
+ {
+ "epoch": 24.223385689354277,
+ "grad_norm": 0.054737839847803116,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.981072425842285,
+ "step": 1744
+ },
+ {
+ "epoch": 24.237347294938917,
+ "grad_norm": 0.05696691572666168,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.9708170890808105,
+ "step": 1745
+ },
+ {
+ "epoch": 24.25130890052356,
+ "grad_norm": 0.05677401274442673,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.985995292663574,
+ "step": 1746
+ },
+ {
+ "epoch": 24.2652705061082,
+ "grad_norm": 0.05895088613033295,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.991887092590332,
+ "step": 1747
+ },
+ {
+ "epoch": 24.279232111692846,
+ "grad_norm": 0.05758961662650108,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.9695687294006348,
+ "step": 1748
+ },
+ {
+ "epoch": 24.293193717277486,
+ "grad_norm": 0.05174405500292778,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.959138870239258,
+ "step": 1749
+ },
+ {
+ "epoch": 24.30715532286213,
+ "grad_norm": 0.055284611880779266,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 4.012698173522949,
+ "step": 1750
+ },
+ {
+ "epoch": 24.32111692844677,
+ "grad_norm": 0.06042446941137314,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.9892425537109375,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335078534031414,
+ "grad_norm": 0.05489135906100273,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.9962222576141357,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349040139616054,
+ "grad_norm": 0.0542326495051384,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.9787094593048096,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363001745200698,
+ "grad_norm": 0.05612313374876976,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.9881339073181152,
+ "step": 1754
+ },
+ {
+ "epoch": 24.376963350785342,
+ "grad_norm": 0.05661709979176521,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 4.009440898895264,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390924956369982,
+ "grad_norm": 0.055537596344947815,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 4.00221061706543,
+ "step": 1756
+ },
+ {
+ "epoch": 24.404886561954626,
+ "grad_norm": 0.054210033267736435,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.9807987213134766,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418848167539267,
+ "grad_norm": 0.05847020074725151,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.960838794708252,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43280977312391,
+ "grad_norm": 0.05386386811733246,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 4.017860412597656,
+ "step": 1759
+ },
+ {
+ "epoch": 24.44677137870855,
+ "grad_norm": 0.056131865829229355,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.980543613433838,
+ "step": 1760
+ },
+ {
+ "epoch": 24.460732984293195,
+ "grad_norm": 0.05648239329457283,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.9948508739471436,
+ "step": 1761
+ },
+ {
+ "epoch": 24.474694589877835,
+ "grad_norm": 0.054138120263814926,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 4.002998352050781,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48865619546248,
+ "grad_norm": 0.05725586414337158,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 4.01864767074585,
+ "step": 1763
+ },
+ {
+ "epoch": 24.50261780104712,
+ "grad_norm": 0.05680665746331215,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 4.023143768310547,
+ "step": 1764
+ },
+ {
+ "epoch": 24.516579406631763,
+ "grad_norm": 0.052627697587013245,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 4.006389617919922,
+ "step": 1765
+ },
+ {
+ "epoch": 24.530541012216403,
+ "grad_norm": 0.053234782069921494,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 4.007419586181641,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544502617801047,
+ "grad_norm": 0.05756475403904915,
+ "learning_rate": 0.000325372061241796,
+ "loss": 4.0226945877075195,
+ "step": 1767
+ },
+ {
+ "epoch": 24.55846422338569,
+ "grad_norm": 0.055328067392110825,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 4.022767066955566,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57242582897033,
+ "grad_norm": 0.05378621071577072,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 4.005022048950195,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586387434554975,
+ "grad_norm": 0.056049644947052,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 4.025008201599121,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600349040139616,
+ "grad_norm": 0.057860117405653,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 4.003593444824219,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61431064572426,
+ "grad_norm": 0.056685276329517365,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.997868299484253,
+ "step": 1772
+ },
+ {
+ "epoch": 24.6282722513089,
+ "grad_norm": 0.053294967859983444,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 4.0211992263793945,
+ "step": 1773
+ },
+ {
+ "epoch": 24.642233856893544,
+ "grad_norm": 0.053967639803886414,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 4.022367477416992,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656195462478184,
+ "grad_norm": 0.052357546985149384,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 4.021792411804199,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670157068062828,
+ "grad_norm": 0.0547296479344368,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 4.008403778076172,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68411867364747,
+ "grad_norm": 0.056104280054569244,
+ "learning_rate": 0.000322712903929477,
+ "loss": 4.033827781677246,
+ "step": 1777
+ },
+ {
+ "epoch": 24.698080279232112,
+ "grad_norm": 0.055415406823158264,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 4.015371322631836,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712041884816752,
+ "grad_norm": 0.054506298154592514,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 4.0000152587890625,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726003490401396,
+ "grad_norm": 0.05468682199716568,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.9905481338500977,
+ "step": 1780
+ },
+ {
+ "epoch": 24.739965095986037,
+ "grad_norm": 0.05323269963264465,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 4.040729999542236,
+ "step": 1781
+ },
+ {
+ "epoch": 24.75392670157068,
+ "grad_norm": 0.052569612860679626,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 4.028299331665039,
+ "step": 1782
+ },
+ {
+ "epoch": 24.767888307155324,
+ "grad_norm": 0.052494753152132034,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 4.028860569000244,
+ "step": 1783
+ },
+ {
+ "epoch": 24.781849912739965,
+ "grad_norm": 0.05346379801630974,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 4.000349044799805,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79581151832461,
+ "grad_norm": 0.051543332636356354,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 4.016782760620117,
+ "step": 1785
+ },
+ {
+ "epoch": 24.80977312390925,
+ "grad_norm": 0.05330873280763626,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 4.024371147155762,
+ "step": 1786
+ },
+ {
+ "epoch": 24.823734729493893,
+ "grad_norm": 0.05306831747293472,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 4.0101213455200195,
+ "step": 1787
+ },
+ {
+ "epoch": 24.837696335078533,
+ "grad_norm": 0.05148417130112648,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 4.020768642425537,
+ "step": 1788
+ },
+ {
+ "epoch": 24.851657940663177,
+ "grad_norm": 0.050848908722400665,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 4.030246734619141,
+ "step": 1789
+ },
+ {
+ "epoch": 24.865619546247817,
+ "grad_norm": 0.05279014632105827,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 4.04910135269165,
+ "step": 1790
+ },
+ {
+ "epoch": 24.87958115183246,
+ "grad_norm": 0.0514594204723835,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 4.025290489196777,
+ "step": 1791
+ },
+ {
+ "epoch": 24.8935427574171,
+ "grad_norm": 0.05200032889842987,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 4.019560813903809,
+ "step": 1792
+ },
+ {
+ "epoch": 24.907504363001745,
+ "grad_norm": 0.052474118769168854,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 4.035177230834961,
+ "step": 1793
+ },
+ {
+ "epoch": 24.921465968586386,
+ "grad_norm": 0.05411219224333763,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.9920859336853027,
+ "step": 1794
+ },
+ {
+ "epoch": 24.93542757417103,
+ "grad_norm": 0.0543992817401886,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 4.017391204833984,
+ "step": 1795
+ },
+ {
+ "epoch": 24.949389179755673,
+ "grad_norm": 0.05187678337097168,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 4.042266368865967,
+ "step": 1796
+ },
+ {
+ "epoch": 24.963350785340314,
+ "grad_norm": 0.053762417286634445,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 4.029013633728027,
+ "step": 1797
+ },
+ {
+ "epoch": 24.977312390924958,
+ "grad_norm": 0.051723796874284744,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 4.038886070251465,
+ "step": 1798
+ },
+ {
+ "epoch": 24.991273996509598,
+ "grad_norm": 0.05242295190691948,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 4.056063652038574,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.03998562693595886,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 2.534770965576172,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 0.6009227633476257,
+ "eval_runtime": 58.9421,
+ "eval_samples_per_second": 41.43,
+ "eval_steps_per_second": 0.662,
+ "step": 1800
+ },
+ {
+ "epoch": 25.013961605584644,
+ "grad_norm": 0.05450186878442764,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 3.936816692352295,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027923211169284,
+ "grad_norm": 0.06389589607715607,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 3.9368391036987305,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041884816753928,
+ "grad_norm": 0.06028348580002785,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 3.953296661376953,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05584642233857,
+ "grad_norm": 0.05359262600541115,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 3.924558639526367,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069808027923212,
+ "grad_norm": 0.059837762266397476,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 3.9608402252197266,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083769633507853,
+ "grad_norm": 0.05523517727851868,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 3.943776845932007,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097731239092496,
+ "grad_norm": 0.05617433041334152,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 3.9635634422302246,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111692844677137,
+ "grad_norm": 0.06066269427537918,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 3.9647364616394043,
+ "step": 1808
+ },
+ {
+ "epoch": 25.12565445026178,
+ "grad_norm": 0.05491397902369499,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 3.95222806930542,
+ "step": 1809
+ },
+ {
+ "epoch": 25.13961605584642,
+ "grad_norm": 0.05731203034520149,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 3.9440197944641113,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153577661431065,
+ "grad_norm": 0.05428403243422508,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 3.9428441524505615,
+ "step": 1811
+ },
+ {
+ "epoch": 25.167539267015705,
+ "grad_norm": 0.05487453565001488,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 3.9489784240722656,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18150087260035,
+ "grad_norm": 0.05725884810090065,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 3.9643478393554688,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195462478184993,
+ "grad_norm": 0.05752462521195412,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 3.963301181793213,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209424083769633,
+ "grad_norm": 0.057387594133615494,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 3.9427294731140137,
+ "step": 1815
+ },
+ {
+ "epoch": 25.223385689354277,
+ "grad_norm": 0.057492151856422424,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 3.9229259490966797,
+ "step": 1816
+ },
+ {
+ "epoch": 25.237347294938917,
+ "grad_norm": 0.05202317237854004,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 3.938037395477295,
+ "step": 1817
+ },
+ {
+ "epoch": 25.25130890052356,
+ "grad_norm": 0.05574658885598183,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 3.9532949924468994,
+ "step": 1818
+ },
+ {
+ "epoch": 25.2652705061082,
+ "grad_norm": 0.05714752897620201,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 3.962207794189453,
+ "step": 1819
+ },
+ {
+ "epoch": 25.279232111692846,
+ "grad_norm": 0.05540093779563904,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 3.96329402923584,
+ "step": 1820
+ },
+ {
+ "epoch": 25.293193717277486,
+ "grad_norm": 0.056597232818603516,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 3.948793411254883,
+ "step": 1821
+ },
+ {
+ "epoch": 25.30715532286213,
+ "grad_norm": 0.055902402848005295,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 3.9111557006835938,
+ "step": 1822
+ },
+ {
+ "epoch": 25.32111692844677,
+ "grad_norm": 0.05509798601269722,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 3.929591655731201,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335078534031414,
+ "grad_norm": 0.05763949453830719,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 3.9758448600769043,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349040139616054,
+ "grad_norm": 0.053838349878787994,
+ "learning_rate": 0.000309930621834911,
+ "loss": 3.92457914352417,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363001745200698,
+ "grad_norm": 0.05467558279633522,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 3.9501285552978516,
+ "step": 1826
+ },
+ {
+ "epoch": 25.376963350785342,
+ "grad_norm": 0.05715392529964447,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 3.9734368324279785,
+ "step": 1827
+ },
+ {
+ "epoch": 25.390924956369982,
+ "grad_norm": 0.0566536970436573,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 3.9933152198791504,
+ "step": 1828
+ },
+ {
+ "epoch": 25.404886561954626,
+ "grad_norm": 0.053221236914396286,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 3.967787265777588,
+ "step": 1829
+ },
+ {
+ "epoch": 25.418848167539267,
+ "grad_norm": 0.05779031664133072,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 3.9415433406829834,
+ "step": 1830
+ },
+ {
+ "epoch": 25.43280977312391,
+ "grad_norm": 0.062013521790504456,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 3.9801456928253174,
+ "step": 1831
+ },
+ {
+ "epoch": 25.44677137870855,
+ "grad_norm": 0.05819004401564598,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 3.966928005218506,
+ "step": 1832
+ },
+ {
+ "epoch": 25.460732984293195,
+ "grad_norm": 0.05754564702510834,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 3.9612770080566406,
+ "step": 1833
+ },
+ {
+ "epoch": 25.474694589877835,
+ "grad_norm": 0.05707791447639465,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 3.935521125793457,
+ "step": 1834
+ },
+ {
+ "epoch": 25.48865619546248,
+ "grad_norm": 0.056380245834589005,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 4.006120681762695,
+ "step": 1835
+ },
+ {
+ "epoch": 25.50261780104712,
+ "grad_norm": 0.0551077239215374,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 3.9544754028320312,
+ "step": 1836
+ },
+ {
+ "epoch": 25.516579406631763,
+ "grad_norm": 0.05720360204577446,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 3.9564664363861084,
+ "step": 1837
+ },
+ {
+ "epoch": 25.530541012216403,
+ "grad_norm": 0.05522238463163376,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 4.0157880783081055,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544502617801047,
+ "grad_norm": 0.05304659530520439,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 3.957207202911377,
+ "step": 1839
+ },
+ {
+ "epoch": 25.55846422338569,
+ "grad_norm": 0.05569292604923248,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 3.9909608364105225,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57242582897033,
+ "grad_norm": 0.05347870662808418,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 3.9537010192871094,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586387434554975,
+ "grad_norm": 0.05572560802102089,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 3.9818642139434814,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600349040139616,
+ "grad_norm": 0.05624900385737419,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 3.995117425918579,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61431064572426,
+ "grad_norm": 0.05600332096219063,
+ "learning_rate": 0.000304866093757771,
+ "loss": 3.9480154514312744,
+ "step": 1844
+ },
+ {
+ "epoch": 25.6282722513089,
+ "grad_norm": 0.05488704890012741,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 3.993014335632324,
+ "step": 1845
+ },
+ {
+ "epoch": 25.642233856893544,
+ "grad_norm": 0.05647030100226402,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 3.994615077972412,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656195462478184,
+ "grad_norm": 0.05460594967007637,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 3.9714808464050293,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670157068062828,
+ "grad_norm": 0.05629213526844978,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 3.962033271789551,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68411867364747,
+ "grad_norm": 0.05666861683130264,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 3.995985269546509,
+ "step": 1849
+ },
+ {
+ "epoch": 25.698080279232112,
+ "grad_norm": 0.05580414459109306,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 4.001766681671143,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712041884816752,
+ "grad_norm": 0.05638023093342781,
+ "learning_rate": 0.000303,
+ "loss": 4.005331993103027,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726003490401396,
+ "grad_norm": 0.05684564262628555,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 4.017581939697266,
+ "step": 1852
+ },
+ {
+ "epoch": 25.739965095986037,
+ "grad_norm": 0.057602159678936005,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 3.9668312072753906,
+ "step": 1853
+ },
+ {
+ "epoch": 25.75392670157068,
+ "grad_norm": 0.055540211498737335,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 3.9885573387145996,
+ "step": 1854
+ },
+ {
+ "epoch": 25.767888307155324,
+ "grad_norm": 0.05655810981988907,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 3.990997314453125,
+ "step": 1855
+ },
+ {
+ "epoch": 25.781849912739965,
+ "grad_norm": 0.05443689972162247,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 3.9554123878479004,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79581151832461,
+ "grad_norm": 0.05570529028773308,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 3.983283519744873,
+ "step": 1857
+ },
+ {
+ "epoch": 25.80977312390925,
+ "grad_norm": 0.05399419739842415,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 4.004507064819336,
+ "step": 1858
+ },
+ {
+ "epoch": 25.823734729493893,
+ "grad_norm": 0.05587748438119888,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 4.00900936126709,
+ "step": 1859
+ },
+ {
+ "epoch": 25.837696335078533,
+ "grad_norm": 0.05465575307607651,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 3.9826579093933105,
+ "step": 1860
+ },
+ {
+ "epoch": 25.851657940663177,
+ "grad_norm": 0.05286499112844467,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 3.973453998565674,
+ "step": 1861
+ },
+ {
+ "epoch": 25.865619546247817,
+ "grad_norm": 0.05495590344071388,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 4.003095626831055,
+ "step": 1862
+ },
+ {
+ "epoch": 25.87958115183246,
+ "grad_norm": 0.053388018161058426,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 4.00042200088501,
+ "step": 1863
+ },
+ {
+ "epoch": 25.8935427574171,
+ "grad_norm": 0.05323237553238869,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 3.992929458618164,
+ "step": 1864
+ },
+ {
+ "epoch": 25.907504363001745,
+ "grad_norm": 0.053946711122989655,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 3.996519088745117,
+ "step": 1865
+ },
+ {
+ "epoch": 25.921465968586386,
+ "grad_norm": 0.05455862358212471,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 4.007123947143555,
+ "step": 1866
+ },
+ {
+ "epoch": 25.93542757417103,
+ "grad_norm": 0.05290355160832405,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 4.001724720001221,
+ "step": 1867
+ },
+ {
+ "epoch": 25.949389179755673,
+ "grad_norm": 0.052808016538619995,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 3.9964821338653564,
+ "step": 1868
+ },
+ {
+ "epoch": 25.963350785340314,
+ "grad_norm": 0.05416771396994591,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 3.98549222946167,
+ "step": 1869
+ },
+ {
+ "epoch": 25.977312390924958,
+ "grad_norm": 0.05264521390199661,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 3.9729766845703125,
+ "step": 1870
+ },
+ {
+ "epoch": 25.991273996509598,
+ "grad_norm": 0.05483962222933769,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 4.018877029418945,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.03927824646234512,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 2.456951856613159,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 0.6026532053947449,
+ "eval_runtime": 60.4079,
+ "eval_samples_per_second": 40.425,
+ "eval_steps_per_second": 0.646,
+ "step": 1872
+ },
+ {
+ "epoch": 26.013961605584644,
+ "grad_norm": 0.05756894499063492,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 3.93898606300354,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027923211169284,
+ "grad_norm": 0.06288876384496689,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 3.9142136573791504,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041884816753928,
+ "grad_norm": 0.0600322000682354,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 3.918700695037842,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05584642233857,
+ "grad_norm": 0.05729348957538605,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 3.9040162563323975,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069808027923212,
+ "grad_norm": 0.05658194422721863,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 3.8809690475463867,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083769633507853,
+ "grad_norm": 0.06274262815713882,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 3.928222179412842,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097731239092496,
+ "grad_norm": 0.06313030421733856,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 3.9395527839660645,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111692844677137,
+ "grad_norm": 0.0554068423807621,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 3.9314563274383545,
+ "step": 1880
+ },
+ {
+ "epoch": 26.12565445026178,
+ "grad_norm": 0.05697336047887802,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 3.9071288108825684,
+ "step": 1881
+ },
+ {
+ "epoch": 26.13961605584642,
+ "grad_norm": 0.05603424087166786,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 3.932054042816162,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153577661431065,
+ "grad_norm": 0.055603109300136566,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 3.9169063568115234,
+ "step": 1883
+ },
+ {
+ "epoch": 26.167539267015705,
+ "grad_norm": 0.055022306740283966,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 3.8985838890075684,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18150087260035,
+ "grad_norm": 0.05383468046784401,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 3.896359443664551,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195462478184993,
+ "grad_norm": 0.055076900869607925,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 3.905672073364258,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209424083769633,
+ "grad_norm": 0.05629882961511612,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 3.9222521781921387,
+ "step": 1887
+ },
+ {
+ "epoch": 26.223385689354277,
+ "grad_norm": 0.056736111640930176,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 3.8918867111206055,
+ "step": 1888
+ },
+ {
+ "epoch": 26.237347294938917,
+ "grad_norm": 0.05625338479876518,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 3.9249935150146484,
+ "step": 1889
+ },
+ {
+ "epoch": 26.25130890052356,
+ "grad_norm": 0.06145375967025757,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 3.904536724090576,
+ "step": 1890
+ },
+ {
+ "epoch": 26.2652705061082,
+ "grad_norm": 0.05601578950881958,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 3.8845863342285156,
+ "step": 1891
+ },
+ {
+ "epoch": 26.279232111692846,
+ "grad_norm": 0.056899815797805786,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 3.917966842651367,
+ "step": 1892
+ },
+ {
+ "epoch": 26.293193717277486,
+ "grad_norm": 0.060314927250146866,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 3.9230940341949463,
+ "step": 1893
+ },
+ {
+ "epoch": 26.30715532286213,
+ "grad_norm": 0.05449546128511429,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 3.9082441329956055,
+ "step": 1894
+ },
+ {
+ "epoch": 26.32111692844677,
+ "grad_norm": 0.05601051449775696,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 3.9174978733062744,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335078534031414,
+ "grad_norm": 0.05770742893218994,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 3.9183058738708496,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349040139616054,
+ "grad_norm": 0.05718272924423218,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 3.9438772201538086,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363001745200698,
+ "grad_norm": 0.052873238921165466,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 3.9554827213287354,
+ "step": 1898
+ },
+ {
+ "epoch": 26.376963350785342,
+ "grad_norm": 0.056853555142879486,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 3.886812210083008,
+ "step": 1899
+ },
+ {
+ "epoch": 26.390924956369982,
+ "grad_norm": 0.05662025138735771,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 3.9384284019470215,
+ "step": 1900
+ },
+ {
+ "epoch": 26.404886561954626,
+ "grad_norm": 0.05693329498171806,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 3.9123764038085938,
+ "step": 1901
+ },
+ {
+ "epoch": 26.418848167539267,
+ "grad_norm": 0.05425408110022545,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 3.9441678524017334,
+ "step": 1902
+ },
+ {
+ "epoch": 26.43280977312391,
+ "grad_norm": 0.05533420667052269,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 3.9086289405822754,
+ "step": 1903
+ },
+ {
+ "epoch": 26.44677137870855,
+ "grad_norm": 0.05527901649475098,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 3.9469594955444336,
+ "step": 1904
+ },
+ {
+ "epoch": 26.460732984293195,
+ "grad_norm": 0.052982039749622345,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 3.9182326793670654,
+ "step": 1905
+ },
+ {
+ "epoch": 26.474694589877835,
+ "grad_norm": 0.05520346015691757,
+ "learning_rate": 0.000288343693342466,
+ "loss": 3.954651117324829,
+ "step": 1906
+ },
+ {
+ "epoch": 26.48865619546248,
+ "grad_norm": 0.05564865842461586,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 3.951840400695801,
+ "step": 1907
+ },
+ {
+ "epoch": 26.50261780104712,
+ "grad_norm": 0.05588578060269356,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 3.933762550354004,
+ "step": 1908
+ },
+ {
+ "epoch": 26.516579406631763,
+ "grad_norm": 0.05727043375372887,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 3.9299421310424805,
+ "step": 1909
+ },
+ {
+ "epoch": 26.530541012216403,
+ "grad_norm": 0.05533654987812042,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 3.941525936126709,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544502617801047,
+ "grad_norm": 0.05524713918566704,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 3.935666084289551,
+ "step": 1911
+ },
+ {
+ "epoch": 26.55846422338569,
+ "grad_norm": 0.055435653775930405,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 3.951631546020508,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57242582897033,
+ "grad_norm": 0.05678093433380127,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 3.960881233215332,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586387434554975,
+ "grad_norm": 0.054896436631679535,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 3.9262163639068604,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600349040139616,
+ "grad_norm": 0.055213335901498795,
+ "learning_rate": 0.000285947841605349,
+ "loss": 3.9284324645996094,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61431064572426,
+ "grad_norm": 0.05644496902823448,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 3.9727559089660645,
+ "step": 1916
+ },
+ {
+ "epoch": 26.6282722513089,
+ "grad_norm": 0.05483568087220192,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 3.9410910606384277,
+ "step": 1917
+ },
+ {
+ "epoch": 26.642233856893544,
+ "grad_norm": 0.054636143147945404,
+ "learning_rate": 0.000285149463934261,
+ "loss": 3.945554733276367,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656195462478184,
+ "grad_norm": 0.056062210351228714,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 3.950249671936035,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670157068062828,
+ "grad_norm": 0.05734739080071449,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 3.9527649879455566,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68411867364747,
+ "grad_norm": 0.05562867224216461,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 3.9563350677490234,
+ "step": 1921
+ },
+ {
+ "epoch": 26.698080279232112,
+ "grad_norm": 0.05565498396754265,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 3.9478225708007812,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712041884816752,
+ "grad_norm": 0.057103481143713,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 3.9302189350128174,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726003490401396,
+ "grad_norm": 0.058241210877895355,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 3.9773902893066406,
+ "step": 1924
+ },
+ {
+ "epoch": 26.739965095986037,
+ "grad_norm": 0.055981557816267014,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 3.923896074295044,
+ "step": 1925
+ },
+ {
+ "epoch": 26.75392670157068,
+ "grad_norm": 0.05711016431450844,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 3.972334861755371,
+ "step": 1926
+ },
+ {
+ "epoch": 26.767888307155324,
+ "grad_norm": 0.05707632377743721,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 3.958393096923828,
+ "step": 1927
+ },
+ {
+ "epoch": 26.781849912739965,
+ "grad_norm": 0.055073633790016174,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 3.9449703693389893,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79581151832461,
+ "grad_norm": 0.05765518546104431,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 3.9442358016967773,
+ "step": 1929
+ },
+ {
+ "epoch": 26.80977312390925,
+ "grad_norm": 0.05692309886217117,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 3.9665679931640625,
+ "step": 1930
+ },
+ {
+ "epoch": 26.823734729493893,
+ "grad_norm": 0.05776236951351166,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 3.9153542518615723,
+ "step": 1931
+ },
+ {
+ "epoch": 26.837696335078533,
+ "grad_norm": 0.05465872213244438,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 3.919874429702759,
+ "step": 1932
+ },
+ {
+ "epoch": 26.851657940663177,
+ "grad_norm": 0.057827629148960114,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 3.9578394889831543,
+ "step": 1933
+ },
+ {
+ "epoch": 26.865619546247817,
+ "grad_norm": 0.055902253836393356,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 3.995711326599121,
+ "step": 1934
+ },
+ {
+ "epoch": 26.87958115183246,
+ "grad_norm": 0.05868638679385185,
+ "learning_rate": 0.000280627938758204,
+ "loss": 3.96519136428833,
+ "step": 1935
+ },
+ {
+ "epoch": 26.8935427574171,
+ "grad_norm": 0.059722673147916794,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 3.9566941261291504,
+ "step": 1936
+ },
+ {
+ "epoch": 26.907504363001745,
+ "grad_norm": 0.055209770798683167,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 3.9929540157318115,
+ "step": 1937
+ },
+ {
+ "epoch": 26.921465968586386,
+ "grad_norm": 0.05694293603301048,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 3.9529929161071777,
+ "step": 1938
+ },
+ {
+ "epoch": 26.93542757417103,
+ "grad_norm": 0.058492012321949005,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 3.9465792179107666,
+ "step": 1939
+ },
+ {
+ "epoch": 26.949389179755673,
+ "grad_norm": 0.056036073714494705,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 3.9590697288513184,
+ "step": 1940
+ },
+ {
+ "epoch": 26.963350785340314,
+ "grad_norm": 0.0547766387462616,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 3.969907522201538,
+ "step": 1941
+ },
+ {
+ "epoch": 26.977312390924958,
+ "grad_norm": 0.05592350661754608,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 3.9509897232055664,
+ "step": 1942
+ },
+ {
+ "epoch": 26.991273996509598,
+ "grad_norm": 0.05589321628212929,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 3.9648609161376953,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.03934016078710556,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 2.4784152507781982,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 0.6035652160644531,
+ "eval_runtime": 60.0656,
+ "eval_samples_per_second": 40.656,
+ "eval_steps_per_second": 0.649,
+ "step": 1944
+ },
+ {
+ "epoch": 27.013961605584644,
+ "grad_norm": 0.05773136392235756,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 3.8625826835632324,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027923211169284,
+ "grad_norm": 0.05844702199101448,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 3.8796658515930176,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041884816753928,
+ "grad_norm": 0.05718810483813286,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 3.8808112144470215,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05584642233857,
+ "grad_norm": 0.05264506861567497,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 3.8776121139526367,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069808027923212,
+ "grad_norm": 0.055468373000621796,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 3.866316318511963,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083769633507853,
+ "grad_norm": 0.053565360605716705,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 3.8870849609375,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097731239092496,
+ "grad_norm": 0.05468900501728058,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 3.8581550121307373,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111692844677137,
+ "grad_norm": 0.05454741418361664,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 3.8752541542053223,
+ "step": 1952
+ },
+ {
+ "epoch": 27.12565445026178,
+ "grad_norm": 0.05647914111614227,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 3.8818912506103516,
+ "step": 1953
+ },
+ {
+ "epoch": 27.13961605584642,
+ "grad_norm": 0.05325528234243393,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 3.8562560081481934,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153577661431065,
+ "grad_norm": 0.05581580847501755,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 3.879669189453125,
+ "step": 1955
+ },
+ {
+ "epoch": 27.167539267015705,
+ "grad_norm": 0.058219097554683685,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 3.8530635833740234,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18150087260035,
+ "grad_norm": 0.05613216012716293,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 3.8905630111694336,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195462478184993,
+ "grad_norm": 0.05486270785331726,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 3.8945157527923584,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209424083769633,
+ "grad_norm": 0.05338508263230324,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 3.909775733947754,
+ "step": 1959
+ },
+ {
+ "epoch": 27.223385689354277,
+ "grad_norm": 0.05602312833070755,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 3.8643651008605957,
+ "step": 1960
+ },
+ {
+ "epoch": 27.237347294938917,
+ "grad_norm": 0.05373037979006767,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 3.8915491104125977,
+ "step": 1961
+ },
+ {
+ "epoch": 27.25130890052356,
+ "grad_norm": 0.05805065110325813,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 3.8836984634399414,
+ "step": 1962
+ },
+ {
+ "epoch": 27.2652705061082,
+ "grad_norm": 0.05469847097992897,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 3.8812215328216553,
+ "step": 1963
+ },
+ {
+ "epoch": 27.279232111692846,
+ "grad_norm": 0.05575987696647644,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 3.8887271881103516,
+ "step": 1964
+ },
+ {
+ "epoch": 27.293193717277486,
+ "grad_norm": 0.05577889457345009,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 3.8917694091796875,
+ "step": 1965
+ },
+ {
+ "epoch": 27.30715532286213,
+ "grad_norm": 0.05604076758027077,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 3.881073474884033,
+ "step": 1966
+ },
+ {
+ "epoch": 27.32111692844677,
+ "grad_norm": 0.055674705654382706,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 3.9303488731384277,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335078534031414,
+ "grad_norm": 0.05300940200686455,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 3.871678113937378,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349040139616054,
+ "grad_norm": 0.05647176504135132,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 3.8848423957824707,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363001745200698,
+ "grad_norm": 0.05647560954093933,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 3.931574583053589,
+ "step": 1970
+ },
+ {
+ "epoch": 27.376963350785342,
+ "grad_norm": 0.055338893085718155,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 3.908456563949585,
+ "step": 1971
+ },
+ {
+ "epoch": 27.390924956369982,
+ "grad_norm": 0.05508219450712204,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 3.8933262825012207,
+ "step": 1972
+ },
+ {
+ "epoch": 27.404886561954626,
+ "grad_norm": 0.05787097290158272,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 3.8908863067626953,
+ "step": 1973
+ },
+ {
+ "epoch": 27.418848167539267,
+ "grad_norm": 0.05587272346019745,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 3.9261322021484375,
+ "step": 1974
+ },
+ {
+ "epoch": 27.43280977312391,
+ "grad_norm": 0.05573277547955513,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 3.9129767417907715,
+ "step": 1975
+ },
+ {
+ "epoch": 27.44677137870855,
+ "grad_norm": 0.05628194287419319,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 3.906001091003418,
+ "step": 1976
+ },
+ {
+ "epoch": 27.460732984293195,
+ "grad_norm": 0.055410996079444885,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 3.9163825511932373,
+ "step": 1977
+ },
+ {
+ "epoch": 27.474694589877835,
+ "grad_norm": 0.05579163134098053,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 3.9286627769470215,
+ "step": 1978
+ },
+ {
+ "epoch": 27.48865619546248,
+ "grad_norm": 0.05793057754635811,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 3.8955917358398438,
+ "step": 1979
+ },
+ {
+ "epoch": 27.50261780104712,
+ "grad_norm": 0.05592824146151543,
+ "learning_rate": 0.00026868712586269,
+ "loss": 3.922128677368164,
+ "step": 1980
+ },
+ {
+ "epoch": 27.516579406631763,
+ "grad_norm": 0.05768568441271782,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 3.897780418395996,
+ "step": 1981
+ },
+ {
+ "epoch": 27.530541012216403,
+ "grad_norm": 0.055304404348134995,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 3.886094570159912,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544502617801047,
+ "grad_norm": 0.054745838046073914,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 3.9272961616516113,
+ "step": 1983
+ },
+ {
+ "epoch": 27.55846422338569,
+ "grad_norm": 0.05616277828812599,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 3.8794174194335938,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57242582897033,
+ "grad_norm": 0.05605581775307655,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 3.9165992736816406,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586387434554975,
+ "grad_norm": 0.057485803961753845,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 3.902085542678833,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600349040139616,
+ "grad_norm": 0.0575849711894989,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 3.92405366897583,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61431064572426,
+ "grad_norm": 0.05563105270266533,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 3.8700900077819824,
+ "step": 1988
+ },
+ {
+ "epoch": 27.6282722513089,
+ "grad_norm": 0.05615481734275818,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 3.9526686668395996,
+ "step": 1989
+ },
+ {
+ "epoch": 27.642233856893544,
+ "grad_norm": 0.056746870279312134,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 3.9000678062438965,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656195462478184,
+ "grad_norm": 0.057345934212207794,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 3.910407543182373,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670157068062828,
+ "grad_norm": 0.05434483289718628,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 3.9132578372955322,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68411867364747,
+ "grad_norm": 0.0582718588411808,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 3.9188318252563477,
+ "step": 1993
+ },
+ {
+ "epoch": 27.698080279232112,
+ "grad_norm": 0.05566178634762764,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 3.8843226432800293,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712041884816752,
+ "grad_norm": 0.05550086498260498,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 3.904834032058716,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726003490401396,
+ "grad_norm": 0.05517689883708954,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 3.9082438945770264,
+ "step": 1996
+ },
+ {
+ "epoch": 27.739965095986037,
+ "grad_norm": 0.05598316714167595,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 3.909560203552246,
+ "step": 1997
+ },
+ {
+ "epoch": 27.75392670157068,
+ "grad_norm": 0.054991044104099274,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 3.903012275695801,
+ "step": 1998
+ },
+ {
+ "epoch": 27.767888307155324,
+ "grad_norm": 0.05612454190850258,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 3.9183692932128906,
+ "step": 1999
+ },
+ {
+ "epoch": 27.781849912739965,
+ "grad_norm": 0.055587366223335266,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 3.918454647064209,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79581151832461,
+ "grad_norm": 0.05595824122428894,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 3.895993232727051,
+ "step": 2001
+ },
+ {
+ "epoch": 27.80977312390925,
+ "grad_norm": 0.054730597883462906,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 3.9305458068847656,
+ "step": 2002
+ },
+ {
+ "epoch": 27.823734729493893,
+ "grad_norm": 0.05543222650885582,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 3.9375224113464355,
+ "step": 2003
+ },
+ {
+ "epoch": 27.837696335078533,
+ "grad_norm": 0.055638376623392105,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 3.9242658615112305,
+ "step": 2004
+ },
+ {
+ "epoch": 27.851657940663177,
+ "grad_norm": 0.05635111406445503,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 3.889681339263916,
+ "step": 2005
+ },
+ {
+ "epoch": 27.865619546247817,
+ "grad_norm": 0.056509099900722504,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 3.9191479682922363,
+ "step": 2006
+ },
+ {
+ "epoch": 27.87958115183246,
+ "grad_norm": 0.055455707013607025,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 3.936049222946167,
+ "step": 2007
+ },
+ {
+ "epoch": 27.8935427574171,
+ "grad_norm": 0.05508789047598839,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 3.901210069656372,
+ "step": 2008
+ },
+ {
+ "epoch": 27.907504363001745,
+ "grad_norm": 0.05542607232928276,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 3.9348654747009277,
+ "step": 2009
+ },
+ {
+ "epoch": 27.921465968586386,
+ "grad_norm": 0.05693232640624046,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 3.940847158432007,
+ "step": 2010
+ },
+ {
+ "epoch": 27.93542757417103,
+ "grad_norm": 0.05707051232457161,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 3.9295883178710938,
+ "step": 2011
+ },
+ {
+ "epoch": 27.949389179755673,
+ "grad_norm": 0.05778150260448456,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 3.9317305088043213,
+ "step": 2012
+ },
+ {
+ "epoch": 27.963350785340314,
+ "grad_norm": 0.05821850523352623,
+ "learning_rate": 0.00025996500713765,
+ "loss": 3.9197943210601807,
+ "step": 2013
+ },
+ {
+ "epoch": 27.977312390924958,
+ "grad_norm": 0.05592802166938782,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 3.9471435546875,
+ "step": 2014
+ },
+ {
+ "epoch": 27.991273996509598,
+ "grad_norm": 0.059447627514600754,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 3.9220168590545654,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.04082372784614563,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 2.433501958847046,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 0.6050393581390381,
+ "eval_runtime": 57.4222,
+ "eval_samples_per_second": 42.527,
+ "eval_steps_per_second": 0.679,
+ "step": 2016
+ },
+ {
+ "epoch": 28.013961605584644,
+ "grad_norm": 0.059585411101579666,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 3.8270139694213867,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027923211169284,
+ "grad_norm": 0.0652594342827797,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 3.842228412628174,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041884816753928,
+ "grad_norm": 0.05607042834162712,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 3.8498799800872803,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05584642233857,
+ "grad_norm": 0.05797959491610527,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 3.8537564277648926,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069808027923212,
+ "grad_norm": 0.06011393293738365,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 3.865638256072998,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083769633507853,
+ "grad_norm": 0.05946647375822067,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 3.8406717777252197,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097731239092496,
+ "grad_norm": 0.05657580494880676,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 3.8516461849212646,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111692844677137,
+ "grad_norm": 0.06000044196844101,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 3.859069347381592,
+ "step": 2024
+ },
+ {
+ "epoch": 28.12565445026178,
+ "grad_norm": 0.05713335797190666,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 3.8558874130249023,
+ "step": 2025
+ },
+ {
+ "epoch": 28.13961605584642,
+ "grad_norm": 0.05555511265993118,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 3.882650852203369,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153577661431065,
+ "grad_norm": 0.05772312358021736,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 3.847323179244995,
+ "step": 2027
+ },
+ {
+ "epoch": 28.167539267015705,
+ "grad_norm": 0.056661207228899,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 3.8922581672668457,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18150087260035,
+ "grad_norm": 0.05714283883571625,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 3.849281072616577,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195462478184993,
+ "grad_norm": 0.05582018196582794,
+ "learning_rate": 0.000255486047794226,
+ "loss": 3.8659892082214355,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209424083769633,
+ "grad_norm": 0.05674750730395317,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 3.8591721057891846,
+ "step": 2031
+ },
+ {
+ "epoch": 28.223385689354277,
+ "grad_norm": 0.060421206057071686,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 3.8599629402160645,
+ "step": 2032
+ },
+ {
+ "epoch": 28.237347294938917,
+ "grad_norm": 0.05727452412247658,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 3.8559670448303223,
+ "step": 2033
+ },
+ {
+ "epoch": 28.25130890052356,
+ "grad_norm": 0.05708890035748482,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 3.8643712997436523,
+ "step": 2034
+ },
+ {
+ "epoch": 28.2652705061082,
+ "grad_norm": 0.05768970027565956,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 3.842012405395508,
+ "step": 2035
+ },
+ {
+ "epoch": 28.279232111692846,
+ "grad_norm": 0.05973518267273903,
+ "learning_rate": 0.000253907826333243,
+ "loss": 3.841464042663574,
+ "step": 2036
+ },
+ {
+ "epoch": 28.293193717277486,
+ "grad_norm": 0.057536788284778595,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 3.8655552864074707,
+ "step": 2037
+ },
+ {
+ "epoch": 28.30715532286213,
+ "grad_norm": 0.05864274501800537,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 3.852839469909668,
+ "step": 2038
+ },
+ {
+ "epoch": 28.32111692844677,
+ "grad_norm": 0.056395065039396286,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 3.859243392944336,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335078534031414,
+ "grad_norm": 0.05933082476258278,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 3.843926191329956,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349040139616054,
+ "grad_norm": 0.05862351134419441,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 3.8284354209899902,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363001745200698,
+ "grad_norm": 0.05564470961689949,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 3.8802285194396973,
+ "step": 2042
+ },
+ {
+ "epoch": 28.376963350785342,
+ "grad_norm": 0.06139146164059639,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 3.867100954055786,
+ "step": 2043
+ },
+ {
+ "epoch": 28.390924956369982,
+ "grad_norm": 0.058729205280542374,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 3.8322975635528564,
+ "step": 2044
+ },
+ {
+ "epoch": 28.404886561954626,
+ "grad_norm": 0.05771554261445999,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 3.855440616607666,
+ "step": 2045
+ },
+ {
+ "epoch": 28.418848167539267,
+ "grad_norm": 0.06000557541847229,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 3.8510828018188477,
+ "step": 2046
+ },
+ {
+ "epoch": 28.43280977312391,
+ "grad_norm": 0.05935550108551979,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 3.8572049140930176,
+ "step": 2047
+ },
+ {
+ "epoch": 28.44677137870855,
+ "grad_norm": 0.056745223701000214,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 3.8892972469329834,
+ "step": 2048
+ },
+ {
+ "epoch": 28.460732984293195,
+ "grad_norm": 0.0588800348341465,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 3.865957260131836,
+ "step": 2049
+ },
+ {
+ "epoch": 28.474694589877835,
+ "grad_norm": 0.05756748467683792,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 3.8530519008636475,
+ "step": 2050
+ },
+ {
+ "epoch": 28.48865619546248,
+ "grad_norm": 0.057398825883865356,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 3.9022018909454346,
+ "step": 2051
+ },
+ {
+ "epoch": 28.50261780104712,
+ "grad_norm": 0.05809267610311508,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 3.853647232055664,
+ "step": 2052
+ },
+ {
+ "epoch": 28.516579406631763,
+ "grad_norm": 0.05625335872173309,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 3.890363931655884,
+ "step": 2053
+ },
+ {
+ "epoch": 28.530541012216403,
+ "grad_norm": 0.05852413922548294,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 3.8836288452148438,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544502617801047,
+ "grad_norm": 0.05545232817530632,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 3.902006149291992,
+ "step": 2055
+ },
+ {
+ "epoch": 28.55846422338569,
+ "grad_norm": 0.05776599794626236,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 3.8910727500915527,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57242582897033,
+ "grad_norm": 0.05847984179854393,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 3.890106201171875,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586387434554975,
+ "grad_norm": 0.056724466383457184,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 3.836899995803833,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600349040139616,
+ "grad_norm": 0.05690842866897583,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 3.834292411804199,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61431064572426,
+ "grad_norm": 0.058501504361629486,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 3.891890048980713,
+ "step": 2060
+ },
+ {
+ "epoch": 28.6282722513089,
+ "grad_norm": 0.059110380709171295,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 3.883753538131714,
+ "step": 2061
+ },
+ {
+ "epoch": 28.642233856893544,
+ "grad_norm": 0.0574255995452404,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 3.8829569816589355,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656195462478184,
+ "grad_norm": 0.05842957645654678,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 3.845698118209839,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670157068062828,
+ "grad_norm": 0.05536387115716934,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 3.9008209705352783,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68411867364747,
+ "grad_norm": 0.05769932642579079,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 3.890972137451172,
+ "step": 2065
+ },
+ {
+ "epoch": 28.698080279232112,
+ "grad_norm": 0.058036014437675476,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 3.873745918273926,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712041884816752,
+ "grad_norm": 0.05829216167330742,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 3.879685401916504,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726003490401396,
+ "grad_norm": 0.06169239804148674,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 3.9156579971313477,
+ "step": 2068
+ },
+ {
+ "epoch": 28.739965095986037,
+ "grad_norm": 0.05861479043960571,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 3.876373052597046,
+ "step": 2069
+ },
+ {
+ "epoch": 28.75392670157068,
+ "grad_norm": 0.05878114700317383,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 3.918684959411621,
+ "step": 2070
+ },
+ {
+ "epoch": 28.767888307155324,
+ "grad_norm": 0.059904664754867554,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 3.8893160820007324,
+ "step": 2071
+ },
+ {
+ "epoch": 28.781849912739965,
+ "grad_norm": 0.057385317981243134,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 3.880828619003296,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79581151832461,
+ "grad_norm": 0.05815180763602257,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 3.8638038635253906,
+ "step": 2073
+ },
+ {
+ "epoch": 28.80977312390925,
+ "grad_norm": 0.05871085450053215,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 3.8427436351776123,
+ "step": 2074
+ },
+ {
+ "epoch": 28.823734729493893,
+ "grad_norm": 0.056275319308042526,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 3.858442544937134,
+ "step": 2075
+ },
+ {
+ "epoch": 28.837696335078533,
+ "grad_norm": 0.05850474536418915,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 3.871570110321045,
+ "step": 2076
+ },
+ {
+ "epoch": 28.851657940663177,
+ "grad_norm": 0.0580497644841671,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 3.89687442779541,
+ "step": 2077
+ },
+ {
+ "epoch": 28.865619546247817,
+ "grad_norm": 0.05944157764315605,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 3.8877720832824707,
+ "step": 2078
+ },
+ {
+ "epoch": 28.87958115183246,
+ "grad_norm": 0.05749465152621269,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 3.8787431716918945,
+ "step": 2079
+ },
+ {
+ "epoch": 28.8935427574171,
+ "grad_norm": 0.05645301938056946,
+ "learning_rate": 0.000242380656502223,
+ "loss": 3.9244091510772705,
+ "step": 2080
+ },
+ {
+ "epoch": 28.907504363001745,
+ "grad_norm": 0.05700884759426117,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 3.9192817211151123,
+ "step": 2081
+ },
+ {
+ "epoch": 28.921465968586386,
+ "grad_norm": 0.05823809653520584,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 3.905806064605713,
+ "step": 2082
+ },
+ {
+ "epoch": 28.93542757417103,
+ "grad_norm": 0.05736592411994934,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 3.8907763957977295,
+ "step": 2083
+ },
+ {
+ "epoch": 28.949389179755673,
+ "grad_norm": 0.05646584555506706,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 3.892714500427246,
+ "step": 2084
+ },
+ {
+ "epoch": 28.963350785340314,
+ "grad_norm": 0.06081141158938408,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 3.864286184310913,
+ "step": 2085
+ },
+ {
+ "epoch": 28.977312390924958,
+ "grad_norm": 0.05633515864610672,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 3.8791451454162598,
+ "step": 2086
+ },
+ {
+ "epoch": 28.991273996509598,
+ "grad_norm": 0.05825566127896309,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 3.8895483016967773,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.042954832315444946,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 2.408613681793213,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 0.6065697073936462,
+ "eval_runtime": 59.1276,
+ "eval_samples_per_second": 41.301,
+ "eval_steps_per_second": 0.66,
+ "step": 2088
+ },
+ {
+ "epoch": 29.013961605584644,
+ "grad_norm": 0.06000809744000435,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 3.8419973850250244,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027923211169284,
+ "grad_norm": 0.06233612447977066,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 3.83025860786438,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041884816753928,
+ "grad_norm": 0.0551699623465538,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 3.8283262252807617,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05584642233857,
+ "grad_norm": 0.06225701421499252,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 3.822310447692871,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069808027923212,
+ "grad_norm": 0.05747336149215698,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 3.8208513259887695,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083769633507853,
+ "grad_norm": 0.0582275427877903,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 3.8013901710510254,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097731239092496,
+ "grad_norm": 0.059701140969991684,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 3.829238176345825,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111692844677137,
+ "grad_norm": 0.057550523430109024,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 3.7963826656341553,
+ "step": 2096
+ },
+ {
+ "epoch": 29.12565445026178,
+ "grad_norm": 0.05855116620659828,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 3.824028491973877,
+ "step": 2097
+ },
+ {
+ "epoch": 29.13961605584642,
+ "grad_norm": 0.056533779948949814,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 3.8151779174804688,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153577661431065,
+ "grad_norm": 0.0573289655148983,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 3.7990314960479736,
+ "step": 2099
+ },
+ {
+ "epoch": 29.167539267015705,
+ "grad_norm": 0.060061171650886536,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 3.8221707344055176,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18150087260035,
+ "grad_norm": 0.05851743370294571,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 3.793701648712158,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195462478184993,
+ "grad_norm": 0.059191569685935974,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 3.8194010257720947,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209424083769633,
+ "grad_norm": 0.05996483936905861,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 3.8316292762756348,
+ "step": 2103
+ },
+ {
+ "epoch": 29.223385689354277,
+ "grad_norm": 0.0555657260119915,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 3.821362018585205,
+ "step": 2104
+ },
+ {
+ "epoch": 29.237347294938917,
+ "grad_norm": 0.058476634323596954,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 3.8098394870758057,
+ "step": 2105
+ },
+ {
+ "epoch": 29.25130890052356,
+ "grad_norm": 0.05669668689370155,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 3.843484878540039,
+ "step": 2106
+ },
+ {
+ "epoch": 29.2652705061082,
+ "grad_norm": 0.05648602545261383,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 3.811458110809326,
+ "step": 2107
+ },
+ {
+ "epoch": 29.279232111692846,
+ "grad_norm": 0.05574130266904831,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 3.837676525115967,
+ "step": 2108
+ },
+ {
+ "epoch": 29.293193717277486,
+ "grad_norm": 0.05538785457611084,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 3.810798168182373,
+ "step": 2109
+ },
+ {
+ "epoch": 29.30715532286213,
+ "grad_norm": 0.05477031692862511,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 3.8156371116638184,
+ "step": 2110
+ },
+ {
+ "epoch": 29.32111692844677,
+ "grad_norm": 0.05664678290486336,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 3.8472559452056885,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335078534031414,
+ "grad_norm": 0.05649891495704651,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 3.824376106262207,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349040139616054,
+ "grad_norm": 0.05665554851293564,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 3.7963147163391113,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363001745200698,
+ "grad_norm": 0.054111722856760025,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 3.819735527038574,
+ "step": 2114
+ },
+ {
+ "epoch": 29.376963350785342,
+ "grad_norm": 0.05805648863315582,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 3.790189743041992,
+ "step": 2115
+ },
+ {
+ "epoch": 29.390924956369982,
+ "grad_norm": 0.05634011700749397,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 3.852161169052124,
+ "step": 2116
+ },
+ {
+ "epoch": 29.404886561954626,
+ "grad_norm": 0.0555867962539196,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 3.846884250640869,
+ "step": 2117
+ },
+ {
+ "epoch": 29.418848167539267,
+ "grad_norm": 0.057229746133089066,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 3.8432483673095703,
+ "step": 2118
+ },
+ {
+ "epoch": 29.43280977312391,
+ "grad_norm": 0.05664287880063057,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 3.842709541320801,
+ "step": 2119
+ },
+ {
+ "epoch": 29.44677137870855,
+ "grad_norm": 0.0567023791372776,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 3.844529628753662,
+ "step": 2120
+ },
+ {
+ "epoch": 29.460732984293195,
+ "grad_norm": 0.057870302349328995,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 3.853799343109131,
+ "step": 2121
+ },
+ {
+ "epoch": 29.474694589877835,
+ "grad_norm": 0.05681667849421501,
+ "learning_rate": 0.000231465389734324,
+ "loss": 3.849405288696289,
+ "step": 2122
+ },
+ {
+ "epoch": 29.48865619546248,
+ "grad_norm": 0.05738653615117073,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 3.8083410263061523,
+ "step": 2123
+ },
+ {
+ "epoch": 29.50261780104712,
+ "grad_norm": 0.058900732547044754,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 3.840458869934082,
+ "step": 2124
+ },
+ {
+ "epoch": 29.516579406631763,
+ "grad_norm": 0.05823393166065216,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 3.821415662765503,
+ "step": 2125
+ },
+ {
+ "epoch": 29.530541012216403,
+ "grad_norm": 0.057413019239902496,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 3.8713812828063965,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544502617801047,
+ "grad_norm": 0.05935502052307129,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 3.8488903045654297,
+ "step": 2127
+ },
+ {
+ "epoch": 29.55846422338569,
+ "grad_norm": 0.055285241454839706,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 3.8196067810058594,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57242582897033,
+ "grad_norm": 0.05894080176949501,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 3.8444557189941406,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586387434554975,
+ "grad_norm": 0.05878668278455734,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 3.8731908798217773,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600349040139616,
+ "grad_norm": 0.05855663865804672,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 3.831028938293457,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61431064572426,
+ "grad_norm": 0.057419586926698685,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 3.849435806274414,
+ "step": 2132
+ },
+ {
+ "epoch": 29.6282722513089,
+ "grad_norm": 0.058863408863544464,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 3.8495519161224365,
+ "step": 2133
+ },
+ {
+ "epoch": 29.642233856893544,
+ "grad_norm": 0.06081194058060646,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 3.844421863555908,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656195462478184,
+ "grad_norm": 0.05597590655088425,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 3.8413243293762207,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670157068062828,
+ "grad_norm": 0.058956339955329895,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 3.861703872680664,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68411867364747,
+ "grad_norm": 0.05883771926164627,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 3.8756394386291504,
+ "step": 2137
+ },
+ {
+ "epoch": 29.698080279232112,
+ "grad_norm": 0.06036122143268585,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 3.855403423309326,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712041884816752,
+ "grad_norm": 0.061243657022714615,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 3.861691474914551,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726003490401396,
+ "grad_norm": 0.05917691811919212,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 3.870516777038574,
+ "step": 2140
+ },
+ {
+ "epoch": 29.739965095986037,
+ "grad_norm": 0.06157418712973595,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 3.8292267322540283,
+ "step": 2141
+ },
+ {
+ "epoch": 29.75392670157068,
+ "grad_norm": 0.0605802945792675,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 3.836577892303467,
+ "step": 2142
+ },
+ {
+ "epoch": 29.767888307155324,
+ "grad_norm": 0.05750525742769241,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 3.8472204208374023,
+ "step": 2143
+ },
+ {
+ "epoch": 29.781849912739965,
+ "grad_norm": 0.057982541620731354,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 3.8682217597961426,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79581151832461,
+ "grad_norm": 0.057891640812158585,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 3.8397727012634277,
+ "step": 2145
+ },
+ {
+ "epoch": 29.80977312390925,
+ "grad_norm": 0.058175452053546906,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 3.8578720092773438,
+ "step": 2146
+ },
+ {
+ "epoch": 29.823734729493893,
+ "grad_norm": 0.05686056986451149,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 3.8467812538146973,
+ "step": 2147
+ },
+ {
+ "epoch": 29.837696335078533,
+ "grad_norm": 0.05755440890789032,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 3.844052314758301,
+ "step": 2148
+ },
+ {
+ "epoch": 29.851657940663177,
+ "grad_norm": 0.057706039398908615,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 3.818631172180176,
+ "step": 2149
+ },
+ {
+ "epoch": 29.865619546247817,
+ "grad_norm": 0.058824241161346436,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 3.8732900619506836,
+ "step": 2150
+ },
+ {
+ "epoch": 29.87958115183246,
+ "grad_norm": 0.058583300560712814,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 3.8373632431030273,
+ "step": 2151
+ },
+ {
+ "epoch": 29.8935427574171,
+ "grad_norm": 0.057403020560741425,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 3.8761391639709473,
+ "step": 2152
+ },
+ {
+ "epoch": 29.907504363001745,
+ "grad_norm": 0.0586736761033535,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 3.8518989086151123,
+ "step": 2153
+ },
+ {
+ "epoch": 29.921465968586386,
+ "grad_norm": 0.05727249011397362,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 3.839515209197998,
+ "step": 2154
+ },
+ {
+ "epoch": 29.93542757417103,
+ "grad_norm": 0.0594087690114975,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 3.834868907928467,
+ "step": 2155
+ },
+ {
+ "epoch": 29.949389179755673,
+ "grad_norm": 0.057742297649383545,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 3.861978530883789,
+ "step": 2156
+ },
+ {
+ "epoch": 29.963350785340314,
+ "grad_norm": 0.05707386136054993,
+ "learning_rate": 0.00022244633283095,
+ "loss": 3.886831760406494,
+ "step": 2157
+ },
+ {
+ "epoch": 29.977312390924958,
+ "grad_norm": 0.05631854757666588,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 3.83940052986145,
+ "step": 2158
+ },
+ {
+ "epoch": 29.991273996509598,
+ "grad_norm": 0.057860348373651505,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 3.8477416038513184,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.04080710560083389,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 2.4169907569885254,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 0.6078308820724487,
+ "eval_runtime": 59.4599,
+ "eval_samples_per_second": 41.07,
+ "eval_steps_per_second": 0.656,
+ "step": 2160
+ },
+ {
+ "epoch": 30.013961605584644,
+ "grad_norm": 0.05596645176410675,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 3.798748016357422,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027923211169284,
+ "grad_norm": 0.054763782769441605,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 3.7787699699401855,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041884816753928,
+ "grad_norm": 0.053927499800920486,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 3.779787063598633,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05584642233857,
+ "grad_norm": 0.05513143539428711,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 3.782046318054199,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069808027923212,
+ "grad_norm": 0.055234167724847794,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 3.8064064979553223,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083769633507853,
+ "grad_norm": 0.05320988595485687,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 3.8084359169006348,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097731239092496,
+ "grad_norm": 0.05671925097703934,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 3.7805888652801514,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111692844677137,
+ "grad_norm": 0.05502697452902794,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 3.791616916656494,
+ "step": 2168
+ },
+ {
+ "epoch": 30.12565445026178,
+ "grad_norm": 0.055426280945539474,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 3.7947494983673096,
+ "step": 2169
+ },
+ {
+ "epoch": 30.13961605584642,
+ "grad_norm": 0.05564495921134949,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 3.7958974838256836,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153577661431065,
+ "grad_norm": 0.055089402943849564,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 3.783299446105957,
+ "step": 2171
+ },
+ {
+ "epoch": 30.167539267015705,
+ "grad_norm": 0.055240437388420105,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 3.769829750061035,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18150087260035,
+ "grad_norm": 0.055886462330818176,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 3.805971622467041,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195462478184993,
+ "grad_norm": 0.05560130998492241,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 3.801429271697998,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209424083769633,
+ "grad_norm": 0.054780587553977966,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 3.794417381286621,
+ "step": 2175
+ },
+ {
+ "epoch": 30.223385689354277,
+ "grad_norm": 0.05751654878258705,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 3.7895469665527344,
+ "step": 2176
+ },
+ {
+ "epoch": 30.237347294938917,
+ "grad_norm": 0.054392315447330475,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 3.786479949951172,
+ "step": 2177
+ },
+ {
+ "epoch": 30.25130890052356,
+ "grad_norm": 0.057609833776950836,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 3.809680700302124,
+ "step": 2178
+ },
+ {
+ "epoch": 30.2652705061082,
+ "grad_norm": 0.054882537573575974,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 3.7928991317749023,
+ "step": 2179
+ },
+ {
+ "epoch": 30.279232111692846,
+ "grad_norm": 0.056957781314849854,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 3.790977716445923,
+ "step": 2180
+ },
+ {
+ "epoch": 30.293193717277486,
+ "grad_norm": 0.0580085888504982,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 3.851022958755493,
+ "step": 2181
+ },
+ {
+ "epoch": 30.30715532286213,
+ "grad_norm": 0.055905163288116455,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 3.784822463989258,
+ "step": 2182
+ },
+ {
+ "epoch": 30.32111692844677,
+ "grad_norm": 0.05998636782169342,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 3.811324119567871,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335078534031414,
+ "grad_norm": 0.057061366736888885,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 3.8131840229034424,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349040139616054,
+ "grad_norm": 0.05680330842733383,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 3.806260824203491,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363001745200698,
+ "grad_norm": 0.05539876967668533,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 3.793149471282959,
+ "step": 2186
+ },
+ {
+ "epoch": 30.376963350785342,
+ "grad_norm": 0.05771343782544136,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 3.7772932052612305,
+ "step": 2187
+ },
+ {
+ "epoch": 30.390924956369982,
+ "grad_norm": 0.055599041283130646,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 3.7704858779907227,
+ "step": 2188
+ },
+ {
+ "epoch": 30.404886561954626,
+ "grad_norm": 0.05746382102370262,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 3.8296337127685547,
+ "step": 2189
+ },
+ {
+ "epoch": 30.418848167539267,
+ "grad_norm": 0.05619506537914276,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 3.8134541511535645,
+ "step": 2190
+ },
+ {
+ "epoch": 30.43280977312391,
+ "grad_norm": 0.05699177086353302,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 3.8070058822631836,
+ "step": 2191
+ },
+ {
+ "epoch": 30.44677137870855,
+ "grad_norm": 0.056783925741910934,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 3.7931368350982666,
+ "step": 2192
+ },
+ {
+ "epoch": 30.460732984293195,
+ "grad_norm": 0.055537473410367966,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 3.809480905532837,
+ "step": 2193
+ },
+ {
+ "epoch": 30.474694589877835,
+ "grad_norm": 0.057826802134513855,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 3.8422718048095703,
+ "step": 2194
+ },
+ {
+ "epoch": 30.48865619546248,
+ "grad_norm": 0.0562913753092289,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 3.803966760635376,
+ "step": 2195
+ },
+ {
+ "epoch": 30.50261780104712,
+ "grad_norm": 0.05720103159546852,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 3.767852306365967,
+ "step": 2196
+ },
+ {
+ "epoch": 30.516579406631763,
+ "grad_norm": 0.05813048407435417,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 3.8423361778259277,
+ "step": 2197
+ },
+ {
+ "epoch": 30.530541012216403,
+ "grad_norm": 0.057173099368810654,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 3.8059489727020264,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544502617801047,
+ "grad_norm": 0.059024229645729065,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 3.7705254554748535,
+ "step": 2199
+ },
+ {
+ "epoch": 30.55846422338569,
+ "grad_norm": 0.056807395070791245,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 3.8056063652038574,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57242582897033,
+ "grad_norm": 0.05868417024612427,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 3.829719066619873,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586387434554975,
+ "grad_norm": 0.06023262441158295,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 3.8277781009674072,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600349040139616,
+ "grad_norm": 0.057253073900938034,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 3.8088555335998535,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61431064572426,
+ "grad_norm": 0.05946561321616173,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 3.7907581329345703,
+ "step": 2204
+ },
+ {
+ "epoch": 30.6282722513089,
+ "grad_norm": 0.05909013748168945,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 3.812042236328125,
+ "step": 2205
+ },
+ {
+ "epoch": 30.642233856893544,
+ "grad_norm": 0.05843796953558922,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 3.814551830291748,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656195462478184,
+ "grad_norm": 0.05953318253159523,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 3.8058853149414062,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670157068062828,
+ "grad_norm": 0.0616983063519001,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 3.7986061573028564,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68411867364747,
+ "grad_norm": 0.0588211789727211,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 3.8080081939697266,
+ "step": 2209
+ },
+ {
+ "epoch": 30.698080279232112,
+ "grad_norm": 0.05937618389725685,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 3.7851452827453613,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712041884816752,
+ "grad_norm": 0.05869529768824577,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 3.7893905639648438,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726003490401396,
+ "grad_norm": 0.060255810618400574,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 3.7998719215393066,
+ "step": 2212
+ },
+ {
+ "epoch": 30.739965095986037,
+ "grad_norm": 0.05810222774744034,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 3.800131320953369,
+ "step": 2213
+ },
+ {
+ "epoch": 30.75392670157068,
+ "grad_norm": 0.057548508048057556,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 3.7940425872802734,
+ "step": 2214
+ },
+ {
+ "epoch": 30.767888307155324,
+ "grad_norm": 0.05792916938662529,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 3.8089685440063477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.781849912739965,
+ "grad_norm": 0.06035939231514931,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 3.8355417251586914,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79581151832461,
+ "grad_norm": 0.058805160224437714,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 3.8146378993988037,
+ "step": 2217
+ },
+ {
+ "epoch": 30.80977312390925,
+ "grad_norm": 0.05828496813774109,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 3.8210601806640625,
+ "step": 2218
+ },
+ {
+ "epoch": 30.823734729493893,
+ "grad_norm": 0.059257738292217255,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 3.807644844055176,
+ "step": 2219
+ },
+ {
+ "epoch": 30.837696335078533,
+ "grad_norm": 0.05937729775905609,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 3.8312151432037354,
+ "step": 2220
+ },
+ {
+ "epoch": 30.851657940663177,
+ "grad_norm": 0.058580249547958374,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 3.7979660034179688,
+ "step": 2221
+ },
+ {
+ "epoch": 30.865619546247817,
+ "grad_norm": 0.0585496611893177,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 3.827874183654785,
+ "step": 2222
+ },
+ {
+ "epoch": 30.87958115183246,
+ "grad_norm": 0.05949759483337402,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 3.808143138885498,
+ "step": 2223
+ },
+ {
+ "epoch": 30.8935427574171,
+ "grad_norm": 0.058828581124544144,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 3.8351407051086426,
+ "step": 2224
+ },
+ {
+ "epoch": 30.907504363001745,
+ "grad_norm": 0.0600498728454113,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 3.8025779724121094,
+ "step": 2225
+ },
+ {
+ "epoch": 30.921465968586386,
+ "grad_norm": 0.06004577502608299,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 3.8157856464385986,
+ "step": 2226
+ },
+ {
+ "epoch": 30.93542757417103,
+ "grad_norm": 0.057676251977682114,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 3.8136918544769287,
+ "step": 2227
+ },
+ {
+ "epoch": 30.949389179755673,
+ "grad_norm": 0.0627770945429802,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 3.8470370769500732,
+ "step": 2228
+ },
+ {
+ "epoch": 30.963350785340314,
+ "grad_norm": 0.058836981654167175,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 3.8422586917877197,
+ "step": 2229
+ },
+ {
+ "epoch": 30.977312390924958,
+ "grad_norm": 0.060720525681972504,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 3.7992711067199707,
+ "step": 2230
+ },
+ {
+ "epoch": 30.991273996509598,
+ "grad_norm": 0.05844535678625107,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 3.823277473449707,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.0424773246049881,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 2.3872625827789307,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 0.6090936064720154,
+ "eval_runtime": 58.7367,
+ "eval_samples_per_second": 41.575,
+ "eval_steps_per_second": 0.664,
+ "step": 2232
+ },
+ {
+ "epoch": 31.013961605584644,
+ "grad_norm": 0.057239554822444916,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 3.764132022857666,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027923211169284,
+ "grad_norm": 0.05666050314903259,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 3.767040729522705,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041884816753928,
+ "grad_norm": 0.054038409143686295,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 3.7878732681274414,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05584642233857,
+ "grad_norm": 0.056939832866191864,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 3.7827179431915283,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069808027923212,
+ "grad_norm": 0.05488067492842674,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 3.7544898986816406,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083769633507853,
+ "grad_norm": 0.05604229122400284,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 3.7706053256988525,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097731239092496,
+ "grad_norm": 0.0550883486866951,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 3.755401611328125,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111692844677137,
+ "grad_norm": 0.055959347635507584,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 3.741966724395752,
+ "step": 2240
+ },
+ {
+ "epoch": 31.12565445026178,
+ "grad_norm": 0.05733809620141983,
+ "learning_rate": 0.000201141724176723,
+ "loss": 3.7537970542907715,
+ "step": 2241
+ },
+ {
+ "epoch": 31.13961605584642,
+ "grad_norm": 0.0551145002245903,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 3.763530731201172,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153577661431065,
+ "grad_norm": 0.055499665439128876,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 3.747159481048584,
+ "step": 2243
+ },
+ {
+ "epoch": 31.167539267015705,
+ "grad_norm": 0.0553385354578495,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 3.769044876098633,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18150087260035,
+ "grad_norm": 0.05678795278072357,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 3.7528388500213623,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195462478184993,
+ "grad_norm": 0.055931974202394485,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 3.742429733276367,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209424083769633,
+ "grad_norm": 0.05553249269723892,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 3.7760162353515625,
+ "step": 2247
+ },
+ {
+ "epoch": 31.223385689354277,
+ "grad_norm": 0.05742773041129112,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 3.7682301998138428,
+ "step": 2248
+ },
+ {
+ "epoch": 31.237347294938917,
+ "grad_norm": 0.056315284222364426,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 3.755894660949707,
+ "step": 2249
+ },
+ {
+ "epoch": 31.25130890052356,
+ "grad_norm": 0.05695716291666031,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 3.733503580093384,
+ "step": 2250
+ },
+ {
+ "epoch": 31.2652705061082,
+ "grad_norm": 0.05714774876832962,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 3.7504918575286865,
+ "step": 2251
+ },
+ {
+ "epoch": 31.279232111692846,
+ "grad_norm": 0.058987151831388474,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 3.7579331398010254,
+ "step": 2252
+ },
+ {
+ "epoch": 31.293193717277486,
+ "grad_norm": 0.05853817239403725,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 3.781409740447998,
+ "step": 2253
+ },
+ {
+ "epoch": 31.30715532286213,
+ "grad_norm": 0.057014722377061844,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 3.763247013092041,
+ "step": 2254
+ },
+ {
+ "epoch": 31.32111692844677,
+ "grad_norm": 0.05816958099603653,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 3.7801408767700195,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335078534031414,
+ "grad_norm": 0.05724027752876282,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 3.7466397285461426,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349040139616054,
+ "grad_norm": 0.055836353451013565,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 3.7770347595214844,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363001745200698,
+ "grad_norm": 0.05831955373287201,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 3.7769415378570557,
+ "step": 2258
+ },
+ {
+ "epoch": 31.376963350785342,
+ "grad_norm": 0.056602735072374344,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 3.781635284423828,
+ "step": 2259
+ },
+ {
+ "epoch": 31.390924956369982,
+ "grad_norm": 0.05784839391708374,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 3.7605199813842773,
+ "step": 2260
+ },
+ {
+ "epoch": 31.404886561954626,
+ "grad_norm": 0.057923249900341034,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 3.7671380043029785,
+ "step": 2261
+ },
+ {
+ "epoch": 31.418848167539267,
+ "grad_norm": 0.05663329362869263,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 3.7952239513397217,
+ "step": 2262
+ },
+ {
+ "epoch": 31.43280977312391,
+ "grad_norm": 0.057000353932380676,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 3.7924318313598633,
+ "step": 2263
+ },
+ {
+ "epoch": 31.44677137870855,
+ "grad_norm": 0.059976726770401,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 3.7782416343688965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.460732984293195,
+ "grad_norm": 0.05681120976805687,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 3.7918577194213867,
+ "step": 2265
+ },
+ {
+ "epoch": 31.474694589877835,
+ "grad_norm": 0.05824654549360275,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 3.7725377082824707,
+ "step": 2266
+ },
+ {
+ "epoch": 31.48865619546248,
+ "grad_norm": 0.058513302356004715,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 3.769562244415283,
+ "step": 2267
+ },
+ {
+ "epoch": 31.50261780104712,
+ "grad_norm": 0.05720348283648491,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 3.773402214050293,
+ "step": 2268
+ },
+ {
+ "epoch": 31.516579406631763,
+ "grad_norm": 0.06029265746474266,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 3.7562055587768555,
+ "step": 2269
+ },
+ {
+ "epoch": 31.530541012216403,
+ "grad_norm": 0.05892042815685272,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 3.779987335205078,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544502617801047,
+ "grad_norm": 0.06146431341767311,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 3.749899387359619,
+ "step": 2271
+ },
+ {
+ "epoch": 31.55846422338569,
+ "grad_norm": 0.056403227150440216,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 3.763558864593506,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57242582897033,
+ "grad_norm": 0.058624617755413055,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 3.7811503410339355,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586387434554975,
+ "grad_norm": 0.05978401005268097,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 3.7820377349853516,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600349040139616,
+ "grad_norm": 0.05971162021160126,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 3.7685751914978027,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61431064572426,
+ "grad_norm": 0.05576494708657265,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 3.7588582038879395,
+ "step": 2276
+ },
+ {
+ "epoch": 31.6282722513089,
+ "grad_norm": 0.05953748896718025,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 3.7525811195373535,
+ "step": 2277
+ },
+ {
+ "epoch": 31.642233856893544,
+ "grad_norm": 0.05771994963288307,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 3.7891383171081543,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656195462478184,
+ "grad_norm": 0.05822686105966568,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 3.7933766841888428,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670157068062828,
+ "grad_norm": 0.06044064089655876,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 3.783848762512207,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68411867364747,
+ "grad_norm": 0.05953317880630493,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 3.775938034057617,
+ "step": 2281
+ },
+ {
+ "epoch": 31.698080279232112,
+ "grad_norm": 0.05928120017051697,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 3.7867183685302734,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712041884816752,
+ "grad_norm": 0.05855962261557579,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 3.7872838973999023,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726003490401396,
+ "grad_norm": 0.058542776852846146,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 3.8022043704986572,
+ "step": 2284
+ },
+ {
+ "epoch": 31.739965095986037,
+ "grad_norm": 0.059078946709632874,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 3.79337739944458,
+ "step": 2285
+ },
+ {
+ "epoch": 31.75392670157068,
+ "grad_norm": 0.05854807421565056,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 3.769434928894043,
+ "step": 2286
+ },
+ {
+ "epoch": 31.767888307155324,
+ "grad_norm": 0.058214519172906876,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 3.778744697570801,
+ "step": 2287
+ },
+ {
+ "epoch": 31.781849912739965,
+ "grad_norm": 0.05981272831559181,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 3.8015875816345215,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79581151832461,
+ "grad_norm": 0.05833584442734718,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 3.7718634605407715,
+ "step": 2289
+ },
+ {
+ "epoch": 31.80977312390925,
+ "grad_norm": 0.058713216334581375,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 3.759026050567627,
+ "step": 2290
+ },
+ {
+ "epoch": 31.823734729493893,
+ "grad_norm": 0.06085464730858803,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 3.762263059616089,
+ "step": 2291
+ },
+ {
+ "epoch": 31.837696335078533,
+ "grad_norm": 0.05838872492313385,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 3.7770943641662598,
+ "step": 2292
+ },
+ {
+ "epoch": 31.851657940663177,
+ "grad_norm": 0.05941537022590637,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 3.8041505813598633,
+ "step": 2293
+ },
+ {
+ "epoch": 31.865619546247817,
+ "grad_norm": 0.058818284422159195,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 3.7724623680114746,
+ "step": 2294
+ },
+ {
+ "epoch": 31.87958115183246,
+ "grad_norm": 0.05925480276346207,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 3.7881040573120117,
+ "step": 2295
+ },
+ {
+ "epoch": 31.8935427574171,
+ "grad_norm": 0.058119483292102814,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 3.7992963790893555,
+ "step": 2296
+ },
+ {
+ "epoch": 31.907504363001745,
+ "grad_norm": 0.06139586120843887,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 3.800816297531128,
+ "step": 2297
+ },
+ {
+ "epoch": 31.921465968586386,
+ "grad_norm": 0.05812571942806244,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 3.7683157920837402,
+ "step": 2298
+ },
+ {
+ "epoch": 31.93542757417103,
+ "grad_norm": 0.05929575115442276,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 3.808281660079956,
+ "step": 2299
+ },
+ {
+ "epoch": 31.949389179755673,
+ "grad_norm": 0.057833533734083176,
+ "learning_rate": 0.000186516746349841,
+ "loss": 3.821108818054199,
+ "step": 2300
+ },
+ {
+ "epoch": 31.963350785340314,
+ "grad_norm": 0.05857503414154053,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 3.8082172870635986,
+ "step": 2301
+ },
+ {
+ "epoch": 31.977312390924958,
+ "grad_norm": 0.05893079936504364,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 3.802304267883301,
+ "step": 2302
+ },
+ {
+ "epoch": 31.991273996509598,
+ "grad_norm": 0.057888105511665344,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 3.7589073181152344,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.04306843504309654,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 2.359179735183716,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 0.6101394891738892,
+ "eval_runtime": 59.2767,
+ "eval_samples_per_second": 41.197,
+ "eval_steps_per_second": 0.658,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01396160558464,
+ "grad_norm": 0.059457872062921524,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 3.746081829071045,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02792321116929,
+ "grad_norm": 0.05826259404420853,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 3.7450389862060547,
+ "step": 2306
+ },
+ {
+ "epoch": 32.04188481675393,
+ "grad_norm": 0.05683758482336998,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 3.718315601348877,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05584642233857,
+ "grad_norm": 0.05709684267640114,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 3.7391161918640137,
+ "step": 2308
+ },
+ {
+ "epoch": 32.06980802792321,
+ "grad_norm": 0.0563763789832592,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 3.714268684387207,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083769633507856,
+ "grad_norm": 0.058482736349105835,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 3.7212839126586914,
+ "step": 2310
+ },
+ {
+ "epoch": 32.0977312390925,
+ "grad_norm": 0.05536908283829689,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 3.7573728561401367,
+ "step": 2311
+ },
+ {
+ "epoch": 32.11169284467714,
+ "grad_norm": 0.05950985103845596,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 3.715609312057495,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12565445026178,
+ "grad_norm": 0.05600661039352417,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 3.7144830226898193,
+ "step": 2313
+ },
+ {
+ "epoch": 32.139616055846425,
+ "grad_norm": 0.05814438313245773,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 3.7489242553710938,
+ "step": 2314
+ },
+ {
+ "epoch": 32.153577661431065,
+ "grad_norm": 0.0568280853331089,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 3.729698657989502,
+ "step": 2315
+ },
+ {
+ "epoch": 32.167539267015705,
+ "grad_norm": 0.05647209286689758,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 3.7484517097473145,
+ "step": 2316
+ },
+ {
+ "epoch": 32.181500872600346,
+ "grad_norm": 0.05811113119125366,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 3.7534449100494385,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19546247818499,
+ "grad_norm": 0.055132538080215454,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 3.695220470428467,
+ "step": 2318
+ },
+ {
+ "epoch": 32.20942408376963,
+ "grad_norm": 0.05722372978925705,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 3.7204041481018066,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223385689354274,
+ "grad_norm": 0.05533589795231819,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 3.7357606887817383,
+ "step": 2320
+ },
+ {
+ "epoch": 32.23734729493892,
+ "grad_norm": 0.056150201708078384,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 3.742290735244751,
+ "step": 2321
+ },
+ {
+ "epoch": 32.25130890052356,
+ "grad_norm": 0.05737832188606262,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 3.7484631538391113,
+ "step": 2322
+ },
+ {
+ "epoch": 32.2652705061082,
+ "grad_norm": 0.056017421185970306,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 3.744821071624756,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27923211169284,
+ "grad_norm": 0.05658293142914772,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 3.717287063598633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29319371727749,
+ "grad_norm": 0.05632270127534866,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 3.7172422409057617,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30715532286213,
+ "grad_norm": 0.05777734890580177,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 3.7551779747009277,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32111692844677,
+ "grad_norm": 0.05593368038535118,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 3.7148218154907227,
+ "step": 2327
+ },
+ {
+ "epoch": 32.33507853403141,
+ "grad_norm": 0.05844123661518097,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 3.7418324947357178,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34904013961606,
+ "grad_norm": 0.05616185814142227,
+ "learning_rate": 0.000179445406945268,
+ "loss": 3.7213454246520996,
+ "step": 2329
+ },
+ {
+ "epoch": 32.3630017452007,
+ "grad_norm": 0.058453164994716644,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 3.730037212371826,
+ "step": 2330
+ },
+ {
+ "epoch": 32.37696335078534,
+ "grad_norm": 0.05819960683584213,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 3.7571628093719482,
+ "step": 2331
+ },
+ {
+ "epoch": 32.390924956369986,
+ "grad_norm": 0.05889322608709335,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 3.7306394577026367,
+ "step": 2332
+ },
+ {
+ "epoch": 32.404886561954626,
+ "grad_norm": 0.05727435275912285,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 3.7304887771606445,
+ "step": 2333
+ },
+ {
+ "epoch": 32.41884816753927,
+ "grad_norm": 0.058260343968868256,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 3.714399814605713,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43280977312391,
+ "grad_norm": 0.06092117354273796,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 3.743901491165161,
+ "step": 2335
+ },
+ {
+ "epoch": 32.446771378708554,
+ "grad_norm": 0.05715249851346016,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 3.747819423675537,
+ "step": 2336
+ },
+ {
+ "epoch": 32.460732984293195,
+ "grad_norm": 0.060237716883420944,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 3.770688533782959,
+ "step": 2337
+ },
+ {
+ "epoch": 32.474694589877835,
+ "grad_norm": 0.059161923825740814,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 3.724484443664551,
+ "step": 2338
+ },
+ {
+ "epoch": 32.488656195462475,
+ "grad_norm": 0.05693944916129112,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 3.7228128910064697,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50261780104712,
+ "grad_norm": 0.05745118111371994,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 3.7756295204162598,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51657940663176,
+ "grad_norm": 0.05782153829932213,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 3.757032871246338,
+ "step": 2341
+ },
+ {
+ "epoch": 32.5305410122164,
+ "grad_norm": 0.058188699185848236,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 3.7462215423583984,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544502617801044,
+ "grad_norm": 0.05856512114405632,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 3.748447895050049,
+ "step": 2343
+ },
+ {
+ "epoch": 32.55846422338569,
+ "grad_norm": 0.05581355467438698,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 3.726031541824341,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57242582897033,
+ "grad_norm": 0.05866248160600662,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 3.7235984802246094,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58638743455497,
+ "grad_norm": 0.05732253938913345,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 3.7372875213623047,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60034904013962,
+ "grad_norm": 0.057102739810943604,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 3.768643617630005,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61431064572426,
+ "grad_norm": 0.05914342775940895,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 3.748710870742798,
+ "step": 2348
+ },
+ {
+ "epoch": 32.6282722513089,
+ "grad_norm": 0.05618954077363014,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 3.7640533447265625,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64223385689354,
+ "grad_norm": 0.05810703709721565,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 3.7230985164642334,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65619546247819,
+ "grad_norm": 0.05668202042579651,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 3.736083984375,
+ "step": 2351
+ },
+ {
+ "epoch": 32.67015706806283,
+ "grad_norm": 0.05779054015874863,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 3.7652668952941895,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68411867364747,
+ "grad_norm": 0.05945349857211113,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 3.75258207321167,
+ "step": 2353
+ },
+ {
+ "epoch": 32.69808027923211,
+ "grad_norm": 0.05697475001215935,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 3.7300591468811035,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712041884816756,
+ "grad_norm": 0.05820303037762642,
+ "learning_rate": 0.000173176617304673,
+ "loss": 3.7426252365112305,
+ "step": 2355
+ },
+ {
+ "epoch": 32.726003490401396,
+ "grad_norm": 0.05830353870987892,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 3.7716498374938965,
+ "step": 2356
+ },
+ {
+ "epoch": 32.73996509598604,
+ "grad_norm": 0.05690579116344452,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 3.7565040588378906,
+ "step": 2357
+ },
+ {
+ "epoch": 32.753926701570684,
+ "grad_norm": 0.059615254402160645,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 3.7484030723571777,
+ "step": 2358
+ },
+ {
+ "epoch": 32.767888307155324,
+ "grad_norm": 0.058043934404850006,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 3.76991868019104,
+ "step": 2359
+ },
+ {
+ "epoch": 32.781849912739965,
+ "grad_norm": 0.059192877262830734,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 3.7510385513305664,
+ "step": 2360
+ },
+ {
+ "epoch": 32.795811518324605,
+ "grad_norm": 0.05763391777873039,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 3.7479896545410156,
+ "step": 2361
+ },
+ {
+ "epoch": 32.80977312390925,
+ "grad_norm": 0.05946139991283417,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 3.7780566215515137,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82373472949389,
+ "grad_norm": 0.05837136134505272,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 3.7610299587249756,
+ "step": 2363
+ },
+ {
+ "epoch": 32.83769633507853,
+ "grad_norm": 0.05832294747233391,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 3.7216475009918213,
+ "step": 2364
+ },
+ {
+ "epoch": 32.85165794066317,
+ "grad_norm": 0.05887595936655998,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 3.767125129699707,
+ "step": 2365
+ },
+ {
+ "epoch": 32.86561954624782,
+ "grad_norm": 0.058594878762960434,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 3.751053810119629,
+ "step": 2366
+ },
+ {
+ "epoch": 32.87958115183246,
+ "grad_norm": 0.05895577743649483,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 3.7889444828033447,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8935427574171,
+ "grad_norm": 0.05831810459494591,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 3.7719945907592773,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90750436300174,
+ "grad_norm": 0.06090841069817543,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 3.7372164726257324,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92146596858639,
+ "grad_norm": 0.05728885903954506,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 3.782083511352539,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93542757417103,
+ "grad_norm": 0.061425331979990005,
+ "learning_rate": 0.00016935382741164,
+ "loss": 3.7716455459594727,
+ "step": 2371
+ },
+ {
+ "epoch": 32.94938917975567,
+ "grad_norm": 0.05963108688592911,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 3.7798497676849365,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96335078534032,
+ "grad_norm": 0.05818144232034683,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 3.746483325958252,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97731239092496,
+ "grad_norm": 0.05970250442624092,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 3.7578821182250977,
+ "step": 2374
+ },
+ {
+ "epoch": 32.9912739965096,
+ "grad_norm": 0.05898123234510422,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 3.762807846069336,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.042496208101511,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 2.353346347808838,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 0.611886203289032,
+ "eval_runtime": 58.794,
+ "eval_samples_per_second": 41.535,
+ "eval_steps_per_second": 0.663,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01396160558464,
+ "grad_norm": 0.05598964914679527,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 3.7107529640197754,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02792321116929,
+ "grad_norm": 0.057739973068237305,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 3.6825904846191406,
+ "step": 2378
+ },
+ {
+ "epoch": 33.04188481675393,
+ "grad_norm": 0.05335357040166855,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 3.684080123901367,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05584642233857,
+ "grad_norm": 0.05573783814907074,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 3.69419527053833,
+ "step": 2380
+ },
+ {
+ "epoch": 33.06980802792321,
+ "grad_norm": 0.055793728679418564,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 3.730604648590088,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083769633507856,
+ "grad_norm": 0.054736193269491196,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 3.6922354698181152,
+ "step": 2382
+ },
+ {
+ "epoch": 33.0977312390925,
+ "grad_norm": 0.05806516483426094,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 3.7343029975891113,
+ "step": 2383
+ },
+ {
+ "epoch": 33.11169284467714,
+ "grad_norm": 0.0541117899119854,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 3.7044200897216797,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12565445026178,
+ "grad_norm": 0.057081446051597595,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 3.6954641342163086,
+ "step": 2385
+ },
+ {
+ "epoch": 33.139616055846425,
+ "grad_norm": 0.05561508610844612,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 3.709800958633423,
+ "step": 2386
+ },
+ {
+ "epoch": 33.153577661431065,
+ "grad_norm": 0.05675433203577995,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 3.7057042121887207,
+ "step": 2387
+ },
+ {
+ "epoch": 33.167539267015705,
+ "grad_norm": 0.05598882585763931,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 3.7071051597595215,
+ "step": 2388
+ },
+ {
+ "epoch": 33.181500872600346,
+ "grad_norm": 0.05500704050064087,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 3.7088711261749268,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19546247818499,
+ "grad_norm": 0.057210177183151245,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 3.685708522796631,
+ "step": 2390
+ },
+ {
+ "epoch": 33.20942408376963,
+ "grad_norm": 0.05615527555346489,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 3.695154905319214,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223385689354274,
+ "grad_norm": 0.056796107441186905,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 3.7289557456970215,
+ "step": 2392
+ },
+ {
+ "epoch": 33.23734729493892,
+ "grad_norm": 0.056371185928583145,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 3.7177929878234863,
+ "step": 2393
+ },
+ {
+ "epoch": 33.25130890052356,
+ "grad_norm": 0.05646727234125137,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 3.7265491485595703,
+ "step": 2394
+ },
+ {
+ "epoch": 33.2652705061082,
+ "grad_norm": 0.05717485025525093,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 3.7198410034179688,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27923211169284,
+ "grad_norm": 0.056603338569402695,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 3.692781686782837,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29319371727749,
+ "grad_norm": 0.05612906441092491,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 3.7237164974212646,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30715532286213,
+ "grad_norm": 0.05704745277762413,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 3.698744058609009,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32111692844677,
+ "grad_norm": 0.055712319910526276,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 3.7240562438964844,
+ "step": 2399
+ },
+ {
+ "epoch": 33.33507853403141,
+ "grad_norm": 0.05763087794184685,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 3.709852695465088,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34904013961606,
+ "grad_norm": 0.05521798133850098,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 3.7102928161621094,
+ "step": 2401
+ },
+ {
+ "epoch": 33.3630017452007,
+ "grad_norm": 0.05802507698535919,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 3.697176933288574,
+ "step": 2402
+ },
+ {
+ "epoch": 33.37696335078534,
+ "grad_norm": 0.05547751113772392,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 3.7416112422943115,
+ "step": 2403
+ },
+ {
+ "epoch": 33.390924956369986,
+ "grad_norm": 0.05753115564584732,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 3.7048134803771973,
+ "step": 2404
+ },
+ {
+ "epoch": 33.404886561954626,
+ "grad_norm": 0.05642146244645119,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 3.690187931060791,
+ "step": 2405
+ },
+ {
+ "epoch": 33.41884816753927,
+ "grad_norm": 0.05780908465385437,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 3.7087855339050293,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43280977312391,
+ "grad_norm": 0.05772170051932335,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 3.7156143188476562,
+ "step": 2407
+ },
+ {
+ "epoch": 33.446771378708554,
+ "grad_norm": 0.05813417211174965,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 3.7155094146728516,
+ "step": 2408
+ },
+ {
+ "epoch": 33.460732984293195,
+ "grad_norm": 0.056307602673769,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 3.7018208503723145,
+ "step": 2409
+ },
+ {
+ "epoch": 33.474694589877835,
+ "grad_norm": 0.05811525136232376,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 3.7002341747283936,
+ "step": 2410
+ },
+ {
+ "epoch": 33.488656195462475,
+ "grad_norm": 0.056484196335077286,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 3.715003490447998,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50261780104712,
+ "grad_norm": 0.057813555002212524,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 3.744851589202881,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51657940663176,
+ "grad_norm": 0.05758954957127571,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 3.712819814682007,
+ "step": 2413
+ },
+ {
+ "epoch": 33.5305410122164,
+ "grad_norm": 0.05723990499973297,
+ "learning_rate": 0.000159218843594243,
+ "loss": 3.7235021591186523,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544502617801044,
+ "grad_norm": 0.058723341673612595,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 3.681633949279785,
+ "step": 2415
+ },
+ {
+ "epoch": 33.55846422338569,
+ "grad_norm": 0.05725013092160225,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 3.7118563652038574,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57242582897033,
+ "grad_norm": 0.05801474303007126,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 3.7071094512939453,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58638743455497,
+ "grad_norm": 0.05690481513738632,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 3.7366995811462402,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60034904013962,
+ "grad_norm": 0.05696716532111168,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 3.7125186920166016,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61431064572426,
+ "grad_norm": 0.05725441873073578,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 3.729522705078125,
+ "step": 2420
+ },
+ {
+ "epoch": 33.6282722513089,
+ "grad_norm": 0.05838564410805702,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 3.713085412979126,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64223385689354,
+ "grad_norm": 0.05807472765445709,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 3.722982883453369,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65619546247819,
+ "grad_norm": 0.05865970253944397,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 3.7254886627197266,
+ "step": 2423
+ },
+ {
+ "epoch": 33.67015706806283,
+ "grad_norm": 0.05879775062203407,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 3.7419567108154297,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68411867364747,
+ "grad_norm": 0.058425575494766235,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 3.7334160804748535,
+ "step": 2425
+ },
+ {
+ "epoch": 33.69808027923211,
+ "grad_norm": 0.05795733258128166,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 3.6824467182159424,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712041884816756,
+ "grad_norm": 0.05840831995010376,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 3.7226808071136475,
+ "step": 2427
+ },
+ {
+ "epoch": 33.726003490401396,
+ "grad_norm": 0.05915387347340584,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 3.7192320823669434,
+ "step": 2428
+ },
+ {
+ "epoch": 33.73996509598604,
+ "grad_norm": 0.05898639187216759,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 3.731309175491333,
+ "step": 2429
+ },
+ {
+ "epoch": 33.753926701570684,
+ "grad_norm": 0.058383479714393616,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 3.6850638389587402,
+ "step": 2430
+ },
+ {
+ "epoch": 33.767888307155324,
+ "grad_norm": 0.06078571453690529,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 3.7126986980438232,
+ "step": 2431
+ },
+ {
+ "epoch": 33.781849912739965,
+ "grad_norm": 0.05873129144310951,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 3.726377010345459,
+ "step": 2432
+ },
+ {
+ "epoch": 33.795811518324605,
+ "grad_norm": 0.060784872621297836,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 3.7208423614501953,
+ "step": 2433
+ },
+ {
+ "epoch": 33.80977312390925,
+ "grad_norm": 0.05992938578128815,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 3.7019124031066895,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82373472949389,
+ "grad_norm": 0.060495104640722275,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 3.7536425590515137,
+ "step": 2435
+ },
+ {
+ "epoch": 33.83769633507853,
+ "grad_norm": 0.059453774243593216,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 3.7357583045959473,
+ "step": 2436
+ },
+ {
+ "epoch": 33.85165794066317,
+ "grad_norm": 0.06152774766087532,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 3.7266006469726562,
+ "step": 2437
+ },
+ {
+ "epoch": 33.86561954624782,
+ "grad_norm": 0.05948188528418541,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 3.712332248687744,
+ "step": 2438
+ },
+ {
+ "epoch": 33.87958115183246,
+ "grad_norm": 0.05947373807430267,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 3.696099042892456,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8935427574171,
+ "grad_norm": 0.060062918812036514,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 3.7297117710113525,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90750436300174,
+ "grad_norm": 0.057695310562849045,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 3.7090835571289062,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92146596858639,
+ "grad_norm": 0.05904509872198105,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 3.763152599334717,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93542757417103,
+ "grad_norm": 0.05886756256222725,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 3.7213687896728516,
+ "step": 2443
+ },
+ {
+ "epoch": 33.94938917975567,
+ "grad_norm": 0.06159936264157295,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 3.7479395866394043,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96335078534032,
+ "grad_norm": 0.057995621114969254,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 3.7327473163604736,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97731239092496,
+ "grad_norm": 0.058990150690078735,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 3.699246883392334,
+ "step": 2446
+ },
+ {
+ "epoch": 33.9912739965096,
+ "grad_norm": 0.05924935266375542,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 3.730475425720215,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.04313633218407631,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 2.336765766143799,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 0.6134145855903625,
+ "eval_runtime": 59.4711,
+ "eval_samples_per_second": 41.062,
+ "eval_steps_per_second": 0.656,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01396160558464,
+ "grad_norm": 0.05632878839969635,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 3.667578935623169,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02792321116929,
+ "grad_norm": 0.05653175711631775,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 3.6973605155944824,
+ "step": 2450
+ },
+ {
+ "epoch": 34.04188481675393,
+ "grad_norm": 0.05482367426156998,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 3.675959587097168,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05584642233857,
+ "grad_norm": 0.05636861175298691,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 3.673849582672119,
+ "step": 2452
+ },
+ {
+ "epoch": 34.06980802792321,
+ "grad_norm": 0.05500229448080063,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 3.6991372108459473,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083769633507856,
+ "grad_norm": 0.056783828884363174,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 3.6838951110839844,
+ "step": 2454
+ },
+ {
+ "epoch": 34.0977312390925,
+ "grad_norm": 0.05543818324804306,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 3.6698226928710938,
+ "step": 2455
+ },
+ {
+ "epoch": 34.11169284467714,
+ "grad_norm": 0.05481055751442909,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 3.678802967071533,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12565445026178,
+ "grad_norm": 0.055152181535959244,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 3.706040620803833,
+ "step": 2457
+ },
+ {
+ "epoch": 34.139616055846425,
+ "grad_norm": 0.05652368441224098,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 3.669009208679199,
+ "step": 2458
+ },
+ {
+ "epoch": 34.153577661431065,
+ "grad_norm": 0.05562841519713402,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 3.6889967918395996,
+ "step": 2459
+ },
+ {
+ "epoch": 34.167539267015705,
+ "grad_norm": 0.05579658970236778,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 3.68369197845459,
+ "step": 2460
+ },
+ {
+ "epoch": 34.181500872600346,
+ "grad_norm": 0.055683478713035583,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 3.670243740081787,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19546247818499,
+ "grad_norm": 0.05624355748295784,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 3.633751392364502,
+ "step": 2462
+ },
+ {
+ "epoch": 34.20942408376963,
+ "grad_norm": 0.05518807843327522,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 3.689868927001953,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223385689354274,
+ "grad_norm": 0.05650133267045021,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 3.694890022277832,
+ "step": 2464
+ },
+ {
+ "epoch": 34.23734729493892,
+ "grad_norm": 0.0555979385972023,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 3.6806082725524902,
+ "step": 2465
+ },
+ {
+ "epoch": 34.25130890052356,
+ "grad_norm": 0.0558185949921608,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 3.6637582778930664,
+ "step": 2466
+ },
+ {
+ "epoch": 34.2652705061082,
+ "grad_norm": 0.056465502828359604,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 3.693030834197998,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27923211169284,
+ "grad_norm": 0.0570668987929821,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 3.6703829765319824,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29319371727749,
+ "grad_norm": 0.05592618137598038,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 3.6874935626983643,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30715532286213,
+ "grad_norm": 0.055560462176799774,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 3.656379222869873,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32111692844677,
+ "grad_norm": 0.056786514818668365,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 3.685967445373535,
+ "step": 2471
+ },
+ {
+ "epoch": 34.33507853403141,
+ "grad_norm": 0.055693406611680984,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 3.693864345550537,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34904013961606,
+ "grad_norm": 0.05816012993454933,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 3.7092270851135254,
+ "step": 2473
+ },
+ {
+ "epoch": 34.3630017452007,
+ "grad_norm": 0.05769353732466698,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 3.6529183387756348,
+ "step": 2474
+ },
+ {
+ "epoch": 34.37696335078534,
+ "grad_norm": 0.058298323303461075,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 3.7092885971069336,
+ "step": 2475
+ },
+ {
+ "epoch": 34.390924956369986,
+ "grad_norm": 0.055963121354579926,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 3.6982855796813965,
+ "step": 2476
+ },
+ {
+ "epoch": 34.404886561954626,
+ "grad_norm": 0.057205602526664734,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 3.690000534057617,
+ "step": 2477
+ },
+ {
+ "epoch": 34.41884816753927,
+ "grad_norm": 0.05652588978409767,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 3.6800036430358887,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43280977312391,
+ "grad_norm": 0.059187956154346466,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 3.677968978881836,
+ "step": 2479
+ },
+ {
+ "epoch": 34.446771378708554,
+ "grad_norm": 0.05786960572004318,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 3.7025253772735596,
+ "step": 2480
+ },
+ {
+ "epoch": 34.460732984293195,
+ "grad_norm": 0.06170958653092384,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 3.6797211170196533,
+ "step": 2481
+ },
+ {
+ "epoch": 34.474694589877835,
+ "grad_norm": 0.0579230934381485,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 3.6861648559570312,
+ "step": 2482
+ },
+ {
+ "epoch": 34.488656195462475,
+ "grad_norm": 0.059319235384464264,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 3.6565022468566895,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50261780104712,
+ "grad_norm": 0.05783943459391594,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 3.722566604614258,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51657940663176,
+ "grad_norm": 0.062192801386117935,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 3.6969285011291504,
+ "step": 2485
+ },
+ {
+ "epoch": 34.5305410122164,
+ "grad_norm": 0.05853543430566788,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 3.6851119995117188,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544502617801044,
+ "grad_norm": 0.05946554243564606,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 3.6720314025878906,
+ "step": 2487
+ },
+ {
+ "epoch": 34.55846422338569,
+ "grad_norm": 0.05766003578901291,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 3.665830135345459,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57242582897033,
+ "grad_norm": 0.06046074256300926,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 3.6955373287200928,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58638743455497,
+ "grad_norm": 0.057485196739435196,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 3.7058653831481934,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60034904013962,
+ "grad_norm": 0.0580572746694088,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 3.6663808822631836,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61431064572426,
+ "grad_norm": 0.05744212865829468,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 3.7024950981140137,
+ "step": 2492
+ },
+ {
+ "epoch": 34.6282722513089,
+ "grad_norm": 0.058456238359212875,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 3.6953721046447754,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64223385689354,
+ "grad_norm": 0.05971676856279373,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 3.6731884479522705,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65619546247819,
+ "grad_norm": 0.058337483555078506,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 3.7169854640960693,
+ "step": 2495
+ },
+ {
+ "epoch": 34.67015706806283,
+ "grad_norm": 0.059567082673311234,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 3.6867635250091553,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68411867364747,
+ "grad_norm": 0.05710845813155174,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 3.6897926330566406,
+ "step": 2497
+ },
+ {
+ "epoch": 34.69808027923211,
+ "grad_norm": 0.05975678935647011,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 3.699439764022827,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712041884816756,
+ "grad_norm": 0.05828933045268059,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 3.677755355834961,
+ "step": 2499
+ },
+ {
+ "epoch": 34.726003490401396,
+ "grad_norm": 0.05896460637450218,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 3.687589645385742,
+ "step": 2500
+ },
+ {
+ "epoch": 34.73996509598604,
+ "grad_norm": 0.05770917236804962,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 3.7188401222229004,
+ "step": 2501
+ },
+ {
+ "epoch": 34.753926701570684,
+ "grad_norm": 0.05971298739314079,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 3.6803629398345947,
+ "step": 2502
+ },
+ {
+ "epoch": 34.767888307155324,
+ "grad_norm": 0.05826874077320099,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 3.7006959915161133,
+ "step": 2503
+ },
+ {
+ "epoch": 34.781849912739965,
+ "grad_norm": 0.05863497033715248,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 3.680962562561035,
+ "step": 2504
+ },
+ {
+ "epoch": 34.795811518324605,
+ "grad_norm": 0.058764465153217316,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 3.7000298500061035,
+ "step": 2505
+ },
+ {
+ "epoch": 34.80977312390925,
+ "grad_norm": 0.05830385908484459,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 3.6838178634643555,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82373472949389,
+ "grad_norm": 0.05855685845017433,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 3.709963321685791,
+ "step": 2507
+ },
+ {
+ "epoch": 34.83769633507853,
+ "grad_norm": 0.05872069671750069,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 3.702852725982666,
+ "step": 2508
+ },
+ {
+ "epoch": 34.85165794066317,
+ "grad_norm": 0.05955277755856514,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 3.699869155883789,
+ "step": 2509
+ },
+ {
+ "epoch": 34.86561954624782,
+ "grad_norm": 0.061080168932676315,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 3.7028279304504395,
+ "step": 2510
+ },
+ {
+ "epoch": 34.87958115183246,
+ "grad_norm": 0.05777806416153908,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 3.6681065559387207,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8935427574171,
+ "grad_norm": 0.0598529689013958,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 3.7238996028900146,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90750436300174,
+ "grad_norm": 0.05878746509552002,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 3.696312427520752,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92146596858639,
+ "grad_norm": 0.06028176471590996,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 3.6749677658081055,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93542757417103,
+ "grad_norm": 0.05837864801287651,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 3.675783634185791,
+ "step": 2515
+ },
+ {
+ "epoch": 34.94938917975567,
+ "grad_norm": 0.060354143381118774,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 3.6795129776000977,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96335078534032,
+ "grad_norm": 0.05837022140622139,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 3.711968183517456,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97731239092496,
+ "grad_norm": 0.06035105511546135,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 3.6994686126708984,
+ "step": 2518
+ },
+ {
+ "epoch": 34.9912739965096,
+ "grad_norm": 0.05695690959692001,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 3.6764185428619385,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.04363907873630524,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 2.2967724800109863,
+ "step": 2520
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.0758391632337306e+18,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-2520/training_args.bin b/runs/l2r90-ssl/checkpoint-2520/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..103048067ef8897d0386013ed89ce6764bad5521
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2520/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f320fb44ea3a4a93198f1a6c2405447428148ce142cd94ee283a3c077a327c2d
+size 4792
diff --git a/runs/l2r90-ssl/checkpoint-2880/chat_template.jinja b/runs/l2r90-ssl/checkpoint-2880/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2880/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-ssl/checkpoint-2880/config.json b/runs/l2r90-ssl/checkpoint-2880/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2880/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-ssl/checkpoint-2880/generation_config.json b/runs/l2r90-ssl/checkpoint-2880/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2880/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-ssl/checkpoint-2880/model.safetensors b/runs/l2r90-ssl/checkpoint-2880/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..8914ee4f99499b0af67256460fcc564b86934933
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2880/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:eaddb41f42282fa166ac8b5715ad0b8efc2a9481bceed4f58ec8250a7824f049
+size 583356232
diff --git a/runs/l2r90-ssl/checkpoint-2880/optimizer.pt b/runs/l2r90-ssl/checkpoint-2880/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..aa9a9a765267ca1344a6c5625bc6a76617ad94d5
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2880/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:345b47877ed61f43da4411fe08360c37abd0835ec5cf08e4f3e80503acc2f42d
+size 1166827898
diff --git a/runs/l2r90-ssl/checkpoint-2880/rng_state_0.pth b/runs/l2r90-ssl/checkpoint-2880/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..d0a0edf9dd5d265c92220d4e979a77083deb97af
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2880/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:85c8b26cb86c92e574450b60beab12a61e8d26df90092a5c2f1efeb69e097455
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-2880/rng_state_1.pth b/runs/l2r90-ssl/checkpoint-2880/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..4fa7c978646ed846fd10feda28e54269b8f8067a
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2880/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1ac830d44175ec31da3f7539f6c1a5e88e6ae26472bd7fcea1121a59a58457da
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-2880/scheduler.pt b/runs/l2r90-ssl/checkpoint-2880/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..5c7f6b411db3da6cc58af5b4530e84dddd5abe00
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2880/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2e88932b62643ef7a711b903ac04665eab5638dccef14b4bec61d20b2791e67b
+size 1064
diff --git a/runs/l2r90-ssl/checkpoint-2880/tokenizer.json b/runs/l2r90-ssl/checkpoint-2880/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2880/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-ssl/checkpoint-2880/tokenizer_config.json b/runs/l2r90-ssl/checkpoint-2880/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2880/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-2880/trainer_state.json b/runs/l2r90-ssl/checkpoint-2880/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..83b2f71c67dd4064ea1cb20ffaa81acfcfd501aa
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2880/trainer_state.json
@@ -0,0 +1,20498 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 40.0,
+ "eval_steps": 500,
+ "global_step": 2880,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.2588357925415039,
+ "learning_rate": 0.0,
+ "loss": 3.9982247352600098,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.25455471873283386,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.998683452606201,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.2555367052555084,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.9821574687957764,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.2383691370487213,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.9531917572021484,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2023179680109024,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.9148876667022705,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.17691479623317719,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.875638961791992,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.14372293651103973,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.836796283721924,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1074419692158699,
+ "learning_rate": 4.2e-05,
+ "loss": 3.8036489486694336,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.08204353600740433,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.774808406829834,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.0632256343960762,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.7503671646118164,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.047985248267650604,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.7345473766326904,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.03675895184278488,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.721525192260742,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.02908993512392044,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.713454246520996,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.02443614974617958,
+ "learning_rate": 7.8e-05,
+ "loss": 3.7078704833984375,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02109951712191105,
+ "learning_rate": 8.4e-05,
+ "loss": 3.7034401893615723,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.018616363406181335,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.7009854316711426,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.016399752348661423,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.697768211364746,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.014256286434829235,
+ "learning_rate": 0.000102,
+ "loss": 3.6974306106567383,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.013020666316151619,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.694406509399414,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.011879238300025463,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.691807270050049,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.010290366597473621,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.6937897205352783,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.009482803754508495,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.6902384757995605,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.008425266481935978,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.686815023422241,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.007676936220377684,
+ "learning_rate": 0.000138,
+ "loss": 3.685192823410034,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.007491032127290964,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.6839442253112793,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.007428477052599192,
+ "learning_rate": 0.00015,
+ "loss": 3.6783409118652344,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.006965892389416695,
+ "learning_rate": 0.000156,
+ "loss": 3.6813433170318604,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.008106694556772709,
+ "learning_rate": 0.000162,
+ "loss": 3.6741726398468018,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.0076780058443546295,
+ "learning_rate": 0.000168,
+ "loss": 3.67673921585083,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.007904314436018467,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.6724467277526855,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.008422375656664371,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.673013925552368,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00811714306473732,
+ "learning_rate": 0.000186,
+ "loss": 3.6698389053344727,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.009063836187124252,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.668990135192871,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.01878332532942295,
+ "learning_rate": 0.000198,
+ "loss": 3.6658849716186523,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.01812782697379589,
+ "learning_rate": 0.000204,
+ "loss": 3.6697864532470703,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.020587686449289322,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.662757158279419,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.01515690516680479,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.6558732986450195,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.02510508894920349,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.6495113372802734,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.020487826317548752,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.638803243637085,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.017446046695113182,
+ "learning_rate": 0.000234,
+ "loss": 3.633307456970215,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.030667584389448166,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.6277506351470947,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.023789944127202034,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.6189804077148438,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.02660694345831871,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.610522747039795,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.023814303800463676,
+ "learning_rate": 0.000258,
+ "loss": 3.615910291671753,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.021863561123609543,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.612008571624756,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03705216199159622,
+ "learning_rate": 0.00027,
+ "loss": 3.6066887378692627,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.024143919348716736,
+ "learning_rate": 0.000276,
+ "loss": 3.5987331867218018,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.03251276910305023,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.595367431640625,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06619632989168167,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.59018611907959,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.042714253067970276,
+ "learning_rate": 0.000294,
+ "loss": 3.58981990814209,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.018893254920840263,
+ "learning_rate": 0.0003,
+ "loss": 3.578472375869751,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0749685987830162,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.5828654766082764,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.12142857164144516,
+ "learning_rate": 0.000312,
+ "loss": 3.5840892791748047,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.06944665312767029,
+ "learning_rate": 0.000318,
+ "loss": 3.580961227416992,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.04545000568032265,
+ "learning_rate": 0.000324,
+ "loss": 3.5674569606781006,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.061073049902915955,
+ "learning_rate": 0.00033,
+ "loss": 3.569991111755371,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.04133498668670654,
+ "learning_rate": 0.000336,
+ "loss": 3.5657706260681152,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.055982090532779694,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.559979200363159,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.05259208381175995,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.558039426803589,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.029976001009345055,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.5491275787353516,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.04998327046632767,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.5406670570373535,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.03230690956115723,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.5391530990600586,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.06102178618311882,
+ "learning_rate": 0.000372,
+ "loss": 3.5311684608459473,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.031715717166662216,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.518899440765381,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.06841139495372772,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.514200210571289,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.10323675721883774,
+ "learning_rate": 0.00039,
+ "loss": 3.513646125793457,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.1448228508234024,
+ "learning_rate": 0.000396,
+ "loss": 3.5110087394714355,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.12129563093185425,
+ "learning_rate": 0.000402,
+ "loss": 3.501497507095337,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.030224842950701714,
+ "learning_rate": 0.000408,
+ "loss": 3.4870352745056152,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.07492101192474365,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.482234001159668,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.03942760080099106,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.4746525287628174,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.04967200756072998,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.604069232940674,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8671882152557373,
+ "eval_runtime": 42.6602,
+ "eval_samples_per_second": 57.243,
+ "eval_steps_per_second": 0.469,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.05385325849056244,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.462052822113037,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.04870929941534996,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.459104537963867,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.08701689541339874,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.4533209800720215,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.12219952791929245,
+ "learning_rate": 0.00045,
+ "loss": 3.4558796882629395,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.0823654904961586,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.443049192428589,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.12765446305274963,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.4373486042022705,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.14215047657489777,
+ "learning_rate": 0.000468,
+ "loss": 3.449429988861084,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.03649887070059776,
+ "learning_rate": 0.000474,
+ "loss": 3.42877197265625,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09551949054002762,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.4352874755859375,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.030483486130833626,
+ "learning_rate": 0.000486,
+ "loss": 3.4125874042510986,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.0745147094130516,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.4153835773468018,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.06620422005653381,
+ "learning_rate": 0.000498,
+ "loss": 3.4053854942321777,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.06098725646734238,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.404346466064453,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.047781601548194885,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.3991761207580566,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.04409855231642723,
+ "learning_rate": 0.000516,
+ "loss": 3.396144390106201,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.04355994611978531,
+ "learning_rate": 0.000522,
+ "loss": 3.3928911685943604,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.0670008510351181,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.3831443786621094,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.0812622532248497,
+ "learning_rate": 0.000534,
+ "loss": 3.382413148880005,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.0632157102227211,
+ "learning_rate": 0.00054,
+ "loss": 3.3784892559051514,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.028686825186014175,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.3673207759857178,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.034818850457668304,
+ "learning_rate": 0.000552,
+ "loss": 3.3648972511291504,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.04509204626083374,
+ "learning_rate": 0.000558,
+ "loss": 3.35561203956604,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.054211195558309555,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.3612167835235596,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.03752049803733826,
+ "learning_rate": 0.00057,
+ "loss": 3.357103109359741,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.03620241954922676,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.3469700813293457,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.02454194985330105,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.3429036140441895,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.041030462831258774,
+ "learning_rate": 0.000588,
+ "loss": 3.338247299194336,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.03266424685716629,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.332688331604004,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.04730091616511345,
+ "learning_rate": 0.0006,
+ "loss": 3.335902452468872,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.065738245844841,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.3289852142333984,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.09062394499778748,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.3278567790985107,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.1482805460691452,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.332685947418213,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.20066532492637634,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.363180637359619,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.11500677466392517,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.3308753967285156,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.11754138022661209,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.3383703231811523,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.09672239422798157,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.3267624378204346,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.07190799713134766,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.317737579345703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.0815449133515358,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.3198161125183105,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09612327069044113,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.315814971923828,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.14795289933681488,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.317605495452881,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06667589396238327,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.3127357959747314,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.058537885546684265,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.319807767868042,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.06428155303001404,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.297541618347168,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.03881775587797165,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.297539710998535,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.061801470816135406,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.290297508239746,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.07642851769924164,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.284428596496582,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08123870939016342,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.2820234298706055,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07174684852361679,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.2697296142578125,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.04030010476708412,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.264328956604004,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.05662153288722038,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.2699337005615234,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.03761111572384834,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.255704879760742,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.043027132749557495,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.2579030990600586,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.044409338384866714,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.248547077178955,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.0239730142056942,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.249307632446289,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03628705441951752,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.2451741695404053,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.029564054682850838,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.238445997238159,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.032036975026130676,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.2320892810821533,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.030085794627666473,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.231468915939331,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.025232303887605667,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.232199192047119,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.027053389698266983,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.210745334625244,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.028464114293456078,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.210076332092285,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04616783559322357,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.2103490829467773,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.039081938564777374,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.202662467956543,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.03428936004638672,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.1997475624084473,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.04351917654275894,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.19844913482666,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.0840945914387703,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.2013347148895264,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.36558300256729126,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.26310396194458,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.09432770311832428,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.2881736755371094,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.061970267444849014,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.2840280532836914,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.05112256482243538,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.2632040977478027,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06797203421592712,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.254606246948242,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.051026541739702225,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.429918050765991,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.8117228746414185,
+ "eval_runtime": 42.8572,
+ "eval_samples_per_second": 56.98,
+ "eval_steps_per_second": 0.467,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.0608622208237648,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.2348179817199707,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.05581629276275635,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.22896146774292,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.09162136912345886,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.2343664169311523,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.13750384747982025,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.2507717609405518,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.0872517079114914,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.2440056800842285,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.07161762565374374,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.229494571685791,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04082738608121872,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.2183239459991455,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.04972497373819351,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.218075752258301,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.049254074692726135,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.213379383087158,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.056671254336833954,
+ "learning_rate": 0.000599663984012462,
+ "loss": 3.209568977355957,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.05541909113526344,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.207760810852051,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.03819117695093155,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 3.199732780456543,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.0529443584382534,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.1945419311523438,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.043010979890823364,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.185936212539673,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.07311078161001205,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.189894914627075,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06110258400440216,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 3.1911773681640625,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.06880149245262146,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 3.187437057495117,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.051326874643564224,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 3.1826558113098145,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.03363822400569916,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 3.175527572631836,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.10268989205360413,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 3.172367572784424,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.09723010659217834,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 3.1873984336853027,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.04129317030310631,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 3.165799617767334,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.05292229354381561,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 3.171780824661255,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.03873143345117569,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 3.1680874824523926,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.047018859535455704,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 3.161402940750122,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.03942955285310745,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 3.1530532836914062,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.039801858365535736,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 3.1465344429016113,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.03795155510306358,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 3.1364030838012695,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.0342850424349308,
+ "learning_rate": 0.000599379982511273,
+ "loss": 3.1391525268554688,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.03243280574679375,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 3.1331253051757812,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.029679352417588234,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 3.1346120834350586,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.030230293050408363,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 3.132615804672241,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.02845083735883236,
+ "learning_rate": 0.000599309205504528,
+ "loss": 3.1242547035217285,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.03208369016647339,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 3.119110584259033,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.040268607437610626,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 3.113833427429199,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.06951889395713806,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 3.1186914443969727,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.08879823237657547,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 3.1140987873077393,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07374681532382965,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 3.1147828102111816,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.038190409541130066,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 3.1161253452301025,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.03297388181090355,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 3.1006083488464355,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.05905177444219589,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 3.106564521789551,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.037567321211099625,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 3.0884456634521484,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.02725658379495144,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 3.09000825881958,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.04272424429655075,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 3.090528964996338,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.027587950229644775,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 3.085131883621216,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.03282274678349495,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 3.0880112648010254,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.04064026474952698,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 3.079259157180786,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.023374097421765327,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 3.0747299194335938,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.025003870949149132,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 3.0721259117126465,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.030445197597146034,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 3.0733323097229004,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.025935636833310127,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 3.070709228515625,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.021337421610951424,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 3.0627551078796387,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.040481697767972946,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 3.0529394149780273,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.047608256340026855,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 3.041020631790161,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.03356955572962761,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 3.0532307624816895,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.02101682312786579,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 3.0492985248565674,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.030098581686615944,
+ "learning_rate": 0.000598804365316586,
+ "loss": 3.0527801513671875,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.040166422724723816,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 3.0505762100219727,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.051579318940639496,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 3.0422110557556152,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.06485334783792496,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 3.0435101985931396,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.055590901523828506,
+ "learning_rate": 0.000598706872384039,
+ "loss": 3.0448739528656006,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.03183475881814957,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 3.0416858196258545,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.03995629772543907,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 3.0255980491638184,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.03907827287912369,
+ "learning_rate": 0.000598631251036868,
+ "loss": 3.032339096069336,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.027681387960910797,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 3.025439739227295,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.03470184653997421,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 3.02883243560791,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.031407181173563004,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 3.0289883613586426,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.0371522456407547,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 3.013608932495117,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.03651345893740654,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 3.0120253562927246,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.06004595756530762,
+ "learning_rate": 0.000598473577921366,
+ "loss": 3.017958402633667,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.09979518502950668,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 3.023041248321533,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.10138462483882904,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.275729179382324,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7571707367897034,
+ "eval_runtime": 42.8283,
+ "eval_samples_per_second": 57.018,
+ "eval_steps_per_second": 0.467,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.04982660710811615,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 3.016209602355957,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.06532539427280426,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 3.015043020248413,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.04369623214006424,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 3.0096778869628906,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.05180002748966217,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 3.0121543407440186,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.0440233014523983,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.9937305450439453,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.05645671859383583,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 3.0023865699768066,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.03990716487169266,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 3.0003232955932617,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.03818608075380325,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.994137763977051,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.035388972610235214,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.993170738220215,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.03162284567952156,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.9922614097595215,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.029497336596250534,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.9852986335754395,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.02586926706135273,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.9798672199249268,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.027355050668120384,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.985158920288086,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.0341978520154953,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.982581615447998,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.030727429315447807,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.9701411724090576,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.04429870471358299,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.9789254665374756,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.05775132030248642,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.9758143424987793,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.05635962635278702,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.981480121612549,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.03639192506670952,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.968752861022949,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.028748342767357826,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.9584972858428955,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.035653006285429,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.967545747756958,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.03697885945439339,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.9578099250793457,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.03518654406070709,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.9680328369140625,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.034762222319841385,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.9670262336730957,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03813016042113304,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.957522392272949,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.04372904822230339,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.9540514945983887,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.06646446138620377,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.951028823852539,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.07831092923879623,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.95898699760437,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.05779588967561722,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.965884208679199,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.04414296895265579,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.9510560035705566,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.04797815904021263,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.955803394317627,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.04482145980000496,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.954059362411499,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.037104446440935135,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.94792103767395,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.03616701066493988,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.9500932693481445,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.03882797807455063,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.9558305740356445,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.04356204345822334,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.9237029552459717,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.04535257816314697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.925853729248047,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.04519587755203247,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.924984931945801,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.055922407656908035,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.932561159133911,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.06330762803554535,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.9388322830200195,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.0517115443944931,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.9272942543029785,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.057369694113731384,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.9201431274414062,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.05687880516052246,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.922886371612549,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.055525410920381546,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.9211201667785645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04503652825951576,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.9164583683013916,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.03385302051901817,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.9145395755767822,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03564384952187538,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.916003704071045,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.04188712686300278,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.915839433670044,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03657370060682297,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.903120279312134,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.03126626834273338,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.895824432373047,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.04196290299296379,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.9093422889709473,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.1706087589263916,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.9183316230773926,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.05810421332716942,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.911245107650757,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.061330314725637436,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.9159231185913086,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.06062281131744385,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.9211535453796387,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.05590800568461418,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.9140496253967285,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.04496239125728607,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.9040920734405518,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.03835335746407509,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.9026548862457275,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.05137286335229874,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.889634132385254,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.04091101884841919,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.886838674545288,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.04076292738318443,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.892918109893799,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.045437734574079514,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.893305778503418,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.041114673018455505,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.88517165184021,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.03600773215293884,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.8779795169830322,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.04311220720410347,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.87247896194458,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.04818756878376007,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.889103412628174,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.053460877388715744,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.8899683952331543,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.06763825565576553,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.8823537826538086,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.07304801046848297,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.877340316772461,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.061337269842624664,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.8820111751556396,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.04660911113023758,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.868886709213257,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.053929202258586884,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.159337043762207,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.721762478351593,
+ "eval_runtime": 42.7977,
+ "eval_samples_per_second": 57.059,
+ "eval_steps_per_second": 0.467,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.04259369149804115,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.859579086303711,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.044425517320632935,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.8681139945983887,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.04301869496703148,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.853911876678467,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.04574296995997429,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.8724427223205566,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.050247322767972946,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.8528172969818115,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.046683840453624725,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.853405714035034,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.05966169387102127,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.8605127334594727,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.08935574442148209,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.862339973449707,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.08019179105758667,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.8655967712402344,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.04753892868757248,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.8657939434051514,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06868117302656174,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.861976146697998,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.07040470093488693,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.8562135696411133,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.05683179199695587,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.8576629161834717,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.04918001592159271,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.861943244934082,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.04921915754675865,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.8495378494262695,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.037651967257261276,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.8428618907928467,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.03522057458758354,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.8372769355773926,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.03111186809837818,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.8377151489257812,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.03906773030757904,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.846196174621582,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.02546556107699871,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.8288016319274902,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.03119821846485138,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.8334617614746094,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.030583377927541733,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.8209362030029297,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.029693076387047768,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.8267486095428467,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.028978517279028893,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.819244861602783,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.027871523052453995,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.8289647102355957,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.029287064447999,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.8235747814178467,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.024990517646074295,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.819016218185425,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.030980689451098442,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.828155040740967,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03721801936626434,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.813418388366699,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03738148882985115,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.808328866958618,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.03580411896109581,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.812070846557617,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.039036594331264496,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.807159423828125,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.044666968286037445,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.799992561340332,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.055589817464351654,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.804142951965332,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.06139020994305611,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.8182730674743652,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.07769318670034409,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.8166518211364746,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.06521332263946533,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.8142249584198,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.09083619713783264,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.8166933059692383,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.09737467020750046,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.8111555576324463,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.07148188352584839,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.8168413639068604,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.055516090244054794,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.8183553218841553,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.05338391661643982,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.8127028942108154,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.040785472840070724,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.8124775886535645,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.037135496735572815,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.8065357208251953,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.032455578446388245,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.798459053039551,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.03465662896633148,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.798649549484253,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.032026033848524094,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.7906997203826904,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.026019956916570663,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.791969060897827,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.030298635363578796,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.784593105316162,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.028147678822278976,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.782454013824463,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.023933131247758865,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.787851333618164,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.026789527386426926,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.788067579269409,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.024554884061217308,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.777033567428589,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.028679484501481056,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.786342144012451,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.026410218328237534,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.7790610790252686,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.027413051575422287,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.778944969177246,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.02908242493867874,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.763850212097168,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.029632994905114174,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.7672982215881348,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.027458716183900833,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.7702178955078125,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.03263958916068077,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.7581803798675537,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04427146911621094,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.768141508102417,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.07142394036054611,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.75819730758667,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.08833584934473038,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.782716751098633,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.08168833702802658,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.7763240337371826,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07253874838352203,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.7717690467834473,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.08186902105808258,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.767303466796875,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.05854064226150513,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.761544704437256,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.04602917283773422,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.7669177055358887,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.03920970857143402,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.766085624694824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.04551195353269577,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.7623450756073,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.040861159563064575,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.7471840381622314,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.026907291263341904,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 2.0710244178771973,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6934371590614319,
+ "eval_runtime": 42.6015,
+ "eval_samples_per_second": 57.322,
+ "eval_steps_per_second": 0.469,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.03801899403333664,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.739086627960205,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.03707238286733627,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.741638422012329,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.0387551411986351,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.740777015686035,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.028010359033942223,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.7353010177612305,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.026474518701434135,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.73842453956604,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.02920856513082981,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.736295461654663,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.026074763387441635,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.7295751571655273,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.026862047612667084,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.723040819168091,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.026300370693206787,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.7199485301971436,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.026030950248241425,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.7276217937469482,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.028342435136437416,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.7193045616149902,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.032135091722011566,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.7143118381500244,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.03068603202700615,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.7191050052642822,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.035828713327646255,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.718085765838623,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.03343002125620842,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.707131862640381,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.035051796585321426,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.7091879844665527,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.03395913913846016,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6980862617492676,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.034198079258203506,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.699704647064209,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.0322839580476284,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.702791213989258,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.041531339287757874,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.7018051147460938,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.040806882083415985,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.7096598148345947,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.04156940430402756,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6956775188446045,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.05608797073364258,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.70127272605896,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0729481428861618,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.708817720413208,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.0905836969614029,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.718085289001465,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.08345075696706772,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.7235910892486572,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.08599896728992462,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.7178783416748047,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.07301691174507141,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.714029312133789,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.08231015503406525,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.7196826934814453,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.08286412060260773,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.7176082134246826,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.08590883016586304,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.7242164611816406,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.07531397044658661,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.7225024700164795,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.04711190238595009,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.695887327194214,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.05076327919960022,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.7029781341552734,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.04495669901371002,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.7057900428771973,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.04312257468700409,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.697537899017334,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.043580975383520126,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.67315673828125,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.03360886126756668,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.6883413791656494,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.030360296368598938,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.700927257537842,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.0310810673981905,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.6917803287506104,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.02703232690691948,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.6667580604553223,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.02756045199930668,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.6702821254730225,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.02611083723604679,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.669790744781494,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.02615324780344963,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.6672821044921875,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.026000119745731354,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.6671881675720215,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.02882794849574566,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.67586088180542,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03327213600277901,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.6618690490722656,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.032920945435762405,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.665444850921631,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.02812962979078293,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.6680707931518555,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.025886602699756622,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.652898073196411,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.025222789496183395,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.6560399532318115,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.023131893947720528,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.66127347946167,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.02513662725687027,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.6463732719421387,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.027243830263614655,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.6507630348205566,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.025437965989112854,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.6485800743103027,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.02770340070128441,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.657241106033325,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.03594067320227623,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.6281137466430664,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04486451297998428,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.6340909004211426,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.05102410912513733,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.6463394165039062,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04514491558074951,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.636554002761841,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04469453915953636,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.6453680992126465,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.05421523004770279,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.646010398864746,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04779462888836861,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.6315622329711914,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.06697724014520645,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.6416730880737305,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.07520416378974915,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.641632318496704,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.055692676454782486,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.6289877891540527,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.04539895057678223,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.6363251209259033,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.04829065501689911,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.641608476638794,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.041314542293548584,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.6269643306732178,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.04100198298692703,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.6243174076080322,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.04105962812900543,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.626585006713867,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.03172122687101364,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9550899267196655,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6614077091217041,
+ "eval_runtime": 42.9009,
+ "eval_samples_per_second": 56.922,
+ "eval_steps_per_second": 0.466,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.03258531913161278,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.612126588821411,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.040250636637210846,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.6025328636169434,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.034234337508678436,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.6085972785949707,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.03635786101222038,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.6126222610473633,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.039404451847076416,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.6035327911376953,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.038504838943481445,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5985798835754395,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.03744236007332802,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5873115062713623,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.034184787422418594,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.5899314880371094,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.03766429424285889,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.602734327316284,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.03741588443517685,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.592207193374634,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.03187853842973709,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5850372314453125,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.038447409868240356,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.592654228210449,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.04046263173222542,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.5881710052490234,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.03872602432966232,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5881471633911133,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.03960775211453438,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5826945304870605,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.033132150769233704,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.584038734436035,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.029504230245947838,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.584357261657715,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.03665253147482872,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5677168369293213,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.037208545953035355,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.560795783996582,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.038244519382715225,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.591458559036255,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.043307431042194366,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.5642588138580322,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.055119726806879044,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.57377290725708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.053323689848184586,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5767312049865723,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.0502864345908165,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.562190055847168,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.05568842589855194,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.565556049346924,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.061149828135967255,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5883288383483887,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.059933699667453766,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.5710184574127197,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.06041933596134186,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.572603225708008,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.05482695251703262,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.5667271614074707,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.04712968319654465,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.554041624069214,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.04842432215809822,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.5660672187805176,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.04273047298192978,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.554839849472046,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.05356749892234802,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.5521483421325684,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.06062676012516022,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.5551886558532715,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.05431462824344635,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5654542446136475,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.0505850687623024,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5648036003112793,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.04696856811642647,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.5559377670288086,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.042585279792547226,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.5660319328308105,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.04026484116911888,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.551206588745117,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03382421284914017,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.556361198425293,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.031177924945950508,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.5413894653320312,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.028217196464538574,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.5600497722625732,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.027062954381108284,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5447568893432617,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.03001212701201439,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.540614604949951,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.028869854286313057,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.544583797454834,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.02869918756186962,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.5471596717834473,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03359716013073921,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5416982173919678,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.04000182822346687,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.534921407699585,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.043030042201280594,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5368564128875732,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.04329848289489746,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.539764881134033,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.03740967437624931,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.5246665477752686,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.03733935207128525,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.529362916946411,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.039609868079423904,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.542544364929199,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.03937261551618576,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5366649627685547,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.03261927142739296,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.5250535011291504,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.03904294595122337,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.520721435546875,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.03913239762187004,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.5122013092041016,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.03610851988196373,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.5212767124176025,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.03944548964500427,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.537186622619629,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04762452840805054,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.5204620361328125,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04427091404795647,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5286149978637695,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04366641491651535,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.5398361682891846,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.04644155502319336,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.519331932067871,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.04372382536530495,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.531022787094116,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.04404324293136597,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.521758794784546,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.042115990072488785,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.5146074295043945,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.03796013817191124,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.5040054321289062,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.04327977076172829,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.5209312438964844,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04150385782122612,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.515871286392212,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.04465346783399582,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.5247764587402344,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.05690346285700798,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.501911163330078,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04907930642366409,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8968371152877808,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6358620524406433,
+ "eval_runtime": 42.9991,
+ "eval_samples_per_second": 56.792,
+ "eval_steps_per_second": 0.465,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.03755255416035652,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.486576795578003,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.040234558284282684,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.501087188720703,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.041416723281145096,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4942233562469482,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.039655014872550964,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.503453254699707,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.041707683354616165,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.490978717803955,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.04545537009835243,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4757118225097656,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.04328389838337898,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.504516839981079,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.043312299996614456,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.4920215606689453,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.046466976404190063,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.494760513305664,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.041762981563806534,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4875552654266357,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.03713781014084816,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.495652675628662,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.040603458881378174,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4821407794952393,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.037969645112752914,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.484981060028076,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.03652074560523033,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4736409187316895,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.03783264383673668,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.483346700668335,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.03404324874281883,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.48856782913208,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.03294725716114044,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.483706474304199,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.0347001850605011,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.474888324737549,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.0359107106924057,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.478990077972412,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.03738103806972504,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.4709420204162598,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.03722188249230385,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.465726852416992,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.034907322376966476,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.482231855392456,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.035695310682058334,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4785828590393066,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.04009179025888443,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.4795875549316406,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.038996778428554535,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.478696823120117,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04082407429814339,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.465200901031494,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04856136813759804,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4605681896209717,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.05671004578471184,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.47277569770813,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.0481606163084507,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.476151466369629,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.037689510732889175,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.466142177581787,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.05005146935582161,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.4798035621643066,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.04745829105377197,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.46163272857666,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.04638158157467842,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4728071689605713,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.04530330374836922,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.4621877670288086,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.03971488028764725,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.474224090576172,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.04521876201033592,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.46122670173645,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03923020884394646,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.4603679180145264,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03986494988203049,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.4589664936065674,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.04280020296573639,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.4624319076538086,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.03930043429136276,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.47080135345459,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.0314876064658165,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.458728790283203,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.03189660236239433,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4467148780822754,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03264275938272476,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4613699913024902,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03021484799683094,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.4535279273986816,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.030740290880203247,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.4621875286102295,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03096810355782509,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.458500385284424,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03105205111205578,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.461958646774292,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.03330068290233612,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.449777364730835,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.032831475138664246,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.4616191387176514,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.0387776717543602,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.445256471633911,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.041803423315286636,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.4545960426330566,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.04148292914032936,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.433408737182617,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.04648579657077789,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.449134349822998,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.04186409339308739,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.4449565410614014,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.038569554686546326,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.443675994873047,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03908300772309303,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4626526832580566,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.038897737860679626,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.438591957092285,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.0360480472445488,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.436077117919922,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0429830439388752,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4384970664978027,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.04465422406792641,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.458029270172119,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.037096526473760605,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.436816930770874,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.047823466360569,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.44439959526062,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.05929310992360115,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.4599215984344482,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.053512319922447205,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4387903213500977,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.042625587433576584,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4359588623046875,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.056042976677417755,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.4441447257995605,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05813189223408699,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4399144649505615,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.06313478201627731,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.453447103500366,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.05281798169016838,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.4428963661193848,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.03991910442709923,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4359219074249268,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.047981832176446915,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4506607055664062,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.03646523132920265,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8225739002227783,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6196661591529846,
+ "eval_runtime": 43.8106,
+ "eval_samples_per_second": 55.74,
+ "eval_steps_per_second": 0.457,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.03673036769032478,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.4227209091186523,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.044547371566295624,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.4073169231414795,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.04089285805821419,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4279446601867676,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.03538106009364128,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.4234392642974854,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.03152978792786598,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4214606285095215,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.03324202075600624,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.4085447788238525,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.03352442383766174,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.417776107788086,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.03582802042365074,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.411144256591797,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.0351114459335804,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.405872344970703,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.038898173719644547,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4029338359832764,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.03926656022667885,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4100089073181152,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.04285634309053421,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.410360336303711,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.04179581627249718,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4093637466430664,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.03356701880693436,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4118056297302246,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.038140635937452316,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.4154677391052246,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.04468780383467674,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416846513748169,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.062421441078186035,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4157638549804688,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06317313015460968,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.418593406677246,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.049498945474624634,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4204845428466797,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.046322405338287354,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.403489589691162,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.053081002086400986,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.4148740768432617,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05222945287823677,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.416757583618164,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04893574118614197,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4241795539855957,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04951956123113632,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.401386260986328,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04019852355122566,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.4168622493743896,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.03625848889350891,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4026846885681152,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.03464345261454582,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4021215438842773,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.03750818222761154,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4113571643829346,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.0360100194811821,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.3977622985839844,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.03464623540639877,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.413273334503174,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.03193575143814087,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.423659324645996,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.02710454724729061,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.396085262298584,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.02830432914197445,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.420255422592163,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.029717663303017616,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.3958120346069336,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.030005933716893196,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3908543586730957,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.029054716229438782,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.3873138427734375,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.031556472182273865,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3892760276794434,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03518872708082199,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4113478660583496,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.03471249341964722,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.384857654571533,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03223155066370964,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3840765953063965,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.030012644827365875,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.388247489929199,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.03162285313010216,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.3987512588500977,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.03700442612171173,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.3993430137634277,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.03710822016000748,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.387359380722046,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.028411181643605232,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3925249576568604,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031246855854988098,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.38845157623291,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.034664880484342575,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.392436981201172,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.03354701027274132,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.390890121459961,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.03527764603495598,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.3954997062683105,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03325049206614494,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3847177028656006,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.028021875768899918,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.385910987854004,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.02964954823255539,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3787484169006348,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.03068520873785019,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.400784969329834,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03416873514652252,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.391033411026001,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.03534875437617302,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.3796474933624268,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.03732474148273468,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3831934928894043,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.03648678958415985,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3785600662231445,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03801294043660164,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.387587070465088,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.04186025261878967,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.3893089294433594,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.04416101053357124,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3852195739746094,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.04137178510427475,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.391308307647705,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.05010323226451874,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.3906192779541016,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.049587685614824295,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.3846192359924316,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.04125971347093582,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.384507179260254,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.039676666259765625,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.38167142868042,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.038775283843278885,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.3789167404174805,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.03634221479296684,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.383060932159424,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03464275971055031,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.388552665710449,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.031778786331415176,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.3903298377990723,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.027904309332370758,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.377493381500244,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.028964119032025337,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.381959915161133,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.0233962032943964,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7810559272766113,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6070130467414856,
+ "eval_runtime": 50.871,
+ "eval_samples_per_second": 48.004,
+ "eval_steps_per_second": 0.393,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.030209315940737724,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.3569841384887695,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.04112662002444267,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.376680374145508,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.055858466774225235,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.3621644973754883,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.06746986508369446,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3607563972473145,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.06736797094345093,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3548707962036133,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.0562627948820591,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.364654541015625,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.051833927631378174,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.3748884201049805,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.04824110120534897,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.3502087593078613,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.05518793687224388,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3687515258789062,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.06564056873321533,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.367131233215332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.0605909489095211,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.3589868545532227,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.05744732916355133,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.370117664337158,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.06430432945489883,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.3703484535217285,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.06186915934085846,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3720293045043945,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.05977998673915863,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.3512699604034424,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.06413097679615021,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.377257823944092,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.053935080766677856,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.350717544555664,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.049110863357782364,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.3688902854919434,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.04313219338655472,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3374459743499756,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.039556123316287994,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3482556343078613,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.03796547278761864,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3547937870025635,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.03646104037761688,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.344686269760132,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.03355357423424721,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.3588008880615234,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.03004707582294941,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.352043628692627,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.030186176300048828,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3500709533691406,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.029748762026429176,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.3523058891296387,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.03238696977496147,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3635129928588867,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.032599564641714096,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.346269130706787,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.028626957908272743,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.357151746749878,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.02909751422703266,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3557677268981934,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.029951099306344986,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.34753680229187,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.029768310487270355,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3408985137939453,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.03005973994731903,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.362542152404785,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03360462561249733,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.352529287338257,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.034002047032117844,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.3334460258483887,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.03292316198348999,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3446991443634033,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.0337117575109005,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.350605010986328,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03428212180733681,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.357408046722412,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.03169601410627365,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.349088191986084,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.03389411047101021,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.361576795578003,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.036497049033641815,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.35420560836792,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03574785590171814,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3678526878356934,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.030757993459701538,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3520517349243164,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.0289121363312006,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.337864637374878,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.028012607246637344,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.353999614715576,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.027558492496609688,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.342132568359375,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.02871510572731495,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3425655364990234,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03070886805653572,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3401665687561035,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.03056657873094082,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.345277786254883,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.030907969921827316,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.346494674682617,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0318334735929966,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.3390543460845947,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.029700733721256256,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.3398728370666504,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.029687359929084778,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.3474364280700684,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.027234600856900215,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3457484245300293,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.02823043428361416,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.337700366973877,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03002646192908287,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.335954189300537,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.027246717363595963,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.330207347869873,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.029703574255108833,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.328280448913574,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.028796236962080002,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3448386192321777,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.027944380417466164,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.320234537124634,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.030626650899648666,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.334698438644409,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.029164662584662437,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3450326919555664,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03034045360982418,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.3420588970184326,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.02856968529522419,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3396072387695312,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.02892562374472618,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.340027332305908,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.02938242256641388,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.33550763130188,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.029261089861392975,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3246421813964844,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.030340535566210747,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.3465514183044434,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.033335570245981216,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.3467206954956055,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.03452218323945999,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.343702793121338,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.03326255828142166,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.337631940841675,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025433633476495743,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7402067184448242,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03175730258226395,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.62623405456543,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.04780422896146774,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.623239517211914,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.0504758283495903,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.6369194984436035,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.04481920227408409,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.57640266418457,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0440552793443203,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.630992889404297,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0503336526453495,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.627910137176514,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.053885314613580704,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.637897491455078,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.05481956899166107,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.610715866088867,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.050510745495557785,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.639884948730469,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.04579266160726547,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.590188980102539,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.04817410930991173,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.636295318603516,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.058862220495939255,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.614317417144775,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.06328638643026352,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.655928611755371,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.058752477169036865,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.652004241943359,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.06049015372991562,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.6513800621032715,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.058941151946783066,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.657688140869141,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.05542639642953873,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.637376308441162,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06024057790637016,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.63592529296875,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.06040216609835625,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.6275787353515625,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.059600673615932465,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.652995586395264,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.05379730463027954,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.597358226776123,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05355260521173477,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.6106367111206055,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.06649646162986755,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.654443740844727,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.058309994637966156,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.6869659423828125,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.05403570458292961,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.654750823974609,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04678157716989517,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.653955459594727,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04297948628664017,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.644915580749512,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.04235142841935158,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.662282943725586,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.0405130498111248,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.64569091796875,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.03351007401943207,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.653165817260742,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.03182631731033325,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.606291770935059,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.031552448868751526,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.642643928527832,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027941036969423294,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.623664379119873,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.027744563296437263,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.605886459350586,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02962736040353775,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.606762886047363,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.027251560240983963,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.620492935180664,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.027894044294953346,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.613903999328613,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.02778882533311844,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.634457588195801,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.028082096949219704,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.610980033874512,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.026210976764559746,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.612090110778809,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.026537755504250526,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.581171035766602,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.028709523379802704,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.628799915313721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.028233813121914864,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.586182117462158,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.02966046892106533,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.62320613861084,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.03198636695742607,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.6018524169921875,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.033404890447854996,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.610121726989746,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.029217196628451347,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.601312160491943,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.029335757717490196,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.620820999145508,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.031007781624794006,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.626503944396973,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.028453458100557327,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.589323997497559,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02526312880218029,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.598350524902344,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.027317512780427933,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.62388277053833,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02710605598986149,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.606523513793945,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.027491388842463493,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.586644172668457,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.027158567681908607,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.606039047241211,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.0298276599496603,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.593776702880859,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.026622479781508446,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.59510612487793,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.028238043189048767,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.5899505615234375,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.029481008648872375,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.6092071533203125,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.030138323083519936,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.600069046020508,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.02951825223863125,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.600125789642334,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.029242202639579773,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.590723037719727,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.02941613830626011,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.623342514038086,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.029764151200652122,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.631185531616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.031503792852163315,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.605094909667969,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.03184918314218521,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.592767715454102,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.03223618119955063,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.5930585861206055,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.0340263694524765,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.609347343444824,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.03245621174573898,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.598440170288086,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.03204867243766785,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.619702339172363,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.030866602435708046,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.609165191650391,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.01960863545536995,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8655829429626465,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.593852162361145,
+ "eval_runtime": 60.0971,
+ "eval_samples_per_second": 40.634,
+ "eval_steps_per_second": 0.649,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03552858531475067,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.539240837097168,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.0536109134554863,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.52609395980835,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.061411239206790924,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.536109447479248,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.06352879852056503,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.546850204467773,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.05683520436286926,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.563828468322754,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.05522637069225311,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.550929069519043,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.058990392833948135,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.554411888122559,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.057443976402282715,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.545921325683594,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.05383576825261116,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.5481414794921875,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04797440022230148,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.55556583404541,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04923664405941963,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.545513153076172,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.056191347539424896,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.536705017089844,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.06088989973068237,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.570624351501465,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.04802761599421501,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.526930809020996,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.04018320515751839,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.562432289123535,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.04280981421470642,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.559737205505371,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.04258878529071808,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.561749458312988,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.04411563277244568,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.562010765075684,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.041077401489019394,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.573984622955322,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.04125386103987694,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.550564289093018,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.041974026709795,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.552739143371582,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03689907118678093,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.540107727050781,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03586242347955704,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.554999351501465,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.036602675914764404,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.546337127685547,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.03707926720380783,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.543377876281738,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.03322125971317291,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.574325084686279,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03359243646264076,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.5613603591918945,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.0314965583384037,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.530994415283203,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.031806789338588715,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.56138801574707,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.034988485276699066,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.553149700164795,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.037627559155225754,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.555401802062988,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03723530471324921,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.52452278137207,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.036220453679561615,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.54799747467041,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03228041157126427,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.574814319610596,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.030872676521539688,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.539666652679443,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.033935241401195526,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.579252243041992,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.034338828176259995,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.538324356079102,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.03441951423883438,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.540358543395996,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03328141197562218,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.520340919494629,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03224223107099533,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.5646209716796875,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.02999890223145485,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.504915714263916,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.03014362044632435,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.516615867614746,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.029469173401594162,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.560475826263428,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.027887821197509766,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.586333751678467,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.029970254749059677,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.530396461486816,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.031648170202970505,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.554404258728027,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.0325581431388855,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.544102668762207,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.03555625304579735,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.543701171875,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.033053912222385406,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.508375644683838,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.03059076890349388,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.545750617980957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.029640182852745056,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.54715633392334,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030139951035380363,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.533708572387695,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029753828421235085,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.543603897094727,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.029481034725904465,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.534200668334961,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.028991524130105972,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.5683674812316895,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.028145765885710716,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.543053150177002,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.02999844029545784,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.5482354164123535,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03134117275476456,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.5360107421875,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.029277140274643898,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.5729827880859375,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028516866266727448,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.543548583984375,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028387080878019333,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.552616596221924,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.028977863490581512,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.5456695556640625,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.028010815382003784,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.534921646118164,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.030222605913877487,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.561526298522949,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.03140007704496384,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.500124931335449,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.029269184917211533,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.536653995513916,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.02935507707297802,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.521398544311523,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02801917865872383,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.527442455291748,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.027995768934488297,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.556139945983887,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.028441084548830986,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.555442810058594,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.029100807383656502,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.532083511352539,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020385215058922768,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.8488006591796875,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5902590751647949,
+ "eval_runtime": 60.1424,
+ "eval_samples_per_second": 40.604,
+ "eval_steps_per_second": 0.648,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03513244539499283,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.486500263214111,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.052679311484098434,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.489691734313965,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.06171489134430885,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.471555709838867,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.06164781004190445,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.486891746520996,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.06410388648509979,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.494929313659668,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06399523466825485,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.5334320068359375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.0636502280831337,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.5039381980896,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.06139931082725525,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.512765884399414,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.07460282742977142,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.477229118347168,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.075187548995018,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.484212398529053,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.07584068924188614,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.495598793029785,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.07192158699035645,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.507598400115967,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.05366869643330574,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.514532089233398,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.06397061794996262,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.512920379638672,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.06391509622335434,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.465702056884766,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.059490568935871124,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.5351881980896,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.057501111179590225,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.510270595550537,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.056947533041238785,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.509241104125977,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.050393324345350266,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.50428581237793,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.057608287781476974,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.474919319152832,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.054470717906951904,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.521528244018555,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.054405540227890015,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.510211944580078,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.051328495144844055,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.480720520019531,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.045930180698633194,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.503024101257324,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.04298274591565132,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.49046516418457,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.0406254380941391,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.544064521789551,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.03557496517896652,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.496966361999512,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.0364493802189827,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.5047502517700195,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.034521300345659256,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.485365867614746,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.03305799141526222,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.484884738922119,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.033053625375032425,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.478784561157227,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03426017984747887,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.482428073883057,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.03638783469796181,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.433219909667969,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.03597250208258629,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.522584915161133,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.03086097538471222,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.500803470611572,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.029642771929502487,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.444099426269531,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.03190716728568077,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.493034362792969,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.0312873050570488,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.487887382507324,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03201943263411522,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.479177474975586,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.033088769763708115,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.5020246505737305,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.02975946106016636,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.455426216125488,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03213953226804733,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.501615047454834,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.03262908384203911,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.485121250152588,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.031102091073989868,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.488893508911133,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.02945449948310852,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.490121841430664,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.033680472522974014,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.51833438873291,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.03277525305747986,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.480165481567383,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.03303041681647301,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.501155376434326,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.034927863627672195,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.513175010681152,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.03098093718290329,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.503798484802246,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.029513906687498093,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.49006986618042,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029630785807967186,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.522514343261719,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.030888330191373825,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.4895734786987305,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.032392196357250214,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.471953868865967,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.03215661272406578,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.48297643661499,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.03346327692270279,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.518160820007324,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.031311362981796265,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.484991550445557,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.03056488186120987,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.525470733642578,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.03043230250477791,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.510650634765625,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.02916652522981167,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.496840953826904,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.030673664063215256,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.494868278503418,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030033906921744347,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.479182720184326,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.028828797861933708,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.471983909606934,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028316056355834007,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.5049147605896,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.03158092871308327,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.449270248413086,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.03041706420481205,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.508289813995361,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.02967553399503231,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.465579032897949,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030721619725227356,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.471003532409668,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03066832385957241,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.488987445831299,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.032297998666763306,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.491018295288086,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03221398591995239,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.471237659454346,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.021120507270097733,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.8080053329467773,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5888046026229858,
+ "eval_runtime": 59.1638,
+ "eval_samples_per_second": 41.275,
+ "eval_steps_per_second": 0.659,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.03799329698085785,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.401787757873535,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.05785121023654938,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.408820629119873,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.05964238941669464,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.427263259887695,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05466582626104355,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.403415679931641,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.05613672360777855,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.413750648498535,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.05396881327033043,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.41916561126709,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.05689029023051262,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.395590782165527,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.056090328842401505,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.421253204345703,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.0473211295902729,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.385158061981201,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.04464765638113022,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.431746959686279,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.046421997249126434,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.428816795349121,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.05039866268634796,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.452417850494385,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.05491664633154869,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.447483062744141,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05866196006536484,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.434992790222168,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.05532094091176987,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.450716018676758,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.051083534955978394,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.402632713317871,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.03815016150474548,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.433640956878662,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04203395918011665,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.415888786315918,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.043729111552238464,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.423260688781738,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.039426498115062714,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.428949356079102,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.03837858885526657,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.434070587158203,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.038315922021865845,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.471246719360352,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.03963236138224602,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.4434990882873535,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04027022793889046,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.436992645263672,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.0409243181347847,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.426936149597168,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03404631093144417,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.43462610244751,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.034393735229969025,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.437409400939941,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03635180741548538,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.451286315917969,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03349733352661133,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.419209003448486,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.03238677605986595,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.456454277038574,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.034473925828933716,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.3988728523254395,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.03455805033445358,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.434884071350098,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.037486057728528976,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.466741561889648,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.03617125004529953,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.429305076599121,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.033646389842033386,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.407780170440674,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03401118144392967,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.42977237701416,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.036534134298563004,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.451094627380371,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.0384049266576767,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.414777755737305,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03460627421736717,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.420644760131836,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.033746346831321716,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.461175918579102,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.037153810262680054,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.444699287414551,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03283539041876793,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.431022644042969,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03248969465494156,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.447571277618408,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03327852487564087,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.42236328125,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.031114818528294563,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.418244361877441,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.03231207653880119,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.455004692077637,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.030702251940965652,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.483610153198242,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031027289107441902,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.4470109939575195,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.03129073604941368,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.452799320220947,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.03634301945567131,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4661664962768555,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03718913346529007,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.412006855010986,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.034018341451883316,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.436755180358887,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032502319663763046,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.452866554260254,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.032902952283620834,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.449224472045898,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.0340626984834671,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.445425510406494,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.033442508429288864,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.432218551635742,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.03489428386092186,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.4205827713012695,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.031868621706962585,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.414196968078613,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.032986633479595184,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.46116304397583,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.03348355367779732,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.424936771392822,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03174544870853424,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.434438705444336,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.03293656185269356,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.440502166748047,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.035124607384204865,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.440943717956543,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03438971936702728,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.431408882141113,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.03537093102931976,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.43867301940918,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.03267325833439827,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.462226867675781,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.030995232984423637,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.433740615844727,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.03384929150342941,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.508179664611816,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.032025814056396484,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.407378196716309,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03288578242063522,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.455531120300293,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.032061848789453506,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.452212333679199,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022320276126265526,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7567386627197266,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5878658294677734,
+ "eval_runtime": 59.5054,
+ "eval_samples_per_second": 41.038,
+ "eval_steps_per_second": 0.655,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.039114587008953094,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.347412586212158,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.05164389684796333,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.391279220581055,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055882278829813004,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.355602264404297,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.05625041201710701,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3934855461120605,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05861108750104904,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.364226341247559,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.052431970834732056,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.363334655761719,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.06147193908691406,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.384820938110352,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.06701470166444778,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.373311996459961,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.06714873015880585,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.32745361328125,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.06684907525777817,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.373660087585449,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.05738082900643349,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.362325668334961,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05306725949048996,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.351685047149658,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.05831029266119003,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.374813079833984,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05547972396016121,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.361842632293701,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.048057056963443756,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.383205413818359,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.04682513326406479,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.365846157073975,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.05001377314329147,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.362151145935059,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.05140263959765434,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.376064300537109,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.04640384018421173,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.363801002502441,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04175020009279251,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.3477983474731445,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.04134776070713997,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.337469100952148,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.042178574949502945,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.351490020751953,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.03989115729928017,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.40725040435791,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04086652770638466,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.341119766235352,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.03931201249361038,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.3786468505859375,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.03855905309319496,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.3701605796813965,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.03803952783346176,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.401841163635254,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.036908265203237534,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.410647392272949,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.0398186556994915,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.380431175231934,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03899035602807999,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.374293804168701,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.03946344554424286,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.396918773651123,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.03738604858517647,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.404016017913818,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.03579043596982956,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.396914958953857,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.037139154970645905,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.407739639282227,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.03581099584698677,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.411419868469238,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.03548453748226166,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.332690238952637,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.037561237812042236,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.381314277648926,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.035827480256557465,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.371536731719971,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03538314625620842,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.395467758178711,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.036672890186309814,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.420590400695801,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.038981080055236816,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.395296096801758,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.038234200328588486,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.415799140930176,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.038830097764730453,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.413394927978516,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03990919142961502,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.39515495300293,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.036424681544303894,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.39730978012085,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03488154709339142,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.403696060180664,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.0363851860165596,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.368661403656006,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.035461440682411194,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.396518707275391,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03647598251700401,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.440667152404785,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.03667188808321953,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.3781352043151855,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034516457468271255,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.414007186889648,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.03536032512784004,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.42198371887207,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.037454646080732346,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.409174919128418,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03632275015115738,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3547682762146,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.038922134786844254,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.39993143081665,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.03562692925333977,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.409564971923828,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.033203717321157455,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.421782493591309,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03830202296376228,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.430092811584473,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.03835884481668472,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.379505634307861,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.037440430372953415,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.403937339782715,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03732272610068321,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.440683364868164,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03325957432389259,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.3949294090271,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.035377148538827896,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.400777816772461,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.03725374862551689,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.422494411468506,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03601110726594925,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.4008941650390625,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03193027526140213,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.400361061096191,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.034382354468107224,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.408535003662109,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.03747732564806938,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.412189483642578,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03641800582408905,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.381453037261963,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.0315779373049736,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.4169769287109375,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03583509847521782,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.385746955871582,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02574268914759159,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7456769943237305,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.588273823261261,
+ "eval_runtime": 59.9356,
+ "eval_samples_per_second": 40.744,
+ "eval_steps_per_second": 0.651,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013961605584642,
+ "grad_norm": 0.037516895681619644,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 4.33476448059082,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027923211169284,
+ "grad_norm": 0.048450272530317307,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 4.296454429626465,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041884816753926,
+ "grad_norm": 0.05368480458855629,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 4.314815998077393,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055846422338568,
+ "grad_norm": 0.05520009621977806,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 4.330798625946045,
+ "step": 1084
+ },
+ {
+ "epoch": 15.06980802792321,
+ "grad_norm": 0.05928800627589226,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 4.317663192749023,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083769633507853,
+ "grad_norm": 0.06315027177333832,
+ "learning_rate": 0.000491285979372924,
+ "loss": 4.329221725463867,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097731239092496,
+ "grad_norm": 0.06342213600873947,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 4.32546329498291,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111692844677139,
+ "grad_norm": 0.0667002946138382,
+ "learning_rate": 0.000490873338088198,
+ "loss": 4.309828758239746,
+ "step": 1088
+ },
+ {
+ "epoch": 15.12565445026178,
+ "grad_norm": 0.05264135077595711,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 4.307073593139648,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139616055846423,
+ "grad_norm": 0.051895879209041595,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 4.333526611328125,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153577661431065,
+ "grad_norm": 0.05508396774530411,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 4.330171585083008,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167539267015707,
+ "grad_norm": 0.057272691279649734,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 4.329268455505371,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181500872600349,
+ "grad_norm": 0.058956146240234375,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 4.341983795166016,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195462478184991,
+ "grad_norm": 0.0557059645652771,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 4.348274230957031,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209424083769633,
+ "grad_norm": 0.05661974847316742,
+ "learning_rate": 0.000489424334303338,
+ "loss": 4.347631454467773,
+ "step": 1095
+ },
+ {
+ "epoch": 15.223385689354275,
+ "grad_norm": 0.05338657647371292,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 4.334639072418213,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237347294938917,
+ "grad_norm": 0.05091148987412453,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 4.355465412139893,
+ "step": 1097
+ },
+ {
+ "epoch": 15.25130890052356,
+ "grad_norm": 0.04996965080499649,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 4.346457004547119,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265270506108202,
+ "grad_norm": 0.05573750659823418,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 4.346520900726318,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279232111692846,
+ "grad_norm": 0.05776917189359665,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 4.359007835388184,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293193717277488,
+ "grad_norm": 0.050820041447877884,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 4.331710338592529,
+ "step": 1101
+ },
+ {
+ "epoch": 15.30715532286213,
+ "grad_norm": 0.04441820830106735,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 4.3126420974731445,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321116928446772,
+ "grad_norm": 0.04582618921995163,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 4.312925338745117,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335078534031414,
+ "grad_norm": 0.043592359870672226,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 4.345674514770508,
+ "step": 1104
+ },
+ {
+ "epoch": 15.349040139616056,
+ "grad_norm": 0.04069037735462189,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 4.334713935852051,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363001745200698,
+ "grad_norm": 0.04083041846752167,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 4.358171463012695,
+ "step": 1106
+ },
+ {
+ "epoch": 15.37696335078534,
+ "grad_norm": 0.04119594022631645,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 4.317891597747803,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390924956369982,
+ "grad_norm": 0.042060643434524536,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 4.355810165405273,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404886561954624,
+ "grad_norm": 0.04721749201416969,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 4.353546619415283,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418848167539267,
+ "grad_norm": 0.04657730832695961,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 4.341683387756348,
+ "step": 1110
+ },
+ {
+ "epoch": 15.432809773123909,
+ "grad_norm": 0.04096715524792671,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 4.357192039489746,
+ "step": 1111
+ },
+ {
+ "epoch": 15.44677137870855,
+ "grad_norm": 0.03904155269265175,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 4.354775428771973,
+ "step": 1112
+ },
+ {
+ "epoch": 15.460732984293193,
+ "grad_norm": 0.03828521445393562,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 4.356499671936035,
+ "step": 1113
+ },
+ {
+ "epoch": 15.474694589877837,
+ "grad_norm": 0.03688685595989227,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 4.339485168457031,
+ "step": 1114
+ },
+ {
+ "epoch": 15.488656195462479,
+ "grad_norm": 0.03945618495345116,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 4.363500118255615,
+ "step": 1115
+ },
+ {
+ "epoch": 15.502617801047121,
+ "grad_norm": 0.03821462020277977,
+ "learning_rate": 0.000485033394934934,
+ "loss": 4.353529930114746,
+ "step": 1116
+ },
+ {
+ "epoch": 15.516579406631763,
+ "grad_norm": 0.036265499889850616,
+ "learning_rate": 0.000484822676912974,
+ "loss": 4.337663173675537,
+ "step": 1117
+ },
+ {
+ "epoch": 15.530541012216405,
+ "grad_norm": 0.03803930804133415,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 4.34660530090332,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544502617801047,
+ "grad_norm": 0.038425102829933167,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 4.373246192932129,
+ "step": 1119
+ },
+ {
+ "epoch": 15.55846422338569,
+ "grad_norm": 0.036852262914180756,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 4.331226825714111,
+ "step": 1120
+ },
+ {
+ "epoch": 15.572425828970331,
+ "grad_norm": 0.03700809180736542,
+ "learning_rate": 0.000483978341612154,
+ "loss": 4.375818729400635,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586387434554974,
+ "grad_norm": 0.03628664091229439,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 4.32705020904541,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600349040139616,
+ "grad_norm": 0.036831192672252655,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 4.337127208709717,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614310645724258,
+ "grad_norm": 0.03736541047692299,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 4.36502742767334,
+ "step": 1124
+ },
+ {
+ "epoch": 15.6282722513089,
+ "grad_norm": 0.03658579662442207,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 4.359844207763672,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642233856893542,
+ "grad_norm": 0.03688633441925049,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 4.3355793952941895,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656195462478184,
+ "grad_norm": 0.03477008268237114,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 4.333405494689941,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670157068062828,
+ "grad_norm": 0.03655015677213669,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 4.338570594787598,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68411867364747,
+ "grad_norm": 0.03664765506982803,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 4.352496147155762,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698080279232112,
+ "grad_norm": 0.037019114941358566,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 4.335735321044922,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712041884816754,
+ "grad_norm": 0.03949085995554924,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 4.342585563659668,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726003490401396,
+ "grad_norm": 0.04084350913763046,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 4.329622268676758,
+ "step": 1132
+ },
+ {
+ "epoch": 15.739965095986038,
+ "grad_norm": 0.038439832627773285,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 4.360318660736084,
+ "step": 1133
+ },
+ {
+ "epoch": 15.75392670157068,
+ "grad_norm": 0.03932987526059151,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 4.343446254730225,
+ "step": 1134
+ },
+ {
+ "epoch": 15.767888307155323,
+ "grad_norm": 0.041595373302698135,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 4.351495742797852,
+ "step": 1135
+ },
+ {
+ "epoch": 15.781849912739965,
+ "grad_norm": 0.03917490318417549,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 4.380537033081055,
+ "step": 1136
+ },
+ {
+ "epoch": 15.795811518324607,
+ "grad_norm": 0.038597021251916885,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 4.3527116775512695,
+ "step": 1137
+ },
+ {
+ "epoch": 15.809773123909249,
+ "grad_norm": 0.03651711344718933,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 4.369256496429443,
+ "step": 1138
+ },
+ {
+ "epoch": 15.823734729493891,
+ "grad_norm": 0.035118963569402695,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 4.350616455078125,
+ "step": 1139
+ },
+ {
+ "epoch": 15.837696335078533,
+ "grad_norm": 0.034945692867040634,
+ "learning_rate": 0.000479936075927473,
+ "loss": 4.365446090698242,
+ "step": 1140
+ },
+ {
+ "epoch": 15.851657940663177,
+ "grad_norm": 0.03519820421934128,
+ "learning_rate": 0.000479721889242305,
+ "loss": 4.357390403747559,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86561954624782,
+ "grad_norm": 0.03407098725438118,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 4.350573539733887,
+ "step": 1142
+ },
+ {
+ "epoch": 15.879581151832461,
+ "grad_norm": 0.036511123180389404,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 4.328500270843506,
+ "step": 1143
+ },
+ {
+ "epoch": 15.893542757417103,
+ "grad_norm": 0.039194293320178986,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 4.346471786499023,
+ "step": 1144
+ },
+ {
+ "epoch": 15.907504363001745,
+ "grad_norm": 0.03777911514043808,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 4.358243942260742,
+ "step": 1145
+ },
+ {
+ "epoch": 15.921465968586388,
+ "grad_norm": 0.0354655385017395,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 4.368768692016602,
+ "step": 1146
+ },
+ {
+ "epoch": 15.93542757417103,
+ "grad_norm": 0.03710418567061424,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 4.3482184410095215,
+ "step": 1147
+ },
+ {
+ "epoch": 15.949389179755672,
+ "grad_norm": 0.03545419126749039,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 4.360692501068115,
+ "step": 1148
+ },
+ {
+ "epoch": 15.963350785340314,
+ "grad_norm": 0.035963475704193115,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 4.3592119216918945,
+ "step": 1149
+ },
+ {
+ "epoch": 15.977312390924956,
+ "grad_norm": 0.03425297513604164,
+ "learning_rate": 0.000477787822648007,
+ "loss": 4.329504013061523,
+ "step": 1150
+ },
+ {
+ "epoch": 15.991273996509598,
+ "grad_norm": 0.03453255817294121,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 4.298142433166504,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.025766054168343544,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 2.7026405334472656,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.5888971090316772,
+ "eval_runtime": 58.9721,
+ "eval_samples_per_second": 41.409,
+ "eval_steps_per_second": 0.661,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013961605584644,
+ "grad_norm": 0.04088004305958748,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 4.268601417541504,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027923211169284,
+ "grad_norm": 0.05194641277194023,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 4.267483711242676,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041884816753928,
+ "grad_norm": 0.05488700792193413,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 4.2527570724487305,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05584642233857,
+ "grad_norm": 0.059292279183864594,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 4.283864974975586,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069808027923212,
+ "grad_norm": 0.0567227303981781,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 4.291148662567139,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083769633507853,
+ "grad_norm": 0.05582835525274277,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 4.250524520874023,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097731239092496,
+ "grad_norm": 0.055709999054670334,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 4.259716033935547,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111692844677137,
+ "grad_norm": 0.053324732929468155,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 4.247231483459473,
+ "step": 1160
+ },
+ {
+ "epoch": 16.12565445026178,
+ "grad_norm": 0.05367910861968994,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 4.263905048370361,
+ "step": 1161
+ },
+ {
+ "epoch": 16.13961605584642,
+ "grad_norm": 0.057094886898994446,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 4.266510009765625,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153577661431065,
+ "grad_norm": 0.05921962857246399,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 4.278281211853027,
+ "step": 1163
+ },
+ {
+ "epoch": 16.167539267015705,
+ "grad_norm": 0.060246821492910385,
+ "learning_rate": 0.000474756648870413,
+ "loss": 4.2848052978515625,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18150087260035,
+ "grad_norm": 0.05686050280928612,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 4.286434173583984,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195462478184993,
+ "grad_norm": 0.05207265168428421,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 4.246121406555176,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209424083769633,
+ "grad_norm": 0.0477963425219059,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 4.287411689758301,
+ "step": 1167
+ },
+ {
+ "epoch": 16.223385689354277,
+ "grad_norm": 0.05586806684732437,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 4.2721333503723145,
+ "step": 1168
+ },
+ {
+ "epoch": 16.237347294938917,
+ "grad_norm": 0.05427060276269913,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 4.2751641273498535,
+ "step": 1169
+ },
+ {
+ "epoch": 16.25130890052356,
+ "grad_norm": 0.0518609918653965,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 4.287444114685059,
+ "step": 1170
+ },
+ {
+ "epoch": 16.2652705061082,
+ "grad_norm": 0.053606316447257996,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 4.287125587463379,
+ "step": 1171
+ },
+ {
+ "epoch": 16.279232111692846,
+ "grad_norm": 0.05118813365697861,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 4.291893005371094,
+ "step": 1172
+ },
+ {
+ "epoch": 16.293193717277486,
+ "grad_norm": 0.04774491861462593,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 4.325250148773193,
+ "step": 1173
+ },
+ {
+ "epoch": 16.30715532286213,
+ "grad_norm": 0.04634870961308479,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 4.311764717102051,
+ "step": 1174
+ },
+ {
+ "epoch": 16.32111692844677,
+ "grad_norm": 0.04411927983164787,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 4.277068138122559,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335078534031414,
+ "grad_norm": 0.043819356709718704,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 4.309638977050781,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349040139616054,
+ "grad_norm": 0.0418904684484005,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 4.306234836578369,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363001745200698,
+ "grad_norm": 0.04327298328280449,
+ "learning_rate": 0.000471698352726896,
+ "loss": 4.296874046325684,
+ "step": 1178
+ },
+ {
+ "epoch": 16.376963350785342,
+ "grad_norm": 0.045296184718608856,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 4.288957595825195,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390924956369982,
+ "grad_norm": 0.04447923228144646,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 4.313834190368652,
+ "step": 1180
+ },
+ {
+ "epoch": 16.404886561954626,
+ "grad_norm": 0.04548390209674835,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 4.284562587738037,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418848167539267,
+ "grad_norm": 0.04567817226052284,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 4.295994758605957,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43280977312391,
+ "grad_norm": 0.04277439042925835,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 4.262106895446777,
+ "step": 1183
+ },
+ {
+ "epoch": 16.44677137870855,
+ "grad_norm": 0.043597545474767685,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 4.294536590576172,
+ "step": 1184
+ },
+ {
+ "epoch": 16.460732984293195,
+ "grad_norm": 0.04424033313989639,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 4.297497749328613,
+ "step": 1185
+ },
+ {
+ "epoch": 16.474694589877835,
+ "grad_norm": 0.04354134574532509,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 4.321508884429932,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48865619546248,
+ "grad_norm": 0.04241178184747696,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 4.322726249694824,
+ "step": 1187
+ },
+ {
+ "epoch": 16.50261780104712,
+ "grad_norm": 0.039317622780799866,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 4.311155319213867,
+ "step": 1188
+ },
+ {
+ "epoch": 16.516579406631763,
+ "grad_norm": 0.04126536473631859,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 4.306636810302734,
+ "step": 1189
+ },
+ {
+ "epoch": 16.530541012216403,
+ "grad_norm": 0.040116794407367706,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 4.314671516418457,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544502617801047,
+ "grad_norm": 0.04045999422669411,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 4.301012992858887,
+ "step": 1191
+ },
+ {
+ "epoch": 16.55846422338569,
+ "grad_norm": 0.04013422504067421,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 4.29913330078125,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57242582897033,
+ "grad_norm": 0.03878502920269966,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 4.351600646972656,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586387434554975,
+ "grad_norm": 0.038935888558626175,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 4.27667236328125,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600349040139616,
+ "grad_norm": 0.04055892676115036,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 4.29019021987915,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61431064572426,
+ "grad_norm": 0.03959144279360771,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 4.327725887298584,
+ "step": 1196
+ },
+ {
+ "epoch": 16.6282722513089,
+ "grad_norm": 0.039878539741039276,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 4.297952651977539,
+ "step": 1197
+ },
+ {
+ "epoch": 16.642233856893544,
+ "grad_norm": 0.041209928691387177,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 4.313030242919922,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656195462478184,
+ "grad_norm": 0.04346393793821335,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 4.274901390075684,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670157068062828,
+ "grad_norm": 0.04090269282460213,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 4.330657958984375,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68411867364747,
+ "grad_norm": 0.04004506766796112,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 4.323106288909912,
+ "step": 1201
+ },
+ {
+ "epoch": 16.698080279232112,
+ "grad_norm": 0.041541460901498795,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 4.280947685241699,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712041884816752,
+ "grad_norm": 0.04033653065562248,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 4.341434955596924,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726003490401396,
+ "grad_norm": 0.04104280099272728,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 4.2931671142578125,
+ "step": 1204
+ },
+ {
+ "epoch": 16.739965095986037,
+ "grad_norm": 0.039398156106472015,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 4.316840648651123,
+ "step": 1205
+ },
+ {
+ "epoch": 16.75392670157068,
+ "grad_norm": 0.03980560600757599,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 4.325345039367676,
+ "step": 1206
+ },
+ {
+ "epoch": 16.767888307155324,
+ "grad_norm": 0.040656186640262604,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 4.32675838470459,
+ "step": 1207
+ },
+ {
+ "epoch": 16.781849912739965,
+ "grad_norm": 0.03775986284017563,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 4.314423561096191,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79581151832461,
+ "grad_norm": 0.04043284431099892,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 4.344202041625977,
+ "step": 1209
+ },
+ {
+ "epoch": 16.80977312390925,
+ "grad_norm": 0.04074503108859062,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 4.350123405456543,
+ "step": 1210
+ },
+ {
+ "epoch": 16.823734729493893,
+ "grad_norm": 0.04117420315742493,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 4.350917816162109,
+ "step": 1211
+ },
+ {
+ "epoch": 16.837696335078533,
+ "grad_norm": 0.03799610957503319,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 4.289987564086914,
+ "step": 1212
+ },
+ {
+ "epoch": 16.851657940663177,
+ "grad_norm": 0.04071265831589699,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 4.306744575500488,
+ "step": 1213
+ },
+ {
+ "epoch": 16.865619546247817,
+ "grad_norm": 0.04109715297818184,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 4.318753242492676,
+ "step": 1214
+ },
+ {
+ "epoch": 16.87958115183246,
+ "grad_norm": 0.03910145163536072,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 4.281325340270996,
+ "step": 1215
+ },
+ {
+ "epoch": 16.8935427574171,
+ "grad_norm": 0.03752733767032623,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 4.3231048583984375,
+ "step": 1216
+ },
+ {
+ "epoch": 16.907504363001745,
+ "grad_norm": 0.03752103075385094,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 4.351336479187012,
+ "step": 1217
+ },
+ {
+ "epoch": 16.921465968586386,
+ "grad_norm": 0.03675422817468643,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 4.282256126403809,
+ "step": 1218
+ },
+ {
+ "epoch": 16.93542757417103,
+ "grad_norm": 0.03850258141756058,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 4.322597503662109,
+ "step": 1219
+ },
+ {
+ "epoch": 16.949389179755673,
+ "grad_norm": 0.03692392632365227,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 4.320333957672119,
+ "step": 1220
+ },
+ {
+ "epoch": 16.963350785340314,
+ "grad_norm": 0.035366240888834,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 4.31341552734375,
+ "step": 1221
+ },
+ {
+ "epoch": 16.977312390924958,
+ "grad_norm": 0.037120968103408813,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 4.321141242980957,
+ "step": 1222
+ },
+ {
+ "epoch": 16.991273996509598,
+ "grad_norm": 0.03721754252910614,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 4.305757522583008,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.026698043569922447,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 2.716372489929199,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.5894472599029541,
+ "eval_runtime": 58.927,
+ "eval_samples_per_second": 41.441,
+ "eval_steps_per_second": 0.662,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013961605584644,
+ "grad_norm": 0.044532276690006256,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 4.243244171142578,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027923211169284,
+ "grad_norm": 0.050529688596725464,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 4.221382141113281,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041884816753928,
+ "grad_norm": 0.05690870061516762,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 4.240285873413086,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05584642233857,
+ "grad_norm": 0.054803814738988876,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 4.213047027587891,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069808027923212,
+ "grad_norm": 0.051825862377882004,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 4.226880073547363,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083769633507853,
+ "grad_norm": 0.049802035093307495,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 4.25303840637207,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097731239092496,
+ "grad_norm": 0.05290328338742256,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 4.199422836303711,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111692844677137,
+ "grad_norm": 0.0531814880669117,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 4.201648712158203,
+ "step": 1232
+ },
+ {
+ "epoch": 17.12565445026178,
+ "grad_norm": 0.048972588032484055,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 4.207578182220459,
+ "step": 1233
+ },
+ {
+ "epoch": 17.13961605584642,
+ "grad_norm": 0.051345713436603546,
+ "learning_rate": 0.000459203665939781,
+ "loss": 4.206205368041992,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153577661431065,
+ "grad_norm": 0.059012677520513535,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 4.228360176086426,
+ "step": 1235
+ },
+ {
+ "epoch": 17.167539267015705,
+ "grad_norm": 0.05984492972493172,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 4.234362602233887,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18150087260035,
+ "grad_norm": 0.060886185616254807,
+ "learning_rate": 0.000458522886595554,
+ "loss": 4.253525257110596,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195462478184993,
+ "grad_norm": 0.06476911902427673,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 4.249889373779297,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209424083769633,
+ "grad_norm": 0.059773124754428864,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 4.225249290466309,
+ "step": 1239
+ },
+ {
+ "epoch": 17.223385689354277,
+ "grad_norm": 0.06063077598810196,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 4.25465202331543,
+ "step": 1240
+ },
+ {
+ "epoch": 17.237347294938917,
+ "grad_norm": 0.059922896325588226,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 4.252101898193359,
+ "step": 1241
+ },
+ {
+ "epoch": 17.25130890052356,
+ "grad_norm": 0.05849695950746536,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 4.268218040466309,
+ "step": 1242
+ },
+ {
+ "epoch": 17.2652705061082,
+ "grad_norm": 0.05769657716155052,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 4.2395782470703125,
+ "step": 1243
+ },
+ {
+ "epoch": 17.279232111692846,
+ "grad_norm": 0.058189719915390015,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 4.257606506347656,
+ "step": 1244
+ },
+ {
+ "epoch": 17.293193717277486,
+ "grad_norm": 0.059679049998521805,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 4.278588771820068,
+ "step": 1245
+ },
+ {
+ "epoch": 17.30715532286213,
+ "grad_norm": 0.0579286590218544,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 4.25511360168457,
+ "step": 1246
+ },
+ {
+ "epoch": 17.32111692844677,
+ "grad_norm": 0.05269965901970863,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 4.241326332092285,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335078534031414,
+ "grad_norm": 0.04828358441591263,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 4.243182182312012,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349040139616054,
+ "grad_norm": 0.050172045826911926,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 4.295144081115723,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363001745200698,
+ "grad_norm": 0.047297943383455276,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 4.250682830810547,
+ "step": 1250
+ },
+ {
+ "epoch": 17.376963350785342,
+ "grad_norm": 0.04574039950966835,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 4.269444465637207,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390924956369982,
+ "grad_norm": 0.042310427874326706,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 4.255361557006836,
+ "step": 1252
+ },
+ {
+ "epoch": 17.404886561954626,
+ "grad_norm": 0.04268861189484596,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 4.252292156219482,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418848167539267,
+ "grad_norm": 0.0424286350607872,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 4.24439811706543,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43280977312391,
+ "grad_norm": 0.042049869894981384,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 4.236266136169434,
+ "step": 1255
+ },
+ {
+ "epoch": 17.44677137870855,
+ "grad_norm": 0.04185241088271141,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 4.268826484680176,
+ "step": 1256
+ },
+ {
+ "epoch": 17.460732984293195,
+ "grad_norm": 0.04423609375953674,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 4.240240573883057,
+ "step": 1257
+ },
+ {
+ "epoch": 17.474694589877835,
+ "grad_norm": 0.041972566395998,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 4.269453048706055,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48865619546248,
+ "grad_norm": 0.04493343085050583,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 4.270544052124023,
+ "step": 1259
+ },
+ {
+ "epoch": 17.50261780104712,
+ "grad_norm": 0.04502921551465988,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 4.271880149841309,
+ "step": 1260
+ },
+ {
+ "epoch": 17.516579406631763,
+ "grad_norm": 0.044518254697322845,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 4.249786376953125,
+ "step": 1261
+ },
+ {
+ "epoch": 17.530541012216403,
+ "grad_norm": 0.044482357800006866,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 4.300779819488525,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544502617801047,
+ "grad_norm": 0.0424671471118927,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 4.248597621917725,
+ "step": 1263
+ },
+ {
+ "epoch": 17.55846422338569,
+ "grad_norm": 0.04136480763554573,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 4.276763916015625,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57242582897033,
+ "grad_norm": 0.04094843938946724,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 4.263628005981445,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586387434554975,
+ "grad_norm": 0.04415388032793999,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 4.257946014404297,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600349040139616,
+ "grad_norm": 0.04177384823560715,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 4.289669036865234,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61431064572426,
+ "grad_norm": 0.040863484144210815,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 4.320903778076172,
+ "step": 1268
+ },
+ {
+ "epoch": 17.6282722513089,
+ "grad_norm": 0.04025879129767418,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 4.268717288970947,
+ "step": 1269
+ },
+ {
+ "epoch": 17.642233856893544,
+ "grad_norm": 0.04161325842142105,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 4.275268077850342,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656195462478184,
+ "grad_norm": 0.04260355979204178,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 4.263361930847168,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670157068062828,
+ "grad_norm": 0.043366871774196625,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 4.254581928253174,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68411867364747,
+ "grad_norm": 0.04299113526940346,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 4.268922328948975,
+ "step": 1273
+ },
+ {
+ "epoch": 17.698080279232112,
+ "grad_norm": 0.041865330189466476,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 4.290097236633301,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712041884816752,
+ "grad_norm": 0.04515853524208069,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 4.261376857757568,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726003490401396,
+ "grad_norm": 0.046030085533857346,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 4.290409564971924,
+ "step": 1276
+ },
+ {
+ "epoch": 17.739965095986037,
+ "grad_norm": 0.045034267008304596,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 4.28201961517334,
+ "step": 1277
+ },
+ {
+ "epoch": 17.75392670157068,
+ "grad_norm": 0.04293077066540718,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 4.282614707946777,
+ "step": 1278
+ },
+ {
+ "epoch": 17.767888307155324,
+ "grad_norm": 0.043866172432899475,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 4.2777557373046875,
+ "step": 1279
+ },
+ {
+ "epoch": 17.781849912739965,
+ "grad_norm": 0.04693951830267906,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 4.272597312927246,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79581151832461,
+ "grad_norm": 0.04072774201631546,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 4.284282684326172,
+ "step": 1281
+ },
+ {
+ "epoch": 17.80977312390925,
+ "grad_norm": 0.04241872951388359,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 4.254406452178955,
+ "step": 1282
+ },
+ {
+ "epoch": 17.823734729493893,
+ "grad_norm": 0.04614942893385887,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 4.297187805175781,
+ "step": 1283
+ },
+ {
+ "epoch": 17.837696335078533,
+ "grad_norm": 0.04326155036687851,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 4.292263984680176,
+ "step": 1284
+ },
+ {
+ "epoch": 17.851657940663177,
+ "grad_norm": 0.04053817689418793,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 4.2925920486450195,
+ "step": 1285
+ },
+ {
+ "epoch": 17.865619546247817,
+ "grad_norm": 0.040867388248443604,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 4.276965141296387,
+ "step": 1286
+ },
+ {
+ "epoch": 17.87958115183246,
+ "grad_norm": 0.041000183671712875,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 4.257638454437256,
+ "step": 1287
+ },
+ {
+ "epoch": 17.8935427574171,
+ "grad_norm": 0.04014866054058075,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 4.2734479904174805,
+ "step": 1288
+ },
+ {
+ "epoch": 17.907504363001745,
+ "grad_norm": 0.041237372905015945,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 4.263461112976074,
+ "step": 1289
+ },
+ {
+ "epoch": 17.921465968586386,
+ "grad_norm": 0.039109110832214355,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 4.295286655426025,
+ "step": 1290
+ },
+ {
+ "epoch": 17.93542757417103,
+ "grad_norm": 0.03911379352211952,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 4.254448413848877,
+ "step": 1291
+ },
+ {
+ "epoch": 17.949389179755673,
+ "grad_norm": 0.03997721150517464,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 4.284714698791504,
+ "step": 1292
+ },
+ {
+ "epoch": 17.963350785340314,
+ "grad_norm": 0.04011402651667595,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 4.278304100036621,
+ "step": 1293
+ },
+ {
+ "epoch": 17.977312390924958,
+ "grad_norm": 0.04040595516562462,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 4.288183212280273,
+ "step": 1294
+ },
+ {
+ "epoch": 17.991273996509598,
+ "grad_norm": 0.038857635110616684,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 4.274664402008057,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.028110472485423088,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 2.6625723838806152,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.5912551283836365,
+ "eval_runtime": 60.273,
+ "eval_samples_per_second": 40.516,
+ "eval_steps_per_second": 0.647,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013961605584644,
+ "grad_norm": 0.041634928435087204,
+ "learning_rate": 0.000444677103403854,
+ "loss": 4.181324005126953,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027923211169284,
+ "grad_norm": 0.051596369594335556,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 4.204722881317139,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041884816753928,
+ "grad_norm": 0.058580417186021805,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 4.207996845245361,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05584642233857,
+ "grad_norm": 0.05986522138118744,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 4.184011459350586,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069808027923212,
+ "grad_norm": 0.051045969128608704,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 4.148118019104004,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083769633507853,
+ "grad_norm": 0.045477766543626785,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 4.207469940185547,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097731239092496,
+ "grad_norm": 0.0488731674849987,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 4.201406478881836,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111692844677137,
+ "grad_norm": 0.05013517662882805,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 4.192239761352539,
+ "step": 1304
+ },
+ {
+ "epoch": 18.12565445026178,
+ "grad_norm": 0.05091651529073715,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 4.188252925872803,
+ "step": 1305
+ },
+ {
+ "epoch": 18.13961605584642,
+ "grad_norm": 0.051825884729623795,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 4.157929420471191,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153577661431065,
+ "grad_norm": 0.049625713378190994,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 4.208776473999023,
+ "step": 1307
+ },
+ {
+ "epoch": 18.167539267015705,
+ "grad_norm": 0.04816461354494095,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 4.180496692657471,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18150087260035,
+ "grad_norm": 0.04720531404018402,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 4.178536415100098,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195462478184993,
+ "grad_norm": 0.0475568026304245,
+ "learning_rate": 0.000441621628895411,
+ "loss": 4.203497886657715,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209424083769633,
+ "grad_norm": 0.04613947868347168,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 4.226914405822754,
+ "step": 1311
+ },
+ {
+ "epoch": 18.223385689354277,
+ "grad_norm": 0.04585929960012436,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 4.175119400024414,
+ "step": 1312
+ },
+ {
+ "epoch": 18.237347294938917,
+ "grad_norm": 0.045777060091495514,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 4.181760311126709,
+ "step": 1313
+ },
+ {
+ "epoch": 18.25130890052356,
+ "grad_norm": 0.04583602771162987,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 4.179868698120117,
+ "step": 1314
+ },
+ {
+ "epoch": 18.2652705061082,
+ "grad_norm": 0.04706503078341484,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 4.203325271606445,
+ "step": 1315
+ },
+ {
+ "epoch": 18.279232111692846,
+ "grad_norm": 0.045122336596250534,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 4.187152862548828,
+ "step": 1316
+ },
+ {
+ "epoch": 18.293193717277486,
+ "grad_norm": 0.04495825991034508,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 4.2080535888671875,
+ "step": 1317
+ },
+ {
+ "epoch": 18.30715532286213,
+ "grad_norm": 0.04767543077468872,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 4.207225322723389,
+ "step": 1318
+ },
+ {
+ "epoch": 18.32111692844677,
+ "grad_norm": 0.053396135568618774,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 4.227478981018066,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335078534031414,
+ "grad_norm": 0.0512089841067791,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 4.1904754638671875,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349040139616054,
+ "grad_norm": 0.04967977851629257,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 4.2079925537109375,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363001745200698,
+ "grad_norm": 0.04800970479846001,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 4.206037998199463,
+ "step": 1322
+ },
+ {
+ "epoch": 18.376963350785342,
+ "grad_norm": 0.04993556812405586,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 4.210054397583008,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390924956369982,
+ "grad_norm": 0.0480492040514946,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 4.229897499084473,
+ "step": 1324
+ },
+ {
+ "epoch": 18.404886561954626,
+ "grad_norm": 0.048699282109737396,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 4.17500114440918,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418848167539267,
+ "grad_norm": 0.04681490734219551,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 4.196526527404785,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43280977312391,
+ "grad_norm": 0.04674449935555458,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 4.246790409088135,
+ "step": 1327
+ },
+ {
+ "epoch": 18.44677137870855,
+ "grad_norm": 0.04650607705116272,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 4.225878715515137,
+ "step": 1328
+ },
+ {
+ "epoch": 18.460732984293195,
+ "grad_norm": 0.047200269997119904,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 4.20685338973999,
+ "step": 1329
+ },
+ {
+ "epoch": 18.474694589877835,
+ "grad_norm": 0.04322655498981476,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 4.212920665740967,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48865619546248,
+ "grad_norm": 0.04489128664135933,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 4.266653060913086,
+ "step": 1331
+ },
+ {
+ "epoch": 18.50261780104712,
+ "grad_norm": 0.048838935792446136,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 4.2364501953125,
+ "step": 1332
+ },
+ {
+ "epoch": 18.516579406631763,
+ "grad_norm": 0.04893286153674126,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 4.202394485473633,
+ "step": 1333
+ },
+ {
+ "epoch": 18.530541012216403,
+ "grad_norm": 0.046900101006031036,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 4.209832191467285,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544502617801047,
+ "grad_norm": 0.044056762009859085,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 4.229903221130371,
+ "step": 1335
+ },
+ {
+ "epoch": 18.55846422338569,
+ "grad_norm": 0.04681328684091568,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 4.22783899307251,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57242582897033,
+ "grad_norm": 0.046087440103292465,
+ "learning_rate": 0.000435215814386134,
+ "loss": 4.2078986167907715,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586387434554975,
+ "grad_norm": 0.04363090172410011,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 4.218686103820801,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600349040139616,
+ "grad_norm": 0.046255532652139664,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 4.2475128173828125,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61431064572426,
+ "grad_norm": 0.045545510947704315,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 4.233272552490234,
+ "step": 1340
+ },
+ {
+ "epoch": 18.6282722513089,
+ "grad_norm": 0.045259326696395874,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 4.252043724060059,
+ "step": 1341
+ },
+ {
+ "epoch": 18.642233856893544,
+ "grad_norm": 0.042144861072301865,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 4.247762203216553,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656195462478184,
+ "grad_norm": 0.04298558086156845,
+ "learning_rate": 0.000433781621332752,
+ "loss": 4.222329616546631,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670157068062828,
+ "grad_norm": 0.044070884585380554,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 4.222488880157471,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68411867364747,
+ "grad_norm": 0.044758375734090805,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 4.238160133361816,
+ "step": 1345
+ },
+ {
+ "epoch": 18.698080279232112,
+ "grad_norm": 0.04324857518076897,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 4.220003128051758,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712041884816752,
+ "grad_norm": 0.04349793121218681,
+ "learning_rate": 0.000432823382695327,
+ "loss": 4.251608848571777,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726003490401396,
+ "grad_norm": 0.04483242705464363,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 4.243319511413574,
+ "step": 1348
+ },
+ {
+ "epoch": 18.739965095986037,
+ "grad_norm": 0.04391994699835777,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 4.237479209899902,
+ "step": 1349
+ },
+ {
+ "epoch": 18.75392670157068,
+ "grad_norm": 0.042584922164678574,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 4.2309980392456055,
+ "step": 1350
+ },
+ {
+ "epoch": 18.767888307155324,
+ "grad_norm": 0.04363991320133209,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 4.216980934143066,
+ "step": 1351
+ },
+ {
+ "epoch": 18.781849912739965,
+ "grad_norm": 0.04662707448005676,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 4.257450103759766,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79581151832461,
+ "grad_norm": 0.0462074875831604,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 4.222329139709473,
+ "step": 1353
+ },
+ {
+ "epoch": 18.80977312390925,
+ "grad_norm": 0.043543554842472076,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 4.231960773468018,
+ "step": 1354
+ },
+ {
+ "epoch": 18.823734729493893,
+ "grad_norm": 0.04444567486643791,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 4.225714683532715,
+ "step": 1355
+ },
+ {
+ "epoch": 18.837696335078533,
+ "grad_norm": 0.04507977142930031,
+ "learning_rate": 0.000430661245733797,
+ "loss": 4.234722137451172,
+ "step": 1356
+ },
+ {
+ "epoch": 18.851657940663177,
+ "grad_norm": 0.043786536902189255,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 4.274998664855957,
+ "step": 1357
+ },
+ {
+ "epoch": 18.865619546247817,
+ "grad_norm": 0.04188325256109238,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 4.260570526123047,
+ "step": 1358
+ },
+ {
+ "epoch": 18.87958115183246,
+ "grad_norm": 0.0439927875995636,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 4.250492572784424,
+ "step": 1359
+ },
+ {
+ "epoch": 18.8935427574171,
+ "grad_norm": 0.04520147293806076,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 4.250911712646484,
+ "step": 1360
+ },
+ {
+ "epoch": 18.907504363001745,
+ "grad_norm": 0.04442249983549118,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 4.255832672119141,
+ "step": 1361
+ },
+ {
+ "epoch": 18.921465968586386,
+ "grad_norm": 0.044756144285202026,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 4.226641654968262,
+ "step": 1362
+ },
+ {
+ "epoch": 18.93542757417103,
+ "grad_norm": 0.04647127911448479,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 4.2110915184021,
+ "step": 1363
+ },
+ {
+ "epoch": 18.949389179755673,
+ "grad_norm": 0.045986004173755646,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 4.243263244628906,
+ "step": 1364
+ },
+ {
+ "epoch": 18.963350785340314,
+ "grad_norm": 0.04533308371901512,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 4.253750801086426,
+ "step": 1365
+ },
+ {
+ "epoch": 18.977312390924958,
+ "grad_norm": 0.04307471215724945,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 4.233024597167969,
+ "step": 1366
+ },
+ {
+ "epoch": 18.991273996509598,
+ "grad_norm": 0.0416078194975853,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 4.264965534210205,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.030926328152418137,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 2.651331901550293,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.592528223991394,
+ "eval_runtime": 59.67,
+ "eval_samples_per_second": 40.925,
+ "eval_steps_per_second": 0.654,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013961605584644,
+ "grad_norm": 0.04524356126785278,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 4.153796672821045,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027923211169284,
+ "grad_norm": 0.05105067789554596,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 4.154603004455566,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041884816753928,
+ "grad_norm": 0.052823506295681,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 4.1599273681640625,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05584642233857,
+ "grad_norm": 0.05688423290848732,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 4.124622344970703,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069808027923212,
+ "grad_norm": 0.053178343921899796,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 4.138644218444824,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083769633507853,
+ "grad_norm": 0.05473732203245163,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 4.138207912445068,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097731239092496,
+ "grad_norm": 0.055207233875989914,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 4.155735015869141,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111692844677137,
+ "grad_norm": 0.05484927073121071,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 4.140205383300781,
+ "step": 1376
+ },
+ {
+ "epoch": 19.12565445026178,
+ "grad_norm": 0.05589596927165985,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 4.160840034484863,
+ "step": 1377
+ },
+ {
+ "epoch": 19.13961605584642,
+ "grad_norm": 0.05222902446985245,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 4.154781341552734,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153577661431065,
+ "grad_norm": 0.051824700087308884,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 4.134243488311768,
+ "step": 1379
+ },
+ {
+ "epoch": 19.167539267015705,
+ "grad_norm": 0.0517590269446373,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 4.125945568084717,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18150087260035,
+ "grad_norm": 0.05399726703763008,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 4.146993160247803,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195462478184993,
+ "grad_norm": 0.054430052638053894,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 4.1310906410217285,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209424083769633,
+ "grad_norm": 0.0551968477666378,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 4.170640468597412,
+ "step": 1383
+ },
+ {
+ "epoch": 19.223385689354277,
+ "grad_norm": 0.06050962209701538,
+ "learning_rate": 0.000423881957237287,
+ "loss": 4.166876316070557,
+ "step": 1384
+ },
+ {
+ "epoch": 19.237347294938917,
+ "grad_norm": 0.055587392300367355,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 4.150339603424072,
+ "step": 1385
+ },
+ {
+ "epoch": 19.25130890052356,
+ "grad_norm": 0.05193328112363815,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 4.171754360198975,
+ "step": 1386
+ },
+ {
+ "epoch": 19.2652705061082,
+ "grad_norm": 0.05272402986884117,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 4.165149688720703,
+ "step": 1387
+ },
+ {
+ "epoch": 19.279232111692846,
+ "grad_norm": 0.056283943355083466,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 4.193273544311523,
+ "step": 1388
+ },
+ {
+ "epoch": 19.293193717277486,
+ "grad_norm": 0.05825977027416229,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 4.192652225494385,
+ "step": 1389
+ },
+ {
+ "epoch": 19.30715532286213,
+ "grad_norm": 0.05539379641413689,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 4.187044143676758,
+ "step": 1390
+ },
+ {
+ "epoch": 19.32111692844677,
+ "grad_norm": 0.05300498753786087,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 4.168220043182373,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335078534031414,
+ "grad_norm": 0.05425756797194481,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 4.185193061828613,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349040139616054,
+ "grad_norm": 0.05527182295918465,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 4.1618332862854,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363001745200698,
+ "grad_norm": 0.054422613233327866,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 4.195295810699463,
+ "step": 1394
+ },
+ {
+ "epoch": 19.376963350785342,
+ "grad_norm": 0.052717432379722595,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 4.143290996551514,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390924956369982,
+ "grad_norm": 0.05154350399971008,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 4.115989685058594,
+ "step": 1396
+ },
+ {
+ "epoch": 19.404886561954626,
+ "grad_norm": 0.0489451140165329,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 4.169506072998047,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418848167539267,
+ "grad_norm": 0.049093786627054214,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 4.164122104644775,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43280977312391,
+ "grad_norm": 0.04690805822610855,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 4.186679840087891,
+ "step": 1399
+ },
+ {
+ "epoch": 19.44677137870855,
+ "grad_norm": 0.046708665788173676,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 4.184325695037842,
+ "step": 1400
+ },
+ {
+ "epoch": 19.460732984293195,
+ "grad_norm": 0.04617927595973015,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 4.201678276062012,
+ "step": 1401
+ },
+ {
+ "epoch": 19.474694589877835,
+ "grad_norm": 0.05017530545592308,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 4.18928337097168,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48865619546248,
+ "grad_norm": 0.05197977274656296,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 4.188801288604736,
+ "step": 1403
+ },
+ {
+ "epoch": 19.50261780104712,
+ "grad_norm": 0.049691297113895416,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 4.209467887878418,
+ "step": 1404
+ },
+ {
+ "epoch": 19.516579406631763,
+ "grad_norm": 0.049709953367710114,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 4.19733190536499,
+ "step": 1405
+ },
+ {
+ "epoch": 19.530541012216403,
+ "grad_norm": 0.05274137482047081,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 4.221798896789551,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544502617801047,
+ "grad_norm": 0.05015392601490021,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 4.210565567016602,
+ "step": 1407
+ },
+ {
+ "epoch": 19.55846422338569,
+ "grad_norm": 0.05023058131337166,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 4.184835910797119,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57242582897033,
+ "grad_norm": 0.0479239858686924,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 4.199846267700195,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586387434554975,
+ "grad_norm": 0.0481228232383728,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 4.185157775878906,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600349040139616,
+ "grad_norm": 0.04909738525748253,
+ "learning_rate": 0.000417272427439646,
+ "loss": 4.207550525665283,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61431064572426,
+ "grad_norm": 0.0475567989051342,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 4.213225841522217,
+ "step": 1412
+ },
+ {
+ "epoch": 19.6282722513089,
+ "grad_norm": 0.048073310405015945,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 4.212669372558594,
+ "step": 1413
+ },
+ {
+ "epoch": 19.642233856893544,
+ "grad_norm": 0.04523796588182449,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 4.1950860023498535,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656195462478184,
+ "grad_norm": 0.04359745234251022,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 4.2136640548706055,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670157068062828,
+ "grad_norm": 0.047282423824071884,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 4.209680557250977,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68411867364747,
+ "grad_norm": 0.0474061481654644,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 4.188494682312012,
+ "step": 1417
+ },
+ {
+ "epoch": 19.698080279232112,
+ "grad_norm": 0.04402102902531624,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 4.207236289978027,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712041884816752,
+ "grad_norm": 0.04398655146360397,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 4.210899353027344,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726003490401396,
+ "grad_norm": 0.045767925679683685,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 4.171738624572754,
+ "step": 1420
+ },
+ {
+ "epoch": 19.739965095986037,
+ "grad_norm": 0.04498986527323723,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 4.222414970397949,
+ "step": 1421
+ },
+ {
+ "epoch": 19.75392670157068,
+ "grad_norm": 0.04282551258802414,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 4.202138900756836,
+ "step": 1422
+ },
+ {
+ "epoch": 19.767888307155324,
+ "grad_norm": 0.04392262175679207,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 4.184439659118652,
+ "step": 1423
+ },
+ {
+ "epoch": 19.781849912739965,
+ "grad_norm": 0.04343154653906822,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 4.204674243927002,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79581151832461,
+ "grad_norm": 0.04496419429779053,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 4.23316764831543,
+ "step": 1425
+ },
+ {
+ "epoch": 19.80977312390925,
+ "grad_norm": 0.04488370940089226,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 4.201659202575684,
+ "step": 1426
+ },
+ {
+ "epoch": 19.823734729493893,
+ "grad_norm": 0.045598775148391724,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 4.187412261962891,
+ "step": 1427
+ },
+ {
+ "epoch": 19.837696335078533,
+ "grad_norm": 0.043846048414707184,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 4.215444564819336,
+ "step": 1428
+ },
+ {
+ "epoch": 19.851657940663177,
+ "grad_norm": 0.043062254786491394,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 4.2074055671691895,
+ "step": 1429
+ },
+ {
+ "epoch": 19.865619546247817,
+ "grad_norm": 0.04544302448630333,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 4.184919357299805,
+ "step": 1430
+ },
+ {
+ "epoch": 19.87958115183246,
+ "grad_norm": 0.046144865453243256,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 4.2175750732421875,
+ "step": 1431
+ },
+ {
+ "epoch": 19.8935427574171,
+ "grad_norm": 0.04465445503592491,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 4.1851115226745605,
+ "step": 1432
+ },
+ {
+ "epoch": 19.907504363001745,
+ "grad_norm": 0.045913565903902054,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 4.2226152420043945,
+ "step": 1433
+ },
+ {
+ "epoch": 19.921465968586386,
+ "grad_norm": 0.04401300847530365,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 4.216772556304932,
+ "step": 1434
+ },
+ {
+ "epoch": 19.93542757417103,
+ "grad_norm": 0.042624711990356445,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 4.2133684158325195,
+ "step": 1435
+ },
+ {
+ "epoch": 19.949389179755673,
+ "grad_norm": 0.04533256217837334,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 4.218449592590332,
+ "step": 1436
+ },
+ {
+ "epoch": 19.963350785340314,
+ "grad_norm": 0.04764309898018837,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 4.191402912139893,
+ "step": 1437
+ },
+ {
+ "epoch": 19.977312390924958,
+ "grad_norm": 0.04689318686723709,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 4.2113037109375,
+ "step": 1438
+ },
+ {
+ "epoch": 19.991273996509598,
+ "grad_norm": 0.04479913413524628,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 4.194989204406738,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.03151386231184006,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 2.6285438537597656,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 0.5938133597373962,
+ "eval_runtime": 58.9656,
+ "eval_samples_per_second": 41.414,
+ "eval_steps_per_second": 0.661,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013961605584644,
+ "grad_norm": 0.04884875938296318,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 4.1257734298706055,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027923211169284,
+ "grad_norm": 0.05968257039785385,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 4.091641426086426,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041884816753928,
+ "grad_norm": 0.06437838077545166,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 4.122465133666992,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05584642233857,
+ "grad_norm": 0.057831212878227234,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 4.129654884338379,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069808027923212,
+ "grad_norm": 0.049407217651605606,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 4.116935729980469,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083769633507853,
+ "grad_norm": 0.051957421004772186,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 4.125039577484131,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097731239092496,
+ "grad_norm": 0.05425718426704407,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 4.113485336303711,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111692844677137,
+ "grad_norm": 0.055318959057331085,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 4.119422912597656,
+ "step": 1448
+ },
+ {
+ "epoch": 20.12565445026178,
+ "grad_norm": 0.0548059344291687,
+ "learning_rate": 0.000407857329622967,
+ "loss": 4.097405910491943,
+ "step": 1449
+ },
+ {
+ "epoch": 20.13961605584642,
+ "grad_norm": 0.051046356558799744,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 4.115051746368408,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153577661431065,
+ "grad_norm": 0.05079492926597595,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 4.126070499420166,
+ "step": 1451
+ },
+ {
+ "epoch": 20.167539267015705,
+ "grad_norm": 0.04777265712618828,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 4.118952751159668,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18150087260035,
+ "grad_norm": 0.04895703122019768,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 4.095477104187012,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195462478184993,
+ "grad_norm": 0.051196496933698654,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 4.116866111755371,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209424083769633,
+ "grad_norm": 0.05625293031334877,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 4.141980171203613,
+ "step": 1455
+ },
+ {
+ "epoch": 20.223385689354277,
+ "grad_norm": 0.05852660909295082,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 4.126821517944336,
+ "step": 1456
+ },
+ {
+ "epoch": 20.237347294938917,
+ "grad_norm": 0.05419151484966278,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 4.10935640335083,
+ "step": 1457
+ },
+ {
+ "epoch": 20.25130890052356,
+ "grad_norm": 0.05251913517713547,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 4.151289463043213,
+ "step": 1458
+ },
+ {
+ "epoch": 20.2652705061082,
+ "grad_norm": 0.05223092809319496,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 4.149935722351074,
+ "step": 1459
+ },
+ {
+ "epoch": 20.279232111692846,
+ "grad_norm": 0.05314765125513077,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 4.106821060180664,
+ "step": 1460
+ },
+ {
+ "epoch": 20.293193717277486,
+ "grad_norm": 0.05472502112388611,
+ "learning_rate": 0.000404858275823277,
+ "loss": 4.120498180389404,
+ "step": 1461
+ },
+ {
+ "epoch": 20.30715532286213,
+ "grad_norm": 0.05233139917254448,
+ "learning_rate": 0.000404607816444578,
+ "loss": 4.1237335205078125,
+ "step": 1462
+ },
+ {
+ "epoch": 20.32111692844677,
+ "grad_norm": 0.05210232734680176,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 4.1271209716796875,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335078534031414,
+ "grad_norm": 0.049779925495386124,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 4.121427536010742,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349040139616054,
+ "grad_norm": 0.05232279747724533,
+ "learning_rate": 0.000403855947934478,
+ "loss": 4.136429786682129,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363001745200698,
+ "grad_norm": 0.05235981196165085,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 4.1808600425720215,
+ "step": 1466
+ },
+ {
+ "epoch": 20.376963350785342,
+ "grad_norm": 0.04902610555291176,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 4.138566017150879,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390924956369982,
+ "grad_norm": 0.04923798143863678,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 4.135126113891602,
+ "step": 1468
+ },
+ {
+ "epoch": 20.404886561954626,
+ "grad_norm": 0.05059202015399933,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 4.1548285484313965,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418848167539267,
+ "grad_norm": 0.05058307573199272,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 4.149041652679443,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43280977312391,
+ "grad_norm": 0.050172705203294754,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 4.148054122924805,
+ "step": 1471
+ },
+ {
+ "epoch": 20.44677137870855,
+ "grad_norm": 0.05158299580216408,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 4.1425018310546875,
+ "step": 1472
+ },
+ {
+ "epoch": 20.460732984293195,
+ "grad_norm": 0.052176836878061295,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 4.129712104797363,
+ "step": 1473
+ },
+ {
+ "epoch": 20.474694589877835,
+ "grad_norm": 0.052095964550971985,
+ "learning_rate": 0.000401595976318565,
+ "loss": 4.134920597076416,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48865619546248,
+ "grad_norm": 0.04944197088479996,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 4.165334701538086,
+ "step": 1475
+ },
+ {
+ "epoch": 20.50261780104712,
+ "grad_norm": 0.049351636320352554,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 4.144362926483154,
+ "step": 1476
+ },
+ {
+ "epoch": 20.516579406631763,
+ "grad_norm": 0.053657080978155136,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 4.164234161376953,
+ "step": 1477
+ },
+ {
+ "epoch": 20.530541012216403,
+ "grad_norm": 0.052578140050172806,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 4.131514549255371,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544502617801047,
+ "grad_norm": 0.049683794379234314,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 4.129961967468262,
+ "step": 1479
+ },
+ {
+ "epoch": 20.55846422338569,
+ "grad_norm": 0.05115316063165665,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 4.164663314819336,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57242582897033,
+ "grad_norm": 0.051620446145534515,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 4.161067485809326,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586387434554975,
+ "grad_norm": 0.04873501881957054,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 4.161318302154541,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600349040139616,
+ "grad_norm": 0.04968130216002464,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 4.151819229125977,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61431064572426,
+ "grad_norm": 0.05476681888103485,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 4.14247989654541,
+ "step": 1484
+ },
+ {
+ "epoch": 20.6282722513089,
+ "grad_norm": 0.05098642408847809,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 4.178083419799805,
+ "step": 1485
+ },
+ {
+ "epoch": 20.642233856893544,
+ "grad_norm": 0.047264356166124344,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 4.15170955657959,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656195462478184,
+ "grad_norm": 0.050382886081933975,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 4.166364669799805,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670157068062828,
+ "grad_norm": 0.04956239461898804,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 4.164124488830566,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68411867364747,
+ "grad_norm": 0.049608584493398666,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 4.130239486694336,
+ "step": 1489
+ },
+ {
+ "epoch": 20.698080279232112,
+ "grad_norm": 0.0489494763314724,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 4.163503646850586,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712041884816752,
+ "grad_norm": 0.04847714304924011,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 4.1744585037231445,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726003490401396,
+ "grad_norm": 0.050256017595529556,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 4.1502227783203125,
+ "step": 1492
+ },
+ {
+ "epoch": 20.739965095986037,
+ "grad_norm": 0.048107463866472244,
+ "learning_rate": 0.000396803974909638,
+ "loss": 4.1790971755981445,
+ "step": 1493
+ },
+ {
+ "epoch": 20.75392670157068,
+ "grad_norm": 0.04828944429755211,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 4.128883361816406,
+ "step": 1494
+ },
+ {
+ "epoch": 20.767888307155324,
+ "grad_norm": 0.04720750078558922,
+ "learning_rate": 0.000396297942393265,
+ "loss": 4.1661577224731445,
+ "step": 1495
+ },
+ {
+ "epoch": 20.781849912739965,
+ "grad_norm": 0.047148339450359344,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 4.172522068023682,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79581151832461,
+ "grad_norm": 0.04612366110086441,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 4.149024963378906,
+ "step": 1497
+ },
+ {
+ "epoch": 20.80977312390925,
+ "grad_norm": 0.048896901309490204,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 4.179235458374023,
+ "step": 1498
+ },
+ {
+ "epoch": 20.823734729493893,
+ "grad_norm": 0.04779870808124542,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 4.1649932861328125,
+ "step": 1499
+ },
+ {
+ "epoch": 20.837696335078533,
+ "grad_norm": 0.04949707165360451,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 4.17645263671875,
+ "step": 1500
+ },
+ {
+ "epoch": 20.851657940663177,
+ "grad_norm": 0.04645023122429848,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 4.216496467590332,
+ "step": 1501
+ },
+ {
+ "epoch": 20.865619546247817,
+ "grad_norm": 0.046520136296749115,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 4.152667999267578,
+ "step": 1502
+ },
+ {
+ "epoch": 20.87958115183246,
+ "grad_norm": 0.04882653057575226,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 4.187082767486572,
+ "step": 1503
+ },
+ {
+ "epoch": 20.8935427574171,
+ "grad_norm": 0.04669157415628433,
+ "learning_rate": 0.000394017098798633,
+ "loss": 4.172348499298096,
+ "step": 1504
+ },
+ {
+ "epoch": 20.907504363001745,
+ "grad_norm": 0.04595336690545082,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 4.180559158325195,
+ "step": 1505
+ },
+ {
+ "epoch": 20.921465968586386,
+ "grad_norm": 0.048478737473487854,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 4.142880439758301,
+ "step": 1506
+ },
+ {
+ "epoch": 20.93542757417103,
+ "grad_norm": 0.04609905555844307,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 4.1459641456604,
+ "step": 1507
+ },
+ {
+ "epoch": 20.949389179755673,
+ "grad_norm": 0.04524408280849457,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 4.185099124908447,
+ "step": 1508
+ },
+ {
+ "epoch": 20.963350785340314,
+ "grad_norm": 0.0475911870598793,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 4.2085161209106445,
+ "step": 1509
+ },
+ {
+ "epoch": 20.977312390924958,
+ "grad_norm": 0.047796521335840225,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 4.165589332580566,
+ "step": 1510
+ },
+ {
+ "epoch": 20.991273996509598,
+ "grad_norm": 0.047358278185129166,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 4.1634368896484375,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.03328625112771988,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 2.6023788452148438,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.5952173471450806,
+ "eval_runtime": 58.6816,
+ "eval_samples_per_second": 41.614,
+ "eval_steps_per_second": 0.665,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013961605584644,
+ "grad_norm": 0.04970651492476463,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 4.077130317687988,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027923211169284,
+ "grad_norm": 0.060882844030857086,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 4.074376106262207,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041884816753928,
+ "grad_norm": 0.06609825044870377,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 4.107635021209717,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05584642233857,
+ "grad_norm": 0.056066352874040604,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 4.068084716796875,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069808027923212,
+ "grad_norm": 0.053013063967227936,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 4.069967269897461,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083769633507853,
+ "grad_norm": 0.059536613523960114,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 4.0639142990112305,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097731239092496,
+ "grad_norm": 0.06084490567445755,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 4.093890190124512,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111692844677137,
+ "grad_norm": 0.05299340933561325,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 4.100567817687988,
+ "step": 1520
+ },
+ {
+ "epoch": 21.12565445026178,
+ "grad_norm": 0.05136161297559738,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 4.08171272277832,
+ "step": 1521
+ },
+ {
+ "epoch": 21.13961605584642,
+ "grad_norm": 0.05201329290866852,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 4.0869975090026855,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153577661431065,
+ "grad_norm": 0.053958263248205185,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 4.0928850173950195,
+ "step": 1523
+ },
+ {
+ "epoch": 21.167539267015705,
+ "grad_norm": 0.053465038537979126,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 4.097828388214111,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18150087260035,
+ "grad_norm": 0.050980277359485626,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 4.101456165313721,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195462478184993,
+ "grad_norm": 0.05415358766913414,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 4.09096097946167,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209424083769633,
+ "grad_norm": 0.05567048117518425,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 4.061287879943848,
+ "step": 1527
+ },
+ {
+ "epoch": 21.223385689354277,
+ "grad_norm": 0.052690375596284866,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 4.080180644989014,
+ "step": 1528
+ },
+ {
+ "epoch": 21.237347294938917,
+ "grad_norm": 0.05448971316218376,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 4.079794883728027,
+ "step": 1529
+ },
+ {
+ "epoch": 21.25130890052356,
+ "grad_norm": 0.055472876876592636,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 4.072025299072266,
+ "step": 1530
+ },
+ {
+ "epoch": 21.2652705061082,
+ "grad_norm": 0.05418326333165169,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 4.103858470916748,
+ "step": 1531
+ },
+ {
+ "epoch": 21.279232111692846,
+ "grad_norm": 0.0524221733212471,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 4.0808916091918945,
+ "step": 1532
+ },
+ {
+ "epoch": 21.293193717277486,
+ "grad_norm": 0.05224563553929329,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 4.090270042419434,
+ "step": 1533
+ },
+ {
+ "epoch": 21.30715532286213,
+ "grad_norm": 0.052917953580617905,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 4.081887245178223,
+ "step": 1534
+ },
+ {
+ "epoch": 21.32111692844677,
+ "grad_norm": 0.05619358271360397,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 4.089571475982666,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335078534031414,
+ "grad_norm": 0.05455326288938522,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 4.096992015838623,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349040139616054,
+ "grad_norm": 0.049311455339193344,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 4.114621639251709,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363001745200698,
+ "grad_norm": 0.05175276845693588,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 4.107684135437012,
+ "step": 1538
+ },
+ {
+ "epoch": 21.376963350785342,
+ "grad_norm": 0.05236738547682762,
+ "learning_rate": 0.00038509205478744,
+ "loss": 4.075709342956543,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390924956369982,
+ "grad_norm": 0.04899081960320473,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 4.108396530151367,
+ "step": 1540
+ },
+ {
+ "epoch": 21.404886561954626,
+ "grad_norm": 0.05129663646221161,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 4.1135711669921875,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418848167539267,
+ "grad_norm": 0.05478592589497566,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 4.11839485168457,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43280977312391,
+ "grad_norm": 0.055830392986536026,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 4.11065149307251,
+ "step": 1543
+ },
+ {
+ "epoch": 21.44677137870855,
+ "grad_norm": 0.05370553582906723,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 4.113103866577148,
+ "step": 1544
+ },
+ {
+ "epoch": 21.460732984293195,
+ "grad_norm": 0.04968908056616783,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 4.119443893432617,
+ "step": 1545
+ },
+ {
+ "epoch": 21.474694589877835,
+ "grad_norm": 0.056149110198020935,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 4.126479625701904,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48865619546248,
+ "grad_norm": 0.053045690059661865,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 4.138535499572754,
+ "step": 1547
+ },
+ {
+ "epoch": 21.50261780104712,
+ "grad_norm": 0.04864223673939705,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 4.099347114562988,
+ "step": 1548
+ },
+ {
+ "epoch": 21.516579406631763,
+ "grad_norm": 0.05234768986701965,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 4.112222194671631,
+ "step": 1549
+ },
+ {
+ "epoch": 21.530541012216403,
+ "grad_norm": 0.050486158579587936,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 4.107913970947266,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544502617801047,
+ "grad_norm": 0.048641227185726166,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 4.114014148712158,
+ "step": 1551
+ },
+ {
+ "epoch": 21.55846422338569,
+ "grad_norm": 0.047759316861629486,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 4.128541946411133,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57242582897033,
+ "grad_norm": 0.046098388731479645,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 4.12142276763916,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586387434554975,
+ "grad_norm": 0.04934769123792648,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 4.13035249710083,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600349040139616,
+ "grad_norm": 0.048911672085523605,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 4.111344337463379,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61431064572426,
+ "grad_norm": 0.049368683248758316,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 4.1345672607421875,
+ "step": 1556
+ },
+ {
+ "epoch": 21.6282722513089,
+ "grad_norm": 0.04890736937522888,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 4.1286821365356445,
+ "step": 1557
+ },
+ {
+ "epoch": 21.642233856893544,
+ "grad_norm": 0.04812583327293396,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 4.09993314743042,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656195462478184,
+ "grad_norm": 0.05025152117013931,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 4.150769233703613,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670157068062828,
+ "grad_norm": 0.05007397383451462,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 4.074525833129883,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68411867364747,
+ "grad_norm": 0.050168465822935104,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 4.0933732986450195,
+ "step": 1561
+ },
+ {
+ "epoch": 21.698080279232112,
+ "grad_norm": 0.04990607872605324,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 4.126741886138916,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712041884816752,
+ "grad_norm": 0.05023440346121788,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 4.120314598083496,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726003490401396,
+ "grad_norm": 0.048100546002388,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 4.1242356300354,
+ "step": 1564
+ },
+ {
+ "epoch": 21.739965095986037,
+ "grad_norm": 0.047999415546655655,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 4.127745628356934,
+ "step": 1565
+ },
+ {
+ "epoch": 21.75392670157068,
+ "grad_norm": 0.05180094763636589,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 4.112565994262695,
+ "step": 1566
+ },
+ {
+ "epoch": 21.767888307155324,
+ "grad_norm": 0.053367964923381805,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 4.109419822692871,
+ "step": 1567
+ },
+ {
+ "epoch": 21.781849912739965,
+ "grad_norm": 0.05205400660634041,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 4.111886024475098,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79581151832461,
+ "grad_norm": 0.0464242659509182,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 4.104648590087891,
+ "step": 1569
+ },
+ {
+ "epoch": 21.80977312390925,
+ "grad_norm": 0.047767721116542816,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 4.156681060791016,
+ "step": 1570
+ },
+ {
+ "epoch": 21.823734729493893,
+ "grad_norm": 0.05028790235519409,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 4.132200241088867,
+ "step": 1571
+ },
+ {
+ "epoch": 21.837696335078533,
+ "grad_norm": 0.04714380204677582,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 4.132688999176025,
+ "step": 1572
+ },
+ {
+ "epoch": 21.851657940663177,
+ "grad_norm": 0.04726320132613182,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 4.173277378082275,
+ "step": 1573
+ },
+ {
+ "epoch": 21.865619546247817,
+ "grad_norm": 0.049558594822883606,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 4.15473747253418,
+ "step": 1574
+ },
+ {
+ "epoch": 21.87958115183246,
+ "grad_norm": 0.04742267727851868,
+ "learning_rate": 0.000375827577564042,
+ "loss": 4.10634708404541,
+ "step": 1575
+ },
+ {
+ "epoch": 21.8935427574171,
+ "grad_norm": 0.048223357647657394,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 4.133949279785156,
+ "step": 1576
+ },
+ {
+ "epoch": 21.907504363001745,
+ "grad_norm": 0.046766068786382675,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 4.153807640075684,
+ "step": 1577
+ },
+ {
+ "epoch": 21.921465968586386,
+ "grad_norm": 0.04586656019091606,
+ "learning_rate": 0.000375051971546236,
+ "loss": 4.169903755187988,
+ "step": 1578
+ },
+ {
+ "epoch": 21.93542757417103,
+ "grad_norm": 0.047912538051605225,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 4.148102283477783,
+ "step": 1579
+ },
+ {
+ "epoch": 21.949389179755673,
+ "grad_norm": 0.048949215561151505,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 4.149707317352295,
+ "step": 1580
+ },
+ {
+ "epoch": 21.963350785340314,
+ "grad_norm": 0.04765952378511429,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 4.128008842468262,
+ "step": 1581
+ },
+ {
+ "epoch": 21.977312390924958,
+ "grad_norm": 0.04726482927799225,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 4.1268720626831055,
+ "step": 1582
+ },
+ {
+ "epoch": 21.991273996509598,
+ "grad_norm": 0.0468299426138401,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 4.1332926750183105,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.03415689244866371,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 2.5875940322875977,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.5966277122497559,
+ "eval_runtime": 59.7704,
+ "eval_samples_per_second": 40.856,
+ "eval_steps_per_second": 0.652,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013961605584644,
+ "grad_norm": 0.04954535514116287,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 4.054527759552002,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027923211169284,
+ "grad_norm": 0.0498427115380764,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 4.048842430114746,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041884816753928,
+ "grad_norm": 0.051013074815273285,
+ "learning_rate": 0.000372722041257003,
+ "loss": 4.036187171936035,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05584642233857,
+ "grad_norm": 0.04973271116614342,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 4.050318717956543,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069808027923212,
+ "grad_norm": 0.051144856959581375,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 4.036319732666016,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083769633507853,
+ "grad_norm": 0.051874030381441116,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 4.017857551574707,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097731239092496,
+ "grad_norm": 0.05181736871600151,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 4.0519819259643555,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111692844677137,
+ "grad_norm": 0.05467049032449722,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 4.064291954040527,
+ "step": 1592
+ },
+ {
+ "epoch": 22.12565445026178,
+ "grad_norm": 0.05251661688089371,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 4.056951522827148,
+ "step": 1593
+ },
+ {
+ "epoch": 22.13961605584642,
+ "grad_norm": 0.05278680846095085,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 4.037125587463379,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153577661431065,
+ "grad_norm": 0.05276667699217796,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 4.053675174713135,
+ "step": 1595
+ },
+ {
+ "epoch": 22.167539267015705,
+ "grad_norm": 0.05007271468639374,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 4.044122695922852,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18150087260035,
+ "grad_norm": 0.051073770970106125,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 4.0265960693359375,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195462478184993,
+ "grad_norm": 0.05045647919178009,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 4.060938835144043,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209424083769633,
+ "grad_norm": 0.05075158178806305,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 4.035902976989746,
+ "step": 1599
+ },
+ {
+ "epoch": 22.223385689354277,
+ "grad_norm": 0.04867244139313698,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 4.055220603942871,
+ "step": 1600
+ },
+ {
+ "epoch": 22.237347294938917,
+ "grad_norm": 0.049976419657468796,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 4.026124954223633,
+ "step": 1601
+ },
+ {
+ "epoch": 22.25130890052356,
+ "grad_norm": 0.04968957230448723,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 4.081721782684326,
+ "step": 1602
+ },
+ {
+ "epoch": 22.2652705061082,
+ "grad_norm": 0.0503058098256588,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 4.0384721755981445,
+ "step": 1603
+ },
+ {
+ "epoch": 22.279232111692846,
+ "grad_norm": 0.050643209367990494,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 4.049397945404053,
+ "step": 1604
+ },
+ {
+ "epoch": 22.293193717277486,
+ "grad_norm": 0.05297046899795532,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 4.046389579772949,
+ "step": 1605
+ },
+ {
+ "epoch": 22.30715532286213,
+ "grad_norm": 0.053165238350629807,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 4.090221405029297,
+ "step": 1606
+ },
+ {
+ "epoch": 22.32111692844677,
+ "grad_norm": 0.05104832351207733,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 4.023227691650391,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335078534031414,
+ "grad_norm": 0.048736415803432465,
+ "learning_rate": 0.000367268105959126,
+ "loss": 4.07478666305542,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349040139616054,
+ "grad_norm": 0.05190873518586159,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 4.068147659301758,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363001745200698,
+ "grad_norm": 0.05409928783774376,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 4.060364723205566,
+ "step": 1610
+ },
+ {
+ "epoch": 22.376963350785342,
+ "grad_norm": 0.05112519487738609,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 4.072710990905762,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390924956369982,
+ "grad_norm": 0.04818312078714371,
+ "learning_rate": 0.000366226612810024,
+ "loss": 4.042757034301758,
+ "step": 1612
+ },
+ {
+ "epoch": 22.404886561954626,
+ "grad_norm": 0.050733599811792374,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 4.064810752868652,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418848167539267,
+ "grad_norm": 0.054070211946964264,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 4.054926872253418,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43280977312391,
+ "grad_norm": 0.05459332838654518,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 4.0652923583984375,
+ "step": 1615
+ },
+ {
+ "epoch": 22.44677137870855,
+ "grad_norm": 0.05064154416322708,
+ "learning_rate": 0.000365184304613104,
+ "loss": 4.076216220855713,
+ "step": 1616
+ },
+ {
+ "epoch": 22.460732984293195,
+ "grad_norm": 0.05071895569562912,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 4.070466041564941,
+ "step": 1617
+ },
+ {
+ "epoch": 22.474694589877835,
+ "grad_norm": 0.05049987882375717,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 4.062691688537598,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48865619546248,
+ "grad_norm": 0.05174451321363449,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 4.07501220703125,
+ "step": 1619
+ },
+ {
+ "epoch": 22.50261780104712,
+ "grad_norm": 0.05360838770866394,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 4.054711818695068,
+ "step": 1620
+ },
+ {
+ "epoch": 22.516579406631763,
+ "grad_norm": 0.051809027791023254,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 4.075811386108398,
+ "step": 1621
+ },
+ {
+ "epoch": 22.530541012216403,
+ "grad_norm": 0.05586795508861542,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 4.0638203620910645,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544502617801047,
+ "grad_norm": 0.05571979284286499,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 4.089414596557617,
+ "step": 1623
+ },
+ {
+ "epoch": 22.55846422338569,
+ "grad_norm": 0.05098723992705345,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 4.091159820556641,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57242582897033,
+ "grad_norm": 0.05359277129173279,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 4.060613632202148,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586387434554975,
+ "grad_norm": 0.05281100794672966,
+ "learning_rate": 0.000362575056487632,
+ "loss": 4.08416748046875,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600349040139616,
+ "grad_norm": 0.053913842886686325,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 4.088703155517578,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61431064572426,
+ "grad_norm": 0.05476798862218857,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 4.064074516296387,
+ "step": 1628
+ },
+ {
+ "epoch": 22.6282722513089,
+ "grad_norm": 0.051888495683670044,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 4.116634368896484,
+ "step": 1629
+ },
+ {
+ "epoch": 22.642233856893544,
+ "grad_norm": 0.051054876297712326,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 4.0800018310546875,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656195462478184,
+ "grad_norm": 0.054063159972429276,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 4.112773418426514,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670157068062828,
+ "grad_norm": 0.05173034593462944,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 4.062599182128906,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68411867364747,
+ "grad_norm": 0.054618336260318756,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 4.110188961029053,
+ "step": 1633
+ },
+ {
+ "epoch": 22.698080279232112,
+ "grad_norm": 0.055602896958589554,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 4.140998840332031,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712041884816752,
+ "grad_norm": 0.05332222580909729,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 4.119804382324219,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726003490401396,
+ "grad_norm": 0.05162122845649719,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 4.0731282234191895,
+ "step": 1636
+ },
+ {
+ "epoch": 22.739965095986037,
+ "grad_norm": 0.0514204278588295,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 4.068007946014404,
+ "step": 1637
+ },
+ {
+ "epoch": 22.75392670157068,
+ "grad_norm": 0.05027637258172035,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 4.097658157348633,
+ "step": 1638
+ },
+ {
+ "epoch": 22.767888307155324,
+ "grad_norm": 0.051812347024679184,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 4.083312511444092,
+ "step": 1639
+ },
+ {
+ "epoch": 22.781849912739965,
+ "grad_norm": 0.05145815759897232,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 4.129471778869629,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79581151832461,
+ "grad_norm": 0.05093684792518616,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 4.091841220855713,
+ "step": 1641
+ },
+ {
+ "epoch": 22.80977312390925,
+ "grad_norm": 0.05100737139582634,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 4.076615810394287,
+ "step": 1642
+ },
+ {
+ "epoch": 22.823734729493893,
+ "grad_norm": 0.05079149082303047,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 4.067737579345703,
+ "step": 1643
+ },
+ {
+ "epoch": 22.837696335078533,
+ "grad_norm": 0.05044509842991829,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 4.118824481964111,
+ "step": 1644
+ },
+ {
+ "epoch": 22.851657940663177,
+ "grad_norm": 0.05010584369301796,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 4.113405227661133,
+ "step": 1645
+ },
+ {
+ "epoch": 22.865619546247817,
+ "grad_norm": 0.052232302725315094,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 4.101553440093994,
+ "step": 1646
+ },
+ {
+ "epoch": 22.87958115183246,
+ "grad_norm": 0.05080661550164223,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 4.116809844970703,
+ "step": 1647
+ },
+ {
+ "epoch": 22.8935427574171,
+ "grad_norm": 0.05051033943891525,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 4.128484725952148,
+ "step": 1648
+ },
+ {
+ "epoch": 22.907504363001745,
+ "grad_norm": 0.051927294582128525,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 4.092285633087158,
+ "step": 1649
+ },
+ {
+ "epoch": 22.921465968586386,
+ "grad_norm": 0.049063827842473984,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 4.088473796844482,
+ "step": 1650
+ },
+ {
+ "epoch": 22.93542757417103,
+ "grad_norm": 0.05048854276537895,
+ "learning_rate": 0.000356031397755195,
+ "loss": 4.081521987915039,
+ "step": 1651
+ },
+ {
+ "epoch": 22.949389179755673,
+ "grad_norm": 0.049307651817798615,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 4.122158050537109,
+ "step": 1652
+ },
+ {
+ "epoch": 22.963350785340314,
+ "grad_norm": 0.048311345279216766,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 4.100488662719727,
+ "step": 1653
+ },
+ {
+ "epoch": 22.977312390924958,
+ "grad_norm": 0.05016734451055527,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 4.088587760925293,
+ "step": 1654
+ },
+ {
+ "epoch": 22.991273996509598,
+ "grad_norm": 0.04944450408220291,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 4.114686012268066,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.03592754527926445,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 2.545376777648926,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.5981184840202332,
+ "eval_runtime": 59.9558,
+ "eval_samples_per_second": 40.73,
+ "eval_steps_per_second": 0.65,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013961605584644,
+ "grad_norm": 0.05112785845994949,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 4.00495719909668,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027923211169284,
+ "grad_norm": 0.05807499587535858,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 4.015654563903809,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041884816753928,
+ "grad_norm": 0.05383148789405823,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.979349136352539,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05584642233857,
+ "grad_norm": 0.05221369490027428,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 4.009493827819824,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069808027923212,
+ "grad_norm": 0.05144324526190758,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 4.045680999755859,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083769633507853,
+ "grad_norm": 0.051422782242298126,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.994011878967285,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097731239092496,
+ "grad_norm": 0.05142863467335701,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 4.03209114074707,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111692844677137,
+ "grad_norm": 0.050954073667526245,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 4.00391960144043,
+ "step": 1664
+ },
+ {
+ "epoch": 23.12565445026178,
+ "grad_norm": 0.05365024879574776,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.9839367866516113,
+ "step": 1665
+ },
+ {
+ "epoch": 23.13961605584642,
+ "grad_norm": 0.051181286573410034,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 4.017707824707031,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153577661431065,
+ "grad_norm": 0.054279860109090805,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.989722490310669,
+ "step": 1667
+ },
+ {
+ "epoch": 23.167539267015705,
+ "grad_norm": 0.052196286618709564,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 4.014932155609131,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18150087260035,
+ "grad_norm": 0.05098043382167816,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 4.026510238647461,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195462478184993,
+ "grad_norm": 0.05241888761520386,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.979708433151245,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209424083769633,
+ "grad_norm": 0.051801104098558426,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 4.029015064239502,
+ "step": 1671
+ },
+ {
+ "epoch": 23.223385689354277,
+ "grad_norm": 0.052666619420051575,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.9932432174682617,
+ "step": 1672
+ },
+ {
+ "epoch": 23.237347294938917,
+ "grad_norm": 0.054418303072452545,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 4.05676794052124,
+ "step": 1673
+ },
+ {
+ "epoch": 23.25130890052356,
+ "grad_norm": 0.055074021220207214,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.999929666519165,
+ "step": 1674
+ },
+ {
+ "epoch": 23.2652705061082,
+ "grad_norm": 0.05626136437058449,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 4.013473987579346,
+ "step": 1675
+ },
+ {
+ "epoch": 23.279232111692846,
+ "grad_norm": 0.054028257727622986,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 4.015802383422852,
+ "step": 1676
+ },
+ {
+ "epoch": 23.293193717277486,
+ "grad_norm": 0.054044198244810104,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 4.020464897155762,
+ "step": 1677
+ },
+ {
+ "epoch": 23.30715532286213,
+ "grad_norm": 0.05465259402990341,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 4.010867118835449,
+ "step": 1678
+ },
+ {
+ "epoch": 23.32111692844677,
+ "grad_norm": 0.05545585975050926,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 4.054412841796875,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335078534031414,
+ "grad_norm": 0.055265430361032486,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 4.020756244659424,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349040139616054,
+ "grad_norm": 0.053363267332315445,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 4.033337116241455,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363001745200698,
+ "grad_norm": 0.05347830802202225,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 4.024794578552246,
+ "step": 1682
+ },
+ {
+ "epoch": 23.376963350785342,
+ "grad_norm": 0.05496951565146446,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 4.013302803039551,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390924956369982,
+ "grad_norm": 0.05179804563522339,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 4.030914306640625,
+ "step": 1684
+ },
+ {
+ "epoch": 23.404886561954626,
+ "grad_norm": 0.054958634078502655,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 4.012727737426758,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418848167539267,
+ "grad_norm": 0.05533262714743614,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 4.0543365478515625,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43280977312391,
+ "grad_norm": 0.052329424768686295,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 4.031486511230469,
+ "step": 1687
+ },
+ {
+ "epoch": 23.44677137870855,
+ "grad_norm": 0.05273434892296791,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 4.040292739868164,
+ "step": 1688
+ },
+ {
+ "epoch": 23.460732984293195,
+ "grad_norm": 0.05445481091737747,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 4.025630950927734,
+ "step": 1689
+ },
+ {
+ "epoch": 23.474694589877835,
+ "grad_norm": 0.052102912217378616,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 4.049562454223633,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48865619546248,
+ "grad_norm": 0.051272716373205185,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 4.0431365966796875,
+ "step": 1691
+ },
+ {
+ "epoch": 23.50261780104712,
+ "grad_norm": 0.051311176270246506,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 4.042119026184082,
+ "step": 1692
+ },
+ {
+ "epoch": 23.516579406631763,
+ "grad_norm": 0.05042742192745209,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 4.044881820678711,
+ "step": 1693
+ },
+ {
+ "epoch": 23.530541012216403,
+ "grad_norm": 0.05103316903114319,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 4.033669471740723,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544502617801047,
+ "grad_norm": 0.05441150814294815,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 4.076536178588867,
+ "step": 1695
+ },
+ {
+ "epoch": 23.55846422338569,
+ "grad_norm": 0.05529651418328285,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 4.073678016662598,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57242582897033,
+ "grad_norm": 0.0526033453643322,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 4.067898750305176,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586387434554975,
+ "grad_norm": 0.052817948162555695,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 4.017403602600098,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600349040139616,
+ "grad_norm": 0.052461300045251846,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 4.026311874389648,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61431064572426,
+ "grad_norm": 0.05097689852118492,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 4.076074600219727,
+ "step": 1700
+ },
+ {
+ "epoch": 23.6282722513089,
+ "grad_norm": 0.05387786403298378,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 4.022763729095459,
+ "step": 1701
+ },
+ {
+ "epoch": 23.642233856893544,
+ "grad_norm": 0.05132855102419853,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 4.077712535858154,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656195462478184,
+ "grad_norm": 0.05380886048078537,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 4.038937091827393,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670157068062828,
+ "grad_norm": 0.053299430757761,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 4.063364028930664,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68411867364747,
+ "grad_norm": 0.051913514733314514,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 4.05495548248291,
+ "step": 1705
+ },
+ {
+ "epoch": 23.698080279232112,
+ "grad_norm": 0.05660073831677437,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 4.057949066162109,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712041884816752,
+ "grad_norm": 0.054757390171289444,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 4.05377721786499,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726003490401396,
+ "grad_norm": 0.0517287515103817,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 4.035928726196289,
+ "step": 1708
+ },
+ {
+ "epoch": 23.739965095986037,
+ "grad_norm": 0.05187627300620079,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 4.043362617492676,
+ "step": 1709
+ },
+ {
+ "epoch": 23.75392670157068,
+ "grad_norm": 0.053502220660448074,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 4.007593154907227,
+ "step": 1710
+ },
+ {
+ "epoch": 23.767888307155324,
+ "grad_norm": 0.05755629390478134,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 4.076575756072998,
+ "step": 1711
+ },
+ {
+ "epoch": 23.781849912739965,
+ "grad_norm": 0.05619359388947487,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 4.072240352630615,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79581151832461,
+ "grad_norm": 0.05604318529367447,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 4.0486650466918945,
+ "step": 1713
+ },
+ {
+ "epoch": 23.80977312390925,
+ "grad_norm": 0.05630200728774071,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 4.086282730102539,
+ "step": 1714
+ },
+ {
+ "epoch": 23.823734729493893,
+ "grad_norm": 0.053262822329998016,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 4.089612007141113,
+ "step": 1715
+ },
+ {
+ "epoch": 23.837696335078533,
+ "grad_norm": 0.05699621140956879,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 4.053062438964844,
+ "step": 1716
+ },
+ {
+ "epoch": 23.851657940663177,
+ "grad_norm": 0.05963658541440964,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 4.072884559631348,
+ "step": 1717
+ },
+ {
+ "epoch": 23.865619546247817,
+ "grad_norm": 0.0547492615878582,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 4.066218852996826,
+ "step": 1718
+ },
+ {
+ "epoch": 23.87958115183246,
+ "grad_norm": 0.054908450692892075,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 4.035344123840332,
+ "step": 1719
+ },
+ {
+ "epoch": 23.8935427574171,
+ "grad_norm": 0.055191367864608765,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 4.08471155166626,
+ "step": 1720
+ },
+ {
+ "epoch": 23.907504363001745,
+ "grad_norm": 0.05334226414561272,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 4.076174736022949,
+ "step": 1721
+ },
+ {
+ "epoch": 23.921465968586386,
+ "grad_norm": 0.05579005926847458,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 4.05291223526001,
+ "step": 1722
+ },
+ {
+ "epoch": 23.93542757417103,
+ "grad_norm": 0.05464969575405121,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 4.07335090637207,
+ "step": 1723
+ },
+ {
+ "epoch": 23.949389179755673,
+ "grad_norm": 0.05240754038095474,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 4.0995073318481445,
+ "step": 1724
+ },
+ {
+ "epoch": 23.963350785340314,
+ "grad_norm": 0.05416056886315346,
+ "learning_rate": 0.000336518346307424,
+ "loss": 4.072793960571289,
+ "step": 1725
+ },
+ {
+ "epoch": 23.977312390924958,
+ "grad_norm": 0.05496834218502045,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 4.059562683105469,
+ "step": 1726
+ },
+ {
+ "epoch": 23.991273996509598,
+ "grad_norm": 0.05537887290120125,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 4.050772666931152,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.036677468568086624,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 2.5407257080078125,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.5995977520942688,
+ "eval_runtime": 60.2704,
+ "eval_samples_per_second": 40.517,
+ "eval_steps_per_second": 0.647,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013961605584644,
+ "grad_norm": 0.05583682656288147,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.9873995780944824,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027923211169284,
+ "grad_norm": 0.0605769157409668,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.999206066131592,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041884816753928,
+ "grad_norm": 0.057975225150585175,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.9923095703125,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05584642233857,
+ "grad_norm": 0.056104276329278946,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.946317672729492,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069808027923212,
+ "grad_norm": 0.05481491982936859,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.9973483085632324,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083769633507853,
+ "grad_norm": 0.05460227280855179,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.984617233276367,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097731239092496,
+ "grad_norm": 0.058481909334659576,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 4.0138773918151855,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111692844677137,
+ "grad_norm": 0.05672750994563103,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.9683098793029785,
+ "step": 1736
+ },
+ {
+ "epoch": 24.12565445026178,
+ "grad_norm": 0.05315157398581505,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.9909560680389404,
+ "step": 1737
+ },
+ {
+ "epoch": 24.13961605584642,
+ "grad_norm": 0.05565650761127472,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.954531192779541,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153577661431065,
+ "grad_norm": 0.05550679191946983,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.97672700881958,
+ "step": 1739
+ },
+ {
+ "epoch": 24.167539267015705,
+ "grad_norm": 0.05608576908707619,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.987941265106201,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18150087260035,
+ "grad_norm": 0.05266846343874931,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.97526216506958,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195462478184993,
+ "grad_norm": 0.05439139902591705,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 4.012054920196533,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209424083769633,
+ "grad_norm": 0.05473192036151886,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 4.022945404052734,
+ "step": 1743
+ },
+ {
+ "epoch": 24.223385689354277,
+ "grad_norm": 0.054737839847803116,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.981072425842285,
+ "step": 1744
+ },
+ {
+ "epoch": 24.237347294938917,
+ "grad_norm": 0.05696691572666168,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.9708170890808105,
+ "step": 1745
+ },
+ {
+ "epoch": 24.25130890052356,
+ "grad_norm": 0.05677401274442673,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.985995292663574,
+ "step": 1746
+ },
+ {
+ "epoch": 24.2652705061082,
+ "grad_norm": 0.05895088613033295,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.991887092590332,
+ "step": 1747
+ },
+ {
+ "epoch": 24.279232111692846,
+ "grad_norm": 0.05758961662650108,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.9695687294006348,
+ "step": 1748
+ },
+ {
+ "epoch": 24.293193717277486,
+ "grad_norm": 0.05174405500292778,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.959138870239258,
+ "step": 1749
+ },
+ {
+ "epoch": 24.30715532286213,
+ "grad_norm": 0.055284611880779266,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 4.012698173522949,
+ "step": 1750
+ },
+ {
+ "epoch": 24.32111692844677,
+ "grad_norm": 0.06042446941137314,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.9892425537109375,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335078534031414,
+ "grad_norm": 0.05489135906100273,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.9962222576141357,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349040139616054,
+ "grad_norm": 0.0542326495051384,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.9787094593048096,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363001745200698,
+ "grad_norm": 0.05612313374876976,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.9881339073181152,
+ "step": 1754
+ },
+ {
+ "epoch": 24.376963350785342,
+ "grad_norm": 0.05661709979176521,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 4.009440898895264,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390924956369982,
+ "grad_norm": 0.055537596344947815,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 4.00221061706543,
+ "step": 1756
+ },
+ {
+ "epoch": 24.404886561954626,
+ "grad_norm": 0.054210033267736435,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.9807987213134766,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418848167539267,
+ "grad_norm": 0.05847020074725151,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.960838794708252,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43280977312391,
+ "grad_norm": 0.05386386811733246,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 4.017860412597656,
+ "step": 1759
+ },
+ {
+ "epoch": 24.44677137870855,
+ "grad_norm": 0.056131865829229355,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.980543613433838,
+ "step": 1760
+ },
+ {
+ "epoch": 24.460732984293195,
+ "grad_norm": 0.05648239329457283,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.9948508739471436,
+ "step": 1761
+ },
+ {
+ "epoch": 24.474694589877835,
+ "grad_norm": 0.054138120263814926,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 4.002998352050781,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48865619546248,
+ "grad_norm": 0.05725586414337158,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 4.01864767074585,
+ "step": 1763
+ },
+ {
+ "epoch": 24.50261780104712,
+ "grad_norm": 0.05680665746331215,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 4.023143768310547,
+ "step": 1764
+ },
+ {
+ "epoch": 24.516579406631763,
+ "grad_norm": 0.052627697587013245,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 4.006389617919922,
+ "step": 1765
+ },
+ {
+ "epoch": 24.530541012216403,
+ "grad_norm": 0.053234782069921494,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 4.007419586181641,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544502617801047,
+ "grad_norm": 0.05756475403904915,
+ "learning_rate": 0.000325372061241796,
+ "loss": 4.0226945877075195,
+ "step": 1767
+ },
+ {
+ "epoch": 24.55846422338569,
+ "grad_norm": 0.055328067392110825,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 4.022767066955566,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57242582897033,
+ "grad_norm": 0.05378621071577072,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 4.005022048950195,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586387434554975,
+ "grad_norm": 0.056049644947052,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 4.025008201599121,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600349040139616,
+ "grad_norm": 0.057860117405653,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 4.003593444824219,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61431064572426,
+ "grad_norm": 0.056685276329517365,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.997868299484253,
+ "step": 1772
+ },
+ {
+ "epoch": 24.6282722513089,
+ "grad_norm": 0.053294967859983444,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 4.0211992263793945,
+ "step": 1773
+ },
+ {
+ "epoch": 24.642233856893544,
+ "grad_norm": 0.053967639803886414,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 4.022367477416992,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656195462478184,
+ "grad_norm": 0.052357546985149384,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 4.021792411804199,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670157068062828,
+ "grad_norm": 0.0547296479344368,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 4.008403778076172,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68411867364747,
+ "grad_norm": 0.056104280054569244,
+ "learning_rate": 0.000322712903929477,
+ "loss": 4.033827781677246,
+ "step": 1777
+ },
+ {
+ "epoch": 24.698080279232112,
+ "grad_norm": 0.055415406823158264,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 4.015371322631836,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712041884816752,
+ "grad_norm": 0.054506298154592514,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 4.0000152587890625,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726003490401396,
+ "grad_norm": 0.05468682199716568,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.9905481338500977,
+ "step": 1780
+ },
+ {
+ "epoch": 24.739965095986037,
+ "grad_norm": 0.05323269963264465,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 4.040729999542236,
+ "step": 1781
+ },
+ {
+ "epoch": 24.75392670157068,
+ "grad_norm": 0.052569612860679626,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 4.028299331665039,
+ "step": 1782
+ },
+ {
+ "epoch": 24.767888307155324,
+ "grad_norm": 0.052494753152132034,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 4.028860569000244,
+ "step": 1783
+ },
+ {
+ "epoch": 24.781849912739965,
+ "grad_norm": 0.05346379801630974,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 4.000349044799805,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79581151832461,
+ "grad_norm": 0.051543332636356354,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 4.016782760620117,
+ "step": 1785
+ },
+ {
+ "epoch": 24.80977312390925,
+ "grad_norm": 0.05330873280763626,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 4.024371147155762,
+ "step": 1786
+ },
+ {
+ "epoch": 24.823734729493893,
+ "grad_norm": 0.05306831747293472,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 4.0101213455200195,
+ "step": 1787
+ },
+ {
+ "epoch": 24.837696335078533,
+ "grad_norm": 0.05148417130112648,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 4.020768642425537,
+ "step": 1788
+ },
+ {
+ "epoch": 24.851657940663177,
+ "grad_norm": 0.050848908722400665,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 4.030246734619141,
+ "step": 1789
+ },
+ {
+ "epoch": 24.865619546247817,
+ "grad_norm": 0.05279014632105827,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 4.04910135269165,
+ "step": 1790
+ },
+ {
+ "epoch": 24.87958115183246,
+ "grad_norm": 0.0514594204723835,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 4.025290489196777,
+ "step": 1791
+ },
+ {
+ "epoch": 24.8935427574171,
+ "grad_norm": 0.05200032889842987,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 4.019560813903809,
+ "step": 1792
+ },
+ {
+ "epoch": 24.907504363001745,
+ "grad_norm": 0.052474118769168854,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 4.035177230834961,
+ "step": 1793
+ },
+ {
+ "epoch": 24.921465968586386,
+ "grad_norm": 0.05411219224333763,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.9920859336853027,
+ "step": 1794
+ },
+ {
+ "epoch": 24.93542757417103,
+ "grad_norm": 0.0543992817401886,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 4.017391204833984,
+ "step": 1795
+ },
+ {
+ "epoch": 24.949389179755673,
+ "grad_norm": 0.05187678337097168,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 4.042266368865967,
+ "step": 1796
+ },
+ {
+ "epoch": 24.963350785340314,
+ "grad_norm": 0.053762417286634445,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 4.029013633728027,
+ "step": 1797
+ },
+ {
+ "epoch": 24.977312390924958,
+ "grad_norm": 0.051723796874284744,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 4.038886070251465,
+ "step": 1798
+ },
+ {
+ "epoch": 24.991273996509598,
+ "grad_norm": 0.05242295190691948,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 4.056063652038574,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.03998562693595886,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 2.534770965576172,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 0.6009227633476257,
+ "eval_runtime": 58.9421,
+ "eval_samples_per_second": 41.43,
+ "eval_steps_per_second": 0.662,
+ "step": 1800
+ },
+ {
+ "epoch": 25.013961605584644,
+ "grad_norm": 0.05450186878442764,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 3.936816692352295,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027923211169284,
+ "grad_norm": 0.06389589607715607,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 3.9368391036987305,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041884816753928,
+ "grad_norm": 0.06028348580002785,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 3.953296661376953,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05584642233857,
+ "grad_norm": 0.05359262600541115,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 3.924558639526367,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069808027923212,
+ "grad_norm": 0.059837762266397476,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 3.9608402252197266,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083769633507853,
+ "grad_norm": 0.05523517727851868,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 3.943776845932007,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097731239092496,
+ "grad_norm": 0.05617433041334152,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 3.9635634422302246,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111692844677137,
+ "grad_norm": 0.06066269427537918,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 3.9647364616394043,
+ "step": 1808
+ },
+ {
+ "epoch": 25.12565445026178,
+ "grad_norm": 0.05491397902369499,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 3.95222806930542,
+ "step": 1809
+ },
+ {
+ "epoch": 25.13961605584642,
+ "grad_norm": 0.05731203034520149,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 3.9440197944641113,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153577661431065,
+ "grad_norm": 0.05428403243422508,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 3.9428441524505615,
+ "step": 1811
+ },
+ {
+ "epoch": 25.167539267015705,
+ "grad_norm": 0.05487453565001488,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 3.9489784240722656,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18150087260035,
+ "grad_norm": 0.05725884810090065,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 3.9643478393554688,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195462478184993,
+ "grad_norm": 0.05752462521195412,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 3.963301181793213,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209424083769633,
+ "grad_norm": 0.057387594133615494,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 3.9427294731140137,
+ "step": 1815
+ },
+ {
+ "epoch": 25.223385689354277,
+ "grad_norm": 0.057492151856422424,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 3.9229259490966797,
+ "step": 1816
+ },
+ {
+ "epoch": 25.237347294938917,
+ "grad_norm": 0.05202317237854004,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 3.938037395477295,
+ "step": 1817
+ },
+ {
+ "epoch": 25.25130890052356,
+ "grad_norm": 0.05574658885598183,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 3.9532949924468994,
+ "step": 1818
+ },
+ {
+ "epoch": 25.2652705061082,
+ "grad_norm": 0.05714752897620201,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 3.962207794189453,
+ "step": 1819
+ },
+ {
+ "epoch": 25.279232111692846,
+ "grad_norm": 0.05540093779563904,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 3.96329402923584,
+ "step": 1820
+ },
+ {
+ "epoch": 25.293193717277486,
+ "grad_norm": 0.056597232818603516,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 3.948793411254883,
+ "step": 1821
+ },
+ {
+ "epoch": 25.30715532286213,
+ "grad_norm": 0.055902402848005295,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 3.9111557006835938,
+ "step": 1822
+ },
+ {
+ "epoch": 25.32111692844677,
+ "grad_norm": 0.05509798601269722,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 3.929591655731201,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335078534031414,
+ "grad_norm": 0.05763949453830719,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 3.9758448600769043,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349040139616054,
+ "grad_norm": 0.053838349878787994,
+ "learning_rate": 0.000309930621834911,
+ "loss": 3.92457914352417,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363001745200698,
+ "grad_norm": 0.05467558279633522,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 3.9501285552978516,
+ "step": 1826
+ },
+ {
+ "epoch": 25.376963350785342,
+ "grad_norm": 0.05715392529964447,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 3.9734368324279785,
+ "step": 1827
+ },
+ {
+ "epoch": 25.390924956369982,
+ "grad_norm": 0.0566536970436573,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 3.9933152198791504,
+ "step": 1828
+ },
+ {
+ "epoch": 25.404886561954626,
+ "grad_norm": 0.053221236914396286,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 3.967787265777588,
+ "step": 1829
+ },
+ {
+ "epoch": 25.418848167539267,
+ "grad_norm": 0.05779031664133072,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 3.9415433406829834,
+ "step": 1830
+ },
+ {
+ "epoch": 25.43280977312391,
+ "grad_norm": 0.062013521790504456,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 3.9801456928253174,
+ "step": 1831
+ },
+ {
+ "epoch": 25.44677137870855,
+ "grad_norm": 0.05819004401564598,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 3.966928005218506,
+ "step": 1832
+ },
+ {
+ "epoch": 25.460732984293195,
+ "grad_norm": 0.05754564702510834,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 3.9612770080566406,
+ "step": 1833
+ },
+ {
+ "epoch": 25.474694589877835,
+ "grad_norm": 0.05707791447639465,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 3.935521125793457,
+ "step": 1834
+ },
+ {
+ "epoch": 25.48865619546248,
+ "grad_norm": 0.056380245834589005,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 4.006120681762695,
+ "step": 1835
+ },
+ {
+ "epoch": 25.50261780104712,
+ "grad_norm": 0.0551077239215374,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 3.9544754028320312,
+ "step": 1836
+ },
+ {
+ "epoch": 25.516579406631763,
+ "grad_norm": 0.05720360204577446,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 3.9564664363861084,
+ "step": 1837
+ },
+ {
+ "epoch": 25.530541012216403,
+ "grad_norm": 0.05522238463163376,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 4.0157880783081055,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544502617801047,
+ "grad_norm": 0.05304659530520439,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 3.957207202911377,
+ "step": 1839
+ },
+ {
+ "epoch": 25.55846422338569,
+ "grad_norm": 0.05569292604923248,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 3.9909608364105225,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57242582897033,
+ "grad_norm": 0.05347870662808418,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 3.9537010192871094,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586387434554975,
+ "grad_norm": 0.05572560802102089,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 3.9818642139434814,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600349040139616,
+ "grad_norm": 0.05624900385737419,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 3.995117425918579,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61431064572426,
+ "grad_norm": 0.05600332096219063,
+ "learning_rate": 0.000304866093757771,
+ "loss": 3.9480154514312744,
+ "step": 1844
+ },
+ {
+ "epoch": 25.6282722513089,
+ "grad_norm": 0.05488704890012741,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 3.993014335632324,
+ "step": 1845
+ },
+ {
+ "epoch": 25.642233856893544,
+ "grad_norm": 0.05647030100226402,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 3.994615077972412,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656195462478184,
+ "grad_norm": 0.05460594967007637,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 3.9714808464050293,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670157068062828,
+ "grad_norm": 0.05629213526844978,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 3.962033271789551,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68411867364747,
+ "grad_norm": 0.05666861683130264,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 3.995985269546509,
+ "step": 1849
+ },
+ {
+ "epoch": 25.698080279232112,
+ "grad_norm": 0.05580414459109306,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 4.001766681671143,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712041884816752,
+ "grad_norm": 0.05638023093342781,
+ "learning_rate": 0.000303,
+ "loss": 4.005331993103027,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726003490401396,
+ "grad_norm": 0.05684564262628555,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 4.017581939697266,
+ "step": 1852
+ },
+ {
+ "epoch": 25.739965095986037,
+ "grad_norm": 0.057602159678936005,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 3.9668312072753906,
+ "step": 1853
+ },
+ {
+ "epoch": 25.75392670157068,
+ "grad_norm": 0.055540211498737335,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 3.9885573387145996,
+ "step": 1854
+ },
+ {
+ "epoch": 25.767888307155324,
+ "grad_norm": 0.05655810981988907,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 3.990997314453125,
+ "step": 1855
+ },
+ {
+ "epoch": 25.781849912739965,
+ "grad_norm": 0.05443689972162247,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 3.9554123878479004,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79581151832461,
+ "grad_norm": 0.05570529028773308,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 3.983283519744873,
+ "step": 1857
+ },
+ {
+ "epoch": 25.80977312390925,
+ "grad_norm": 0.05399419739842415,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 4.004507064819336,
+ "step": 1858
+ },
+ {
+ "epoch": 25.823734729493893,
+ "grad_norm": 0.05587748438119888,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 4.00900936126709,
+ "step": 1859
+ },
+ {
+ "epoch": 25.837696335078533,
+ "grad_norm": 0.05465575307607651,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 3.9826579093933105,
+ "step": 1860
+ },
+ {
+ "epoch": 25.851657940663177,
+ "grad_norm": 0.05286499112844467,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 3.973453998565674,
+ "step": 1861
+ },
+ {
+ "epoch": 25.865619546247817,
+ "grad_norm": 0.05495590344071388,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 4.003095626831055,
+ "step": 1862
+ },
+ {
+ "epoch": 25.87958115183246,
+ "grad_norm": 0.053388018161058426,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 4.00042200088501,
+ "step": 1863
+ },
+ {
+ "epoch": 25.8935427574171,
+ "grad_norm": 0.05323237553238869,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 3.992929458618164,
+ "step": 1864
+ },
+ {
+ "epoch": 25.907504363001745,
+ "grad_norm": 0.053946711122989655,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 3.996519088745117,
+ "step": 1865
+ },
+ {
+ "epoch": 25.921465968586386,
+ "grad_norm": 0.05455862358212471,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 4.007123947143555,
+ "step": 1866
+ },
+ {
+ "epoch": 25.93542757417103,
+ "grad_norm": 0.05290355160832405,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 4.001724720001221,
+ "step": 1867
+ },
+ {
+ "epoch": 25.949389179755673,
+ "grad_norm": 0.052808016538619995,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 3.9964821338653564,
+ "step": 1868
+ },
+ {
+ "epoch": 25.963350785340314,
+ "grad_norm": 0.05416771396994591,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 3.98549222946167,
+ "step": 1869
+ },
+ {
+ "epoch": 25.977312390924958,
+ "grad_norm": 0.05264521390199661,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 3.9729766845703125,
+ "step": 1870
+ },
+ {
+ "epoch": 25.991273996509598,
+ "grad_norm": 0.05483962222933769,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 4.018877029418945,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.03927824646234512,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 2.456951856613159,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 0.6026532053947449,
+ "eval_runtime": 60.4079,
+ "eval_samples_per_second": 40.425,
+ "eval_steps_per_second": 0.646,
+ "step": 1872
+ },
+ {
+ "epoch": 26.013961605584644,
+ "grad_norm": 0.05756894499063492,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 3.93898606300354,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027923211169284,
+ "grad_norm": 0.06288876384496689,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 3.9142136573791504,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041884816753928,
+ "grad_norm": 0.0600322000682354,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 3.918700695037842,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05584642233857,
+ "grad_norm": 0.05729348957538605,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 3.9040162563323975,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069808027923212,
+ "grad_norm": 0.05658194422721863,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 3.8809690475463867,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083769633507853,
+ "grad_norm": 0.06274262815713882,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 3.928222179412842,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097731239092496,
+ "grad_norm": 0.06313030421733856,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 3.9395527839660645,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111692844677137,
+ "grad_norm": 0.0554068423807621,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 3.9314563274383545,
+ "step": 1880
+ },
+ {
+ "epoch": 26.12565445026178,
+ "grad_norm": 0.05697336047887802,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 3.9071288108825684,
+ "step": 1881
+ },
+ {
+ "epoch": 26.13961605584642,
+ "grad_norm": 0.05603424087166786,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 3.932054042816162,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153577661431065,
+ "grad_norm": 0.055603109300136566,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 3.9169063568115234,
+ "step": 1883
+ },
+ {
+ "epoch": 26.167539267015705,
+ "grad_norm": 0.055022306740283966,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 3.8985838890075684,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18150087260035,
+ "grad_norm": 0.05383468046784401,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 3.896359443664551,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195462478184993,
+ "grad_norm": 0.055076900869607925,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 3.905672073364258,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209424083769633,
+ "grad_norm": 0.05629882961511612,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 3.9222521781921387,
+ "step": 1887
+ },
+ {
+ "epoch": 26.223385689354277,
+ "grad_norm": 0.056736111640930176,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 3.8918867111206055,
+ "step": 1888
+ },
+ {
+ "epoch": 26.237347294938917,
+ "grad_norm": 0.05625338479876518,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 3.9249935150146484,
+ "step": 1889
+ },
+ {
+ "epoch": 26.25130890052356,
+ "grad_norm": 0.06145375967025757,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 3.904536724090576,
+ "step": 1890
+ },
+ {
+ "epoch": 26.2652705061082,
+ "grad_norm": 0.05601578950881958,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 3.8845863342285156,
+ "step": 1891
+ },
+ {
+ "epoch": 26.279232111692846,
+ "grad_norm": 0.056899815797805786,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 3.917966842651367,
+ "step": 1892
+ },
+ {
+ "epoch": 26.293193717277486,
+ "grad_norm": 0.060314927250146866,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 3.9230940341949463,
+ "step": 1893
+ },
+ {
+ "epoch": 26.30715532286213,
+ "grad_norm": 0.05449546128511429,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 3.9082441329956055,
+ "step": 1894
+ },
+ {
+ "epoch": 26.32111692844677,
+ "grad_norm": 0.05601051449775696,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 3.9174978733062744,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335078534031414,
+ "grad_norm": 0.05770742893218994,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 3.9183058738708496,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349040139616054,
+ "grad_norm": 0.05718272924423218,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 3.9438772201538086,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363001745200698,
+ "grad_norm": 0.052873238921165466,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 3.9554827213287354,
+ "step": 1898
+ },
+ {
+ "epoch": 26.376963350785342,
+ "grad_norm": 0.056853555142879486,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 3.886812210083008,
+ "step": 1899
+ },
+ {
+ "epoch": 26.390924956369982,
+ "grad_norm": 0.05662025138735771,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 3.9384284019470215,
+ "step": 1900
+ },
+ {
+ "epoch": 26.404886561954626,
+ "grad_norm": 0.05693329498171806,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 3.9123764038085938,
+ "step": 1901
+ },
+ {
+ "epoch": 26.418848167539267,
+ "grad_norm": 0.05425408110022545,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 3.9441678524017334,
+ "step": 1902
+ },
+ {
+ "epoch": 26.43280977312391,
+ "grad_norm": 0.05533420667052269,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 3.9086289405822754,
+ "step": 1903
+ },
+ {
+ "epoch": 26.44677137870855,
+ "grad_norm": 0.05527901649475098,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 3.9469594955444336,
+ "step": 1904
+ },
+ {
+ "epoch": 26.460732984293195,
+ "grad_norm": 0.052982039749622345,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 3.9182326793670654,
+ "step": 1905
+ },
+ {
+ "epoch": 26.474694589877835,
+ "grad_norm": 0.05520346015691757,
+ "learning_rate": 0.000288343693342466,
+ "loss": 3.954651117324829,
+ "step": 1906
+ },
+ {
+ "epoch": 26.48865619546248,
+ "grad_norm": 0.05564865842461586,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 3.951840400695801,
+ "step": 1907
+ },
+ {
+ "epoch": 26.50261780104712,
+ "grad_norm": 0.05588578060269356,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 3.933762550354004,
+ "step": 1908
+ },
+ {
+ "epoch": 26.516579406631763,
+ "grad_norm": 0.05727043375372887,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 3.9299421310424805,
+ "step": 1909
+ },
+ {
+ "epoch": 26.530541012216403,
+ "grad_norm": 0.05533654987812042,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 3.941525936126709,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544502617801047,
+ "grad_norm": 0.05524713918566704,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 3.935666084289551,
+ "step": 1911
+ },
+ {
+ "epoch": 26.55846422338569,
+ "grad_norm": 0.055435653775930405,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 3.951631546020508,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57242582897033,
+ "grad_norm": 0.05678093433380127,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 3.960881233215332,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586387434554975,
+ "grad_norm": 0.054896436631679535,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 3.9262163639068604,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600349040139616,
+ "grad_norm": 0.055213335901498795,
+ "learning_rate": 0.000285947841605349,
+ "loss": 3.9284324645996094,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61431064572426,
+ "grad_norm": 0.05644496902823448,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 3.9727559089660645,
+ "step": 1916
+ },
+ {
+ "epoch": 26.6282722513089,
+ "grad_norm": 0.05483568087220192,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 3.9410910606384277,
+ "step": 1917
+ },
+ {
+ "epoch": 26.642233856893544,
+ "grad_norm": 0.054636143147945404,
+ "learning_rate": 0.000285149463934261,
+ "loss": 3.945554733276367,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656195462478184,
+ "grad_norm": 0.056062210351228714,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 3.950249671936035,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670157068062828,
+ "grad_norm": 0.05734739080071449,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 3.9527649879455566,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68411867364747,
+ "grad_norm": 0.05562867224216461,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 3.9563350677490234,
+ "step": 1921
+ },
+ {
+ "epoch": 26.698080279232112,
+ "grad_norm": 0.05565498396754265,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 3.9478225708007812,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712041884816752,
+ "grad_norm": 0.057103481143713,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 3.9302189350128174,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726003490401396,
+ "grad_norm": 0.058241210877895355,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 3.9773902893066406,
+ "step": 1924
+ },
+ {
+ "epoch": 26.739965095986037,
+ "grad_norm": 0.055981557816267014,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 3.923896074295044,
+ "step": 1925
+ },
+ {
+ "epoch": 26.75392670157068,
+ "grad_norm": 0.05711016431450844,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 3.972334861755371,
+ "step": 1926
+ },
+ {
+ "epoch": 26.767888307155324,
+ "grad_norm": 0.05707632377743721,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 3.958393096923828,
+ "step": 1927
+ },
+ {
+ "epoch": 26.781849912739965,
+ "grad_norm": 0.055073633790016174,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 3.9449703693389893,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79581151832461,
+ "grad_norm": 0.05765518546104431,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 3.9442358016967773,
+ "step": 1929
+ },
+ {
+ "epoch": 26.80977312390925,
+ "grad_norm": 0.05692309886217117,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 3.9665679931640625,
+ "step": 1930
+ },
+ {
+ "epoch": 26.823734729493893,
+ "grad_norm": 0.05776236951351166,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 3.9153542518615723,
+ "step": 1931
+ },
+ {
+ "epoch": 26.837696335078533,
+ "grad_norm": 0.05465872213244438,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 3.919874429702759,
+ "step": 1932
+ },
+ {
+ "epoch": 26.851657940663177,
+ "grad_norm": 0.057827629148960114,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 3.9578394889831543,
+ "step": 1933
+ },
+ {
+ "epoch": 26.865619546247817,
+ "grad_norm": 0.055902253836393356,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 3.995711326599121,
+ "step": 1934
+ },
+ {
+ "epoch": 26.87958115183246,
+ "grad_norm": 0.05868638679385185,
+ "learning_rate": 0.000280627938758204,
+ "loss": 3.96519136428833,
+ "step": 1935
+ },
+ {
+ "epoch": 26.8935427574171,
+ "grad_norm": 0.059722673147916794,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 3.9566941261291504,
+ "step": 1936
+ },
+ {
+ "epoch": 26.907504363001745,
+ "grad_norm": 0.055209770798683167,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 3.9929540157318115,
+ "step": 1937
+ },
+ {
+ "epoch": 26.921465968586386,
+ "grad_norm": 0.05694293603301048,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 3.9529929161071777,
+ "step": 1938
+ },
+ {
+ "epoch": 26.93542757417103,
+ "grad_norm": 0.058492012321949005,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 3.9465792179107666,
+ "step": 1939
+ },
+ {
+ "epoch": 26.949389179755673,
+ "grad_norm": 0.056036073714494705,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 3.9590697288513184,
+ "step": 1940
+ },
+ {
+ "epoch": 26.963350785340314,
+ "grad_norm": 0.0547766387462616,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 3.969907522201538,
+ "step": 1941
+ },
+ {
+ "epoch": 26.977312390924958,
+ "grad_norm": 0.05592350661754608,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 3.9509897232055664,
+ "step": 1942
+ },
+ {
+ "epoch": 26.991273996509598,
+ "grad_norm": 0.05589321628212929,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 3.9648609161376953,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.03934016078710556,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 2.4784152507781982,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 0.6035652160644531,
+ "eval_runtime": 60.0656,
+ "eval_samples_per_second": 40.656,
+ "eval_steps_per_second": 0.649,
+ "step": 1944
+ },
+ {
+ "epoch": 27.013961605584644,
+ "grad_norm": 0.05773136392235756,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 3.8625826835632324,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027923211169284,
+ "grad_norm": 0.05844702199101448,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 3.8796658515930176,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041884816753928,
+ "grad_norm": 0.05718810483813286,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 3.8808112144470215,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05584642233857,
+ "grad_norm": 0.05264506861567497,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 3.8776121139526367,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069808027923212,
+ "grad_norm": 0.055468373000621796,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 3.866316318511963,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083769633507853,
+ "grad_norm": 0.053565360605716705,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 3.8870849609375,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097731239092496,
+ "grad_norm": 0.05468900501728058,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 3.8581550121307373,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111692844677137,
+ "grad_norm": 0.05454741418361664,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 3.8752541542053223,
+ "step": 1952
+ },
+ {
+ "epoch": 27.12565445026178,
+ "grad_norm": 0.05647914111614227,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 3.8818912506103516,
+ "step": 1953
+ },
+ {
+ "epoch": 27.13961605584642,
+ "grad_norm": 0.05325528234243393,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 3.8562560081481934,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153577661431065,
+ "grad_norm": 0.05581580847501755,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 3.879669189453125,
+ "step": 1955
+ },
+ {
+ "epoch": 27.167539267015705,
+ "grad_norm": 0.058219097554683685,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 3.8530635833740234,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18150087260035,
+ "grad_norm": 0.05613216012716293,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 3.8905630111694336,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195462478184993,
+ "grad_norm": 0.05486270785331726,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 3.8945157527923584,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209424083769633,
+ "grad_norm": 0.05338508263230324,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 3.909775733947754,
+ "step": 1959
+ },
+ {
+ "epoch": 27.223385689354277,
+ "grad_norm": 0.05602312833070755,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 3.8643651008605957,
+ "step": 1960
+ },
+ {
+ "epoch": 27.237347294938917,
+ "grad_norm": 0.05373037979006767,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 3.8915491104125977,
+ "step": 1961
+ },
+ {
+ "epoch": 27.25130890052356,
+ "grad_norm": 0.05805065110325813,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 3.8836984634399414,
+ "step": 1962
+ },
+ {
+ "epoch": 27.2652705061082,
+ "grad_norm": 0.05469847097992897,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 3.8812215328216553,
+ "step": 1963
+ },
+ {
+ "epoch": 27.279232111692846,
+ "grad_norm": 0.05575987696647644,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 3.8887271881103516,
+ "step": 1964
+ },
+ {
+ "epoch": 27.293193717277486,
+ "grad_norm": 0.05577889457345009,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 3.8917694091796875,
+ "step": 1965
+ },
+ {
+ "epoch": 27.30715532286213,
+ "grad_norm": 0.05604076758027077,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 3.881073474884033,
+ "step": 1966
+ },
+ {
+ "epoch": 27.32111692844677,
+ "grad_norm": 0.055674705654382706,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 3.9303488731384277,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335078534031414,
+ "grad_norm": 0.05300940200686455,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 3.871678113937378,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349040139616054,
+ "grad_norm": 0.05647176504135132,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 3.8848423957824707,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363001745200698,
+ "grad_norm": 0.05647560954093933,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 3.931574583053589,
+ "step": 1970
+ },
+ {
+ "epoch": 27.376963350785342,
+ "grad_norm": 0.055338893085718155,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 3.908456563949585,
+ "step": 1971
+ },
+ {
+ "epoch": 27.390924956369982,
+ "grad_norm": 0.05508219450712204,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 3.8933262825012207,
+ "step": 1972
+ },
+ {
+ "epoch": 27.404886561954626,
+ "grad_norm": 0.05787097290158272,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 3.8908863067626953,
+ "step": 1973
+ },
+ {
+ "epoch": 27.418848167539267,
+ "grad_norm": 0.05587272346019745,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 3.9261322021484375,
+ "step": 1974
+ },
+ {
+ "epoch": 27.43280977312391,
+ "grad_norm": 0.05573277547955513,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 3.9129767417907715,
+ "step": 1975
+ },
+ {
+ "epoch": 27.44677137870855,
+ "grad_norm": 0.05628194287419319,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 3.906001091003418,
+ "step": 1976
+ },
+ {
+ "epoch": 27.460732984293195,
+ "grad_norm": 0.055410996079444885,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 3.9163825511932373,
+ "step": 1977
+ },
+ {
+ "epoch": 27.474694589877835,
+ "grad_norm": 0.05579163134098053,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 3.9286627769470215,
+ "step": 1978
+ },
+ {
+ "epoch": 27.48865619546248,
+ "grad_norm": 0.05793057754635811,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 3.8955917358398438,
+ "step": 1979
+ },
+ {
+ "epoch": 27.50261780104712,
+ "grad_norm": 0.05592824146151543,
+ "learning_rate": 0.00026868712586269,
+ "loss": 3.922128677368164,
+ "step": 1980
+ },
+ {
+ "epoch": 27.516579406631763,
+ "grad_norm": 0.05768568441271782,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 3.897780418395996,
+ "step": 1981
+ },
+ {
+ "epoch": 27.530541012216403,
+ "grad_norm": 0.055304404348134995,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 3.886094570159912,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544502617801047,
+ "grad_norm": 0.054745838046073914,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 3.9272961616516113,
+ "step": 1983
+ },
+ {
+ "epoch": 27.55846422338569,
+ "grad_norm": 0.05616277828812599,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 3.8794174194335938,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57242582897033,
+ "grad_norm": 0.05605581775307655,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 3.9165992736816406,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586387434554975,
+ "grad_norm": 0.057485803961753845,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 3.902085542678833,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600349040139616,
+ "grad_norm": 0.0575849711894989,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 3.92405366897583,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61431064572426,
+ "grad_norm": 0.05563105270266533,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 3.8700900077819824,
+ "step": 1988
+ },
+ {
+ "epoch": 27.6282722513089,
+ "grad_norm": 0.05615481734275818,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 3.9526686668395996,
+ "step": 1989
+ },
+ {
+ "epoch": 27.642233856893544,
+ "grad_norm": 0.056746870279312134,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 3.9000678062438965,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656195462478184,
+ "grad_norm": 0.057345934212207794,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 3.910407543182373,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670157068062828,
+ "grad_norm": 0.05434483289718628,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 3.9132578372955322,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68411867364747,
+ "grad_norm": 0.0582718588411808,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 3.9188318252563477,
+ "step": 1993
+ },
+ {
+ "epoch": 27.698080279232112,
+ "grad_norm": 0.05566178634762764,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 3.8843226432800293,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712041884816752,
+ "grad_norm": 0.05550086498260498,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 3.904834032058716,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726003490401396,
+ "grad_norm": 0.05517689883708954,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 3.9082438945770264,
+ "step": 1996
+ },
+ {
+ "epoch": 27.739965095986037,
+ "grad_norm": 0.05598316714167595,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 3.909560203552246,
+ "step": 1997
+ },
+ {
+ "epoch": 27.75392670157068,
+ "grad_norm": 0.054991044104099274,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 3.903012275695801,
+ "step": 1998
+ },
+ {
+ "epoch": 27.767888307155324,
+ "grad_norm": 0.05612454190850258,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 3.9183692932128906,
+ "step": 1999
+ },
+ {
+ "epoch": 27.781849912739965,
+ "grad_norm": 0.055587366223335266,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 3.918454647064209,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79581151832461,
+ "grad_norm": 0.05595824122428894,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 3.895993232727051,
+ "step": 2001
+ },
+ {
+ "epoch": 27.80977312390925,
+ "grad_norm": 0.054730597883462906,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 3.9305458068847656,
+ "step": 2002
+ },
+ {
+ "epoch": 27.823734729493893,
+ "grad_norm": 0.05543222650885582,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 3.9375224113464355,
+ "step": 2003
+ },
+ {
+ "epoch": 27.837696335078533,
+ "grad_norm": 0.055638376623392105,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 3.9242658615112305,
+ "step": 2004
+ },
+ {
+ "epoch": 27.851657940663177,
+ "grad_norm": 0.05635111406445503,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 3.889681339263916,
+ "step": 2005
+ },
+ {
+ "epoch": 27.865619546247817,
+ "grad_norm": 0.056509099900722504,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 3.9191479682922363,
+ "step": 2006
+ },
+ {
+ "epoch": 27.87958115183246,
+ "grad_norm": 0.055455707013607025,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 3.936049222946167,
+ "step": 2007
+ },
+ {
+ "epoch": 27.8935427574171,
+ "grad_norm": 0.05508789047598839,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 3.901210069656372,
+ "step": 2008
+ },
+ {
+ "epoch": 27.907504363001745,
+ "grad_norm": 0.05542607232928276,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 3.9348654747009277,
+ "step": 2009
+ },
+ {
+ "epoch": 27.921465968586386,
+ "grad_norm": 0.05693232640624046,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 3.940847158432007,
+ "step": 2010
+ },
+ {
+ "epoch": 27.93542757417103,
+ "grad_norm": 0.05707051232457161,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 3.9295883178710938,
+ "step": 2011
+ },
+ {
+ "epoch": 27.949389179755673,
+ "grad_norm": 0.05778150260448456,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 3.9317305088043213,
+ "step": 2012
+ },
+ {
+ "epoch": 27.963350785340314,
+ "grad_norm": 0.05821850523352623,
+ "learning_rate": 0.00025996500713765,
+ "loss": 3.9197943210601807,
+ "step": 2013
+ },
+ {
+ "epoch": 27.977312390924958,
+ "grad_norm": 0.05592802166938782,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 3.9471435546875,
+ "step": 2014
+ },
+ {
+ "epoch": 27.991273996509598,
+ "grad_norm": 0.059447627514600754,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 3.9220168590545654,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.04082372784614563,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 2.433501958847046,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 0.6050393581390381,
+ "eval_runtime": 57.4222,
+ "eval_samples_per_second": 42.527,
+ "eval_steps_per_second": 0.679,
+ "step": 2016
+ },
+ {
+ "epoch": 28.013961605584644,
+ "grad_norm": 0.059585411101579666,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 3.8270139694213867,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027923211169284,
+ "grad_norm": 0.0652594342827797,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 3.842228412628174,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041884816753928,
+ "grad_norm": 0.05607042834162712,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 3.8498799800872803,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05584642233857,
+ "grad_norm": 0.05797959491610527,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 3.8537564277648926,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069808027923212,
+ "grad_norm": 0.06011393293738365,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 3.865638256072998,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083769633507853,
+ "grad_norm": 0.05946647375822067,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 3.8406717777252197,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097731239092496,
+ "grad_norm": 0.05657580494880676,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 3.8516461849212646,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111692844677137,
+ "grad_norm": 0.06000044196844101,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 3.859069347381592,
+ "step": 2024
+ },
+ {
+ "epoch": 28.12565445026178,
+ "grad_norm": 0.05713335797190666,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 3.8558874130249023,
+ "step": 2025
+ },
+ {
+ "epoch": 28.13961605584642,
+ "grad_norm": 0.05555511265993118,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 3.882650852203369,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153577661431065,
+ "grad_norm": 0.05772312358021736,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 3.847323179244995,
+ "step": 2027
+ },
+ {
+ "epoch": 28.167539267015705,
+ "grad_norm": 0.056661207228899,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 3.8922581672668457,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18150087260035,
+ "grad_norm": 0.05714283883571625,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 3.849281072616577,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195462478184993,
+ "grad_norm": 0.05582018196582794,
+ "learning_rate": 0.000255486047794226,
+ "loss": 3.8659892082214355,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209424083769633,
+ "grad_norm": 0.05674750730395317,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 3.8591721057891846,
+ "step": 2031
+ },
+ {
+ "epoch": 28.223385689354277,
+ "grad_norm": 0.060421206057071686,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 3.8599629402160645,
+ "step": 2032
+ },
+ {
+ "epoch": 28.237347294938917,
+ "grad_norm": 0.05727452412247658,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 3.8559670448303223,
+ "step": 2033
+ },
+ {
+ "epoch": 28.25130890052356,
+ "grad_norm": 0.05708890035748482,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 3.8643712997436523,
+ "step": 2034
+ },
+ {
+ "epoch": 28.2652705061082,
+ "grad_norm": 0.05768970027565956,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 3.842012405395508,
+ "step": 2035
+ },
+ {
+ "epoch": 28.279232111692846,
+ "grad_norm": 0.05973518267273903,
+ "learning_rate": 0.000253907826333243,
+ "loss": 3.841464042663574,
+ "step": 2036
+ },
+ {
+ "epoch": 28.293193717277486,
+ "grad_norm": 0.057536788284778595,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 3.8655552864074707,
+ "step": 2037
+ },
+ {
+ "epoch": 28.30715532286213,
+ "grad_norm": 0.05864274501800537,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 3.852839469909668,
+ "step": 2038
+ },
+ {
+ "epoch": 28.32111692844677,
+ "grad_norm": 0.056395065039396286,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 3.859243392944336,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335078534031414,
+ "grad_norm": 0.05933082476258278,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 3.843926191329956,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349040139616054,
+ "grad_norm": 0.05862351134419441,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 3.8284354209899902,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363001745200698,
+ "grad_norm": 0.05564470961689949,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 3.8802285194396973,
+ "step": 2042
+ },
+ {
+ "epoch": 28.376963350785342,
+ "grad_norm": 0.06139146164059639,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 3.867100954055786,
+ "step": 2043
+ },
+ {
+ "epoch": 28.390924956369982,
+ "grad_norm": 0.058729205280542374,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 3.8322975635528564,
+ "step": 2044
+ },
+ {
+ "epoch": 28.404886561954626,
+ "grad_norm": 0.05771554261445999,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 3.855440616607666,
+ "step": 2045
+ },
+ {
+ "epoch": 28.418848167539267,
+ "grad_norm": 0.06000557541847229,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 3.8510828018188477,
+ "step": 2046
+ },
+ {
+ "epoch": 28.43280977312391,
+ "grad_norm": 0.05935550108551979,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 3.8572049140930176,
+ "step": 2047
+ },
+ {
+ "epoch": 28.44677137870855,
+ "grad_norm": 0.056745223701000214,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 3.8892972469329834,
+ "step": 2048
+ },
+ {
+ "epoch": 28.460732984293195,
+ "grad_norm": 0.0588800348341465,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 3.865957260131836,
+ "step": 2049
+ },
+ {
+ "epoch": 28.474694589877835,
+ "grad_norm": 0.05756748467683792,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 3.8530519008636475,
+ "step": 2050
+ },
+ {
+ "epoch": 28.48865619546248,
+ "grad_norm": 0.057398825883865356,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 3.9022018909454346,
+ "step": 2051
+ },
+ {
+ "epoch": 28.50261780104712,
+ "grad_norm": 0.05809267610311508,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 3.853647232055664,
+ "step": 2052
+ },
+ {
+ "epoch": 28.516579406631763,
+ "grad_norm": 0.05625335872173309,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 3.890363931655884,
+ "step": 2053
+ },
+ {
+ "epoch": 28.530541012216403,
+ "grad_norm": 0.05852413922548294,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 3.8836288452148438,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544502617801047,
+ "grad_norm": 0.05545232817530632,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 3.902006149291992,
+ "step": 2055
+ },
+ {
+ "epoch": 28.55846422338569,
+ "grad_norm": 0.05776599794626236,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 3.8910727500915527,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57242582897033,
+ "grad_norm": 0.05847984179854393,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 3.890106201171875,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586387434554975,
+ "grad_norm": 0.056724466383457184,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 3.836899995803833,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600349040139616,
+ "grad_norm": 0.05690842866897583,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 3.834292411804199,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61431064572426,
+ "grad_norm": 0.058501504361629486,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 3.891890048980713,
+ "step": 2060
+ },
+ {
+ "epoch": 28.6282722513089,
+ "grad_norm": 0.059110380709171295,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 3.883753538131714,
+ "step": 2061
+ },
+ {
+ "epoch": 28.642233856893544,
+ "grad_norm": 0.0574255995452404,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 3.8829569816589355,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656195462478184,
+ "grad_norm": 0.05842957645654678,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 3.845698118209839,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670157068062828,
+ "grad_norm": 0.05536387115716934,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 3.9008209705352783,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68411867364747,
+ "grad_norm": 0.05769932642579079,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 3.890972137451172,
+ "step": 2065
+ },
+ {
+ "epoch": 28.698080279232112,
+ "grad_norm": 0.058036014437675476,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 3.873745918273926,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712041884816752,
+ "grad_norm": 0.05829216167330742,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 3.879685401916504,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726003490401396,
+ "grad_norm": 0.06169239804148674,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 3.9156579971313477,
+ "step": 2068
+ },
+ {
+ "epoch": 28.739965095986037,
+ "grad_norm": 0.05861479043960571,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 3.876373052597046,
+ "step": 2069
+ },
+ {
+ "epoch": 28.75392670157068,
+ "grad_norm": 0.05878114700317383,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 3.918684959411621,
+ "step": 2070
+ },
+ {
+ "epoch": 28.767888307155324,
+ "grad_norm": 0.059904664754867554,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 3.8893160820007324,
+ "step": 2071
+ },
+ {
+ "epoch": 28.781849912739965,
+ "grad_norm": 0.057385317981243134,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 3.880828619003296,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79581151832461,
+ "grad_norm": 0.05815180763602257,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 3.8638038635253906,
+ "step": 2073
+ },
+ {
+ "epoch": 28.80977312390925,
+ "grad_norm": 0.05871085450053215,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 3.8427436351776123,
+ "step": 2074
+ },
+ {
+ "epoch": 28.823734729493893,
+ "grad_norm": 0.056275319308042526,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 3.858442544937134,
+ "step": 2075
+ },
+ {
+ "epoch": 28.837696335078533,
+ "grad_norm": 0.05850474536418915,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 3.871570110321045,
+ "step": 2076
+ },
+ {
+ "epoch": 28.851657940663177,
+ "grad_norm": 0.0580497644841671,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 3.89687442779541,
+ "step": 2077
+ },
+ {
+ "epoch": 28.865619546247817,
+ "grad_norm": 0.05944157764315605,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 3.8877720832824707,
+ "step": 2078
+ },
+ {
+ "epoch": 28.87958115183246,
+ "grad_norm": 0.05749465152621269,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 3.8787431716918945,
+ "step": 2079
+ },
+ {
+ "epoch": 28.8935427574171,
+ "grad_norm": 0.05645301938056946,
+ "learning_rate": 0.000242380656502223,
+ "loss": 3.9244091510772705,
+ "step": 2080
+ },
+ {
+ "epoch": 28.907504363001745,
+ "grad_norm": 0.05700884759426117,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 3.9192817211151123,
+ "step": 2081
+ },
+ {
+ "epoch": 28.921465968586386,
+ "grad_norm": 0.05823809653520584,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 3.905806064605713,
+ "step": 2082
+ },
+ {
+ "epoch": 28.93542757417103,
+ "grad_norm": 0.05736592411994934,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 3.8907763957977295,
+ "step": 2083
+ },
+ {
+ "epoch": 28.949389179755673,
+ "grad_norm": 0.05646584555506706,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 3.892714500427246,
+ "step": 2084
+ },
+ {
+ "epoch": 28.963350785340314,
+ "grad_norm": 0.06081141158938408,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 3.864286184310913,
+ "step": 2085
+ },
+ {
+ "epoch": 28.977312390924958,
+ "grad_norm": 0.05633515864610672,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 3.8791451454162598,
+ "step": 2086
+ },
+ {
+ "epoch": 28.991273996509598,
+ "grad_norm": 0.05825566127896309,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 3.8895483016967773,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.042954832315444946,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 2.408613681793213,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 0.6065697073936462,
+ "eval_runtime": 59.1276,
+ "eval_samples_per_second": 41.301,
+ "eval_steps_per_second": 0.66,
+ "step": 2088
+ },
+ {
+ "epoch": 29.013961605584644,
+ "grad_norm": 0.06000809744000435,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 3.8419973850250244,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027923211169284,
+ "grad_norm": 0.06233612447977066,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 3.83025860786438,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041884816753928,
+ "grad_norm": 0.0551699623465538,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 3.8283262252807617,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05584642233857,
+ "grad_norm": 0.06225701421499252,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 3.822310447692871,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069808027923212,
+ "grad_norm": 0.05747336149215698,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 3.8208513259887695,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083769633507853,
+ "grad_norm": 0.0582275427877903,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 3.8013901710510254,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097731239092496,
+ "grad_norm": 0.059701140969991684,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 3.829238176345825,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111692844677137,
+ "grad_norm": 0.057550523430109024,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 3.7963826656341553,
+ "step": 2096
+ },
+ {
+ "epoch": 29.12565445026178,
+ "grad_norm": 0.05855116620659828,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 3.824028491973877,
+ "step": 2097
+ },
+ {
+ "epoch": 29.13961605584642,
+ "grad_norm": 0.056533779948949814,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 3.8151779174804688,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153577661431065,
+ "grad_norm": 0.0573289655148983,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 3.7990314960479736,
+ "step": 2099
+ },
+ {
+ "epoch": 29.167539267015705,
+ "grad_norm": 0.060061171650886536,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 3.8221707344055176,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18150087260035,
+ "grad_norm": 0.05851743370294571,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 3.793701648712158,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195462478184993,
+ "grad_norm": 0.059191569685935974,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 3.8194010257720947,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209424083769633,
+ "grad_norm": 0.05996483936905861,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 3.8316292762756348,
+ "step": 2103
+ },
+ {
+ "epoch": 29.223385689354277,
+ "grad_norm": 0.0555657260119915,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 3.821362018585205,
+ "step": 2104
+ },
+ {
+ "epoch": 29.237347294938917,
+ "grad_norm": 0.058476634323596954,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 3.8098394870758057,
+ "step": 2105
+ },
+ {
+ "epoch": 29.25130890052356,
+ "grad_norm": 0.05669668689370155,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 3.843484878540039,
+ "step": 2106
+ },
+ {
+ "epoch": 29.2652705061082,
+ "grad_norm": 0.05648602545261383,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 3.811458110809326,
+ "step": 2107
+ },
+ {
+ "epoch": 29.279232111692846,
+ "grad_norm": 0.05574130266904831,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 3.837676525115967,
+ "step": 2108
+ },
+ {
+ "epoch": 29.293193717277486,
+ "grad_norm": 0.05538785457611084,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 3.810798168182373,
+ "step": 2109
+ },
+ {
+ "epoch": 29.30715532286213,
+ "grad_norm": 0.05477031692862511,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 3.8156371116638184,
+ "step": 2110
+ },
+ {
+ "epoch": 29.32111692844677,
+ "grad_norm": 0.05664678290486336,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 3.8472559452056885,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335078534031414,
+ "grad_norm": 0.05649891495704651,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 3.824376106262207,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349040139616054,
+ "grad_norm": 0.05665554851293564,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 3.7963147163391113,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363001745200698,
+ "grad_norm": 0.054111722856760025,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 3.819735527038574,
+ "step": 2114
+ },
+ {
+ "epoch": 29.376963350785342,
+ "grad_norm": 0.05805648863315582,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 3.790189743041992,
+ "step": 2115
+ },
+ {
+ "epoch": 29.390924956369982,
+ "grad_norm": 0.05634011700749397,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 3.852161169052124,
+ "step": 2116
+ },
+ {
+ "epoch": 29.404886561954626,
+ "grad_norm": 0.0555867962539196,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 3.846884250640869,
+ "step": 2117
+ },
+ {
+ "epoch": 29.418848167539267,
+ "grad_norm": 0.057229746133089066,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 3.8432483673095703,
+ "step": 2118
+ },
+ {
+ "epoch": 29.43280977312391,
+ "grad_norm": 0.05664287880063057,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 3.842709541320801,
+ "step": 2119
+ },
+ {
+ "epoch": 29.44677137870855,
+ "grad_norm": 0.0567023791372776,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 3.844529628753662,
+ "step": 2120
+ },
+ {
+ "epoch": 29.460732984293195,
+ "grad_norm": 0.057870302349328995,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 3.853799343109131,
+ "step": 2121
+ },
+ {
+ "epoch": 29.474694589877835,
+ "grad_norm": 0.05681667849421501,
+ "learning_rate": 0.000231465389734324,
+ "loss": 3.849405288696289,
+ "step": 2122
+ },
+ {
+ "epoch": 29.48865619546248,
+ "grad_norm": 0.05738653615117073,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 3.8083410263061523,
+ "step": 2123
+ },
+ {
+ "epoch": 29.50261780104712,
+ "grad_norm": 0.058900732547044754,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 3.840458869934082,
+ "step": 2124
+ },
+ {
+ "epoch": 29.516579406631763,
+ "grad_norm": 0.05823393166065216,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 3.821415662765503,
+ "step": 2125
+ },
+ {
+ "epoch": 29.530541012216403,
+ "grad_norm": 0.057413019239902496,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 3.8713812828063965,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544502617801047,
+ "grad_norm": 0.05935502052307129,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 3.8488903045654297,
+ "step": 2127
+ },
+ {
+ "epoch": 29.55846422338569,
+ "grad_norm": 0.055285241454839706,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 3.8196067810058594,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57242582897033,
+ "grad_norm": 0.05894080176949501,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 3.8444557189941406,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586387434554975,
+ "grad_norm": 0.05878668278455734,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 3.8731908798217773,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600349040139616,
+ "grad_norm": 0.05855663865804672,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 3.831028938293457,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61431064572426,
+ "grad_norm": 0.057419586926698685,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 3.849435806274414,
+ "step": 2132
+ },
+ {
+ "epoch": 29.6282722513089,
+ "grad_norm": 0.058863408863544464,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 3.8495519161224365,
+ "step": 2133
+ },
+ {
+ "epoch": 29.642233856893544,
+ "grad_norm": 0.06081194058060646,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 3.844421863555908,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656195462478184,
+ "grad_norm": 0.05597590655088425,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 3.8413243293762207,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670157068062828,
+ "grad_norm": 0.058956339955329895,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 3.861703872680664,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68411867364747,
+ "grad_norm": 0.05883771926164627,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 3.8756394386291504,
+ "step": 2137
+ },
+ {
+ "epoch": 29.698080279232112,
+ "grad_norm": 0.06036122143268585,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 3.855403423309326,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712041884816752,
+ "grad_norm": 0.061243657022714615,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 3.861691474914551,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726003490401396,
+ "grad_norm": 0.05917691811919212,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 3.870516777038574,
+ "step": 2140
+ },
+ {
+ "epoch": 29.739965095986037,
+ "grad_norm": 0.06157418712973595,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 3.8292267322540283,
+ "step": 2141
+ },
+ {
+ "epoch": 29.75392670157068,
+ "grad_norm": 0.0605802945792675,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 3.836577892303467,
+ "step": 2142
+ },
+ {
+ "epoch": 29.767888307155324,
+ "grad_norm": 0.05750525742769241,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 3.8472204208374023,
+ "step": 2143
+ },
+ {
+ "epoch": 29.781849912739965,
+ "grad_norm": 0.057982541620731354,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 3.8682217597961426,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79581151832461,
+ "grad_norm": 0.057891640812158585,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 3.8397727012634277,
+ "step": 2145
+ },
+ {
+ "epoch": 29.80977312390925,
+ "grad_norm": 0.058175452053546906,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 3.8578720092773438,
+ "step": 2146
+ },
+ {
+ "epoch": 29.823734729493893,
+ "grad_norm": 0.05686056986451149,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 3.8467812538146973,
+ "step": 2147
+ },
+ {
+ "epoch": 29.837696335078533,
+ "grad_norm": 0.05755440890789032,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 3.844052314758301,
+ "step": 2148
+ },
+ {
+ "epoch": 29.851657940663177,
+ "grad_norm": 0.057706039398908615,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 3.818631172180176,
+ "step": 2149
+ },
+ {
+ "epoch": 29.865619546247817,
+ "grad_norm": 0.058824241161346436,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 3.8732900619506836,
+ "step": 2150
+ },
+ {
+ "epoch": 29.87958115183246,
+ "grad_norm": 0.058583300560712814,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 3.8373632431030273,
+ "step": 2151
+ },
+ {
+ "epoch": 29.8935427574171,
+ "grad_norm": 0.057403020560741425,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 3.8761391639709473,
+ "step": 2152
+ },
+ {
+ "epoch": 29.907504363001745,
+ "grad_norm": 0.0586736761033535,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 3.8518989086151123,
+ "step": 2153
+ },
+ {
+ "epoch": 29.921465968586386,
+ "grad_norm": 0.05727249011397362,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 3.839515209197998,
+ "step": 2154
+ },
+ {
+ "epoch": 29.93542757417103,
+ "grad_norm": 0.0594087690114975,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 3.834868907928467,
+ "step": 2155
+ },
+ {
+ "epoch": 29.949389179755673,
+ "grad_norm": 0.057742297649383545,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 3.861978530883789,
+ "step": 2156
+ },
+ {
+ "epoch": 29.963350785340314,
+ "grad_norm": 0.05707386136054993,
+ "learning_rate": 0.00022244633283095,
+ "loss": 3.886831760406494,
+ "step": 2157
+ },
+ {
+ "epoch": 29.977312390924958,
+ "grad_norm": 0.05631854757666588,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 3.83940052986145,
+ "step": 2158
+ },
+ {
+ "epoch": 29.991273996509598,
+ "grad_norm": 0.057860348373651505,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 3.8477416038513184,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.04080710560083389,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 2.4169907569885254,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 0.6078308820724487,
+ "eval_runtime": 59.4599,
+ "eval_samples_per_second": 41.07,
+ "eval_steps_per_second": 0.656,
+ "step": 2160
+ },
+ {
+ "epoch": 30.013961605584644,
+ "grad_norm": 0.05596645176410675,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 3.798748016357422,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027923211169284,
+ "grad_norm": 0.054763782769441605,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 3.7787699699401855,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041884816753928,
+ "grad_norm": 0.053927499800920486,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 3.779787063598633,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05584642233857,
+ "grad_norm": 0.05513143539428711,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 3.782046318054199,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069808027923212,
+ "grad_norm": 0.055234167724847794,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 3.8064064979553223,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083769633507853,
+ "grad_norm": 0.05320988595485687,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 3.8084359169006348,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097731239092496,
+ "grad_norm": 0.05671925097703934,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 3.7805888652801514,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111692844677137,
+ "grad_norm": 0.05502697452902794,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 3.791616916656494,
+ "step": 2168
+ },
+ {
+ "epoch": 30.12565445026178,
+ "grad_norm": 0.055426280945539474,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 3.7947494983673096,
+ "step": 2169
+ },
+ {
+ "epoch": 30.13961605584642,
+ "grad_norm": 0.05564495921134949,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 3.7958974838256836,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153577661431065,
+ "grad_norm": 0.055089402943849564,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 3.783299446105957,
+ "step": 2171
+ },
+ {
+ "epoch": 30.167539267015705,
+ "grad_norm": 0.055240437388420105,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 3.769829750061035,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18150087260035,
+ "grad_norm": 0.055886462330818176,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 3.805971622467041,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195462478184993,
+ "grad_norm": 0.05560130998492241,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 3.801429271697998,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209424083769633,
+ "grad_norm": 0.054780587553977966,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 3.794417381286621,
+ "step": 2175
+ },
+ {
+ "epoch": 30.223385689354277,
+ "grad_norm": 0.05751654878258705,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 3.7895469665527344,
+ "step": 2176
+ },
+ {
+ "epoch": 30.237347294938917,
+ "grad_norm": 0.054392315447330475,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 3.786479949951172,
+ "step": 2177
+ },
+ {
+ "epoch": 30.25130890052356,
+ "grad_norm": 0.057609833776950836,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 3.809680700302124,
+ "step": 2178
+ },
+ {
+ "epoch": 30.2652705061082,
+ "grad_norm": 0.054882537573575974,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 3.7928991317749023,
+ "step": 2179
+ },
+ {
+ "epoch": 30.279232111692846,
+ "grad_norm": 0.056957781314849854,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 3.790977716445923,
+ "step": 2180
+ },
+ {
+ "epoch": 30.293193717277486,
+ "grad_norm": 0.0580085888504982,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 3.851022958755493,
+ "step": 2181
+ },
+ {
+ "epoch": 30.30715532286213,
+ "grad_norm": 0.055905163288116455,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 3.784822463989258,
+ "step": 2182
+ },
+ {
+ "epoch": 30.32111692844677,
+ "grad_norm": 0.05998636782169342,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 3.811324119567871,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335078534031414,
+ "grad_norm": 0.057061366736888885,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 3.8131840229034424,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349040139616054,
+ "grad_norm": 0.05680330842733383,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 3.806260824203491,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363001745200698,
+ "grad_norm": 0.05539876967668533,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 3.793149471282959,
+ "step": 2186
+ },
+ {
+ "epoch": 30.376963350785342,
+ "grad_norm": 0.05771343782544136,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 3.7772932052612305,
+ "step": 2187
+ },
+ {
+ "epoch": 30.390924956369982,
+ "grad_norm": 0.055599041283130646,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 3.7704858779907227,
+ "step": 2188
+ },
+ {
+ "epoch": 30.404886561954626,
+ "grad_norm": 0.05746382102370262,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 3.8296337127685547,
+ "step": 2189
+ },
+ {
+ "epoch": 30.418848167539267,
+ "grad_norm": 0.05619506537914276,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 3.8134541511535645,
+ "step": 2190
+ },
+ {
+ "epoch": 30.43280977312391,
+ "grad_norm": 0.05699177086353302,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 3.8070058822631836,
+ "step": 2191
+ },
+ {
+ "epoch": 30.44677137870855,
+ "grad_norm": 0.056783925741910934,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 3.7931368350982666,
+ "step": 2192
+ },
+ {
+ "epoch": 30.460732984293195,
+ "grad_norm": 0.055537473410367966,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 3.809480905532837,
+ "step": 2193
+ },
+ {
+ "epoch": 30.474694589877835,
+ "grad_norm": 0.057826802134513855,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 3.8422718048095703,
+ "step": 2194
+ },
+ {
+ "epoch": 30.48865619546248,
+ "grad_norm": 0.0562913753092289,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 3.803966760635376,
+ "step": 2195
+ },
+ {
+ "epoch": 30.50261780104712,
+ "grad_norm": 0.05720103159546852,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 3.767852306365967,
+ "step": 2196
+ },
+ {
+ "epoch": 30.516579406631763,
+ "grad_norm": 0.05813048407435417,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 3.8423361778259277,
+ "step": 2197
+ },
+ {
+ "epoch": 30.530541012216403,
+ "grad_norm": 0.057173099368810654,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 3.8059489727020264,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544502617801047,
+ "grad_norm": 0.059024229645729065,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 3.7705254554748535,
+ "step": 2199
+ },
+ {
+ "epoch": 30.55846422338569,
+ "grad_norm": 0.056807395070791245,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 3.8056063652038574,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57242582897033,
+ "grad_norm": 0.05868417024612427,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 3.829719066619873,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586387434554975,
+ "grad_norm": 0.06023262441158295,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 3.8277781009674072,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600349040139616,
+ "grad_norm": 0.057253073900938034,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 3.8088555335998535,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61431064572426,
+ "grad_norm": 0.05946561321616173,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 3.7907581329345703,
+ "step": 2204
+ },
+ {
+ "epoch": 30.6282722513089,
+ "grad_norm": 0.05909013748168945,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 3.812042236328125,
+ "step": 2205
+ },
+ {
+ "epoch": 30.642233856893544,
+ "grad_norm": 0.05843796953558922,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 3.814551830291748,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656195462478184,
+ "grad_norm": 0.05953318253159523,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 3.8058853149414062,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670157068062828,
+ "grad_norm": 0.0616983063519001,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 3.7986061573028564,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68411867364747,
+ "grad_norm": 0.0588211789727211,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 3.8080081939697266,
+ "step": 2209
+ },
+ {
+ "epoch": 30.698080279232112,
+ "grad_norm": 0.05937618389725685,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 3.7851452827453613,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712041884816752,
+ "grad_norm": 0.05869529768824577,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 3.7893905639648438,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726003490401396,
+ "grad_norm": 0.060255810618400574,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 3.7998719215393066,
+ "step": 2212
+ },
+ {
+ "epoch": 30.739965095986037,
+ "grad_norm": 0.05810222774744034,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 3.800131320953369,
+ "step": 2213
+ },
+ {
+ "epoch": 30.75392670157068,
+ "grad_norm": 0.057548508048057556,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 3.7940425872802734,
+ "step": 2214
+ },
+ {
+ "epoch": 30.767888307155324,
+ "grad_norm": 0.05792916938662529,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 3.8089685440063477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.781849912739965,
+ "grad_norm": 0.06035939231514931,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 3.8355417251586914,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79581151832461,
+ "grad_norm": 0.058805160224437714,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 3.8146378993988037,
+ "step": 2217
+ },
+ {
+ "epoch": 30.80977312390925,
+ "grad_norm": 0.05828496813774109,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 3.8210601806640625,
+ "step": 2218
+ },
+ {
+ "epoch": 30.823734729493893,
+ "grad_norm": 0.059257738292217255,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 3.807644844055176,
+ "step": 2219
+ },
+ {
+ "epoch": 30.837696335078533,
+ "grad_norm": 0.05937729775905609,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 3.8312151432037354,
+ "step": 2220
+ },
+ {
+ "epoch": 30.851657940663177,
+ "grad_norm": 0.058580249547958374,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 3.7979660034179688,
+ "step": 2221
+ },
+ {
+ "epoch": 30.865619546247817,
+ "grad_norm": 0.0585496611893177,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 3.827874183654785,
+ "step": 2222
+ },
+ {
+ "epoch": 30.87958115183246,
+ "grad_norm": 0.05949759483337402,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 3.808143138885498,
+ "step": 2223
+ },
+ {
+ "epoch": 30.8935427574171,
+ "grad_norm": 0.058828581124544144,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 3.8351407051086426,
+ "step": 2224
+ },
+ {
+ "epoch": 30.907504363001745,
+ "grad_norm": 0.0600498728454113,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 3.8025779724121094,
+ "step": 2225
+ },
+ {
+ "epoch": 30.921465968586386,
+ "grad_norm": 0.06004577502608299,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 3.8157856464385986,
+ "step": 2226
+ },
+ {
+ "epoch": 30.93542757417103,
+ "grad_norm": 0.057676251977682114,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 3.8136918544769287,
+ "step": 2227
+ },
+ {
+ "epoch": 30.949389179755673,
+ "grad_norm": 0.0627770945429802,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 3.8470370769500732,
+ "step": 2228
+ },
+ {
+ "epoch": 30.963350785340314,
+ "grad_norm": 0.058836981654167175,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 3.8422586917877197,
+ "step": 2229
+ },
+ {
+ "epoch": 30.977312390924958,
+ "grad_norm": 0.060720525681972504,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 3.7992711067199707,
+ "step": 2230
+ },
+ {
+ "epoch": 30.991273996509598,
+ "grad_norm": 0.05844535678625107,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 3.823277473449707,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.0424773246049881,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 2.3872625827789307,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 0.6090936064720154,
+ "eval_runtime": 58.7367,
+ "eval_samples_per_second": 41.575,
+ "eval_steps_per_second": 0.664,
+ "step": 2232
+ },
+ {
+ "epoch": 31.013961605584644,
+ "grad_norm": 0.057239554822444916,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 3.764132022857666,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027923211169284,
+ "grad_norm": 0.05666050314903259,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 3.767040729522705,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041884816753928,
+ "grad_norm": 0.054038409143686295,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 3.7878732681274414,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05584642233857,
+ "grad_norm": 0.056939832866191864,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 3.7827179431915283,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069808027923212,
+ "grad_norm": 0.05488067492842674,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 3.7544898986816406,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083769633507853,
+ "grad_norm": 0.05604229122400284,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 3.7706053256988525,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097731239092496,
+ "grad_norm": 0.0550883486866951,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 3.755401611328125,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111692844677137,
+ "grad_norm": 0.055959347635507584,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 3.741966724395752,
+ "step": 2240
+ },
+ {
+ "epoch": 31.12565445026178,
+ "grad_norm": 0.05733809620141983,
+ "learning_rate": 0.000201141724176723,
+ "loss": 3.7537970542907715,
+ "step": 2241
+ },
+ {
+ "epoch": 31.13961605584642,
+ "grad_norm": 0.0551145002245903,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 3.763530731201172,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153577661431065,
+ "grad_norm": 0.055499665439128876,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 3.747159481048584,
+ "step": 2243
+ },
+ {
+ "epoch": 31.167539267015705,
+ "grad_norm": 0.0553385354578495,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 3.769044876098633,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18150087260035,
+ "grad_norm": 0.05678795278072357,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 3.7528388500213623,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195462478184993,
+ "grad_norm": 0.055931974202394485,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 3.742429733276367,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209424083769633,
+ "grad_norm": 0.05553249269723892,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 3.7760162353515625,
+ "step": 2247
+ },
+ {
+ "epoch": 31.223385689354277,
+ "grad_norm": 0.05742773041129112,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 3.7682301998138428,
+ "step": 2248
+ },
+ {
+ "epoch": 31.237347294938917,
+ "grad_norm": 0.056315284222364426,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 3.755894660949707,
+ "step": 2249
+ },
+ {
+ "epoch": 31.25130890052356,
+ "grad_norm": 0.05695716291666031,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 3.733503580093384,
+ "step": 2250
+ },
+ {
+ "epoch": 31.2652705061082,
+ "grad_norm": 0.05714774876832962,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 3.7504918575286865,
+ "step": 2251
+ },
+ {
+ "epoch": 31.279232111692846,
+ "grad_norm": 0.058987151831388474,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 3.7579331398010254,
+ "step": 2252
+ },
+ {
+ "epoch": 31.293193717277486,
+ "grad_norm": 0.05853817239403725,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 3.781409740447998,
+ "step": 2253
+ },
+ {
+ "epoch": 31.30715532286213,
+ "grad_norm": 0.057014722377061844,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 3.763247013092041,
+ "step": 2254
+ },
+ {
+ "epoch": 31.32111692844677,
+ "grad_norm": 0.05816958099603653,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 3.7801408767700195,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335078534031414,
+ "grad_norm": 0.05724027752876282,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 3.7466397285461426,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349040139616054,
+ "grad_norm": 0.055836353451013565,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 3.7770347595214844,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363001745200698,
+ "grad_norm": 0.05831955373287201,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 3.7769415378570557,
+ "step": 2258
+ },
+ {
+ "epoch": 31.376963350785342,
+ "grad_norm": 0.056602735072374344,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 3.781635284423828,
+ "step": 2259
+ },
+ {
+ "epoch": 31.390924956369982,
+ "grad_norm": 0.05784839391708374,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 3.7605199813842773,
+ "step": 2260
+ },
+ {
+ "epoch": 31.404886561954626,
+ "grad_norm": 0.057923249900341034,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 3.7671380043029785,
+ "step": 2261
+ },
+ {
+ "epoch": 31.418848167539267,
+ "grad_norm": 0.05663329362869263,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 3.7952239513397217,
+ "step": 2262
+ },
+ {
+ "epoch": 31.43280977312391,
+ "grad_norm": 0.057000353932380676,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 3.7924318313598633,
+ "step": 2263
+ },
+ {
+ "epoch": 31.44677137870855,
+ "grad_norm": 0.059976726770401,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 3.7782416343688965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.460732984293195,
+ "grad_norm": 0.05681120976805687,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 3.7918577194213867,
+ "step": 2265
+ },
+ {
+ "epoch": 31.474694589877835,
+ "grad_norm": 0.05824654549360275,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 3.7725377082824707,
+ "step": 2266
+ },
+ {
+ "epoch": 31.48865619546248,
+ "grad_norm": 0.058513302356004715,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 3.769562244415283,
+ "step": 2267
+ },
+ {
+ "epoch": 31.50261780104712,
+ "grad_norm": 0.05720348283648491,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 3.773402214050293,
+ "step": 2268
+ },
+ {
+ "epoch": 31.516579406631763,
+ "grad_norm": 0.06029265746474266,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 3.7562055587768555,
+ "step": 2269
+ },
+ {
+ "epoch": 31.530541012216403,
+ "grad_norm": 0.05892042815685272,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 3.779987335205078,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544502617801047,
+ "grad_norm": 0.06146431341767311,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 3.749899387359619,
+ "step": 2271
+ },
+ {
+ "epoch": 31.55846422338569,
+ "grad_norm": 0.056403227150440216,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 3.763558864593506,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57242582897033,
+ "grad_norm": 0.058624617755413055,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 3.7811503410339355,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586387434554975,
+ "grad_norm": 0.05978401005268097,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 3.7820377349853516,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600349040139616,
+ "grad_norm": 0.05971162021160126,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 3.7685751914978027,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61431064572426,
+ "grad_norm": 0.05576494708657265,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 3.7588582038879395,
+ "step": 2276
+ },
+ {
+ "epoch": 31.6282722513089,
+ "grad_norm": 0.05953748896718025,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 3.7525811195373535,
+ "step": 2277
+ },
+ {
+ "epoch": 31.642233856893544,
+ "grad_norm": 0.05771994963288307,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 3.7891383171081543,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656195462478184,
+ "grad_norm": 0.05822686105966568,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 3.7933766841888428,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670157068062828,
+ "grad_norm": 0.06044064089655876,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 3.783848762512207,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68411867364747,
+ "grad_norm": 0.05953317880630493,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 3.775938034057617,
+ "step": 2281
+ },
+ {
+ "epoch": 31.698080279232112,
+ "grad_norm": 0.05928120017051697,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 3.7867183685302734,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712041884816752,
+ "grad_norm": 0.05855962261557579,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 3.7872838973999023,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726003490401396,
+ "grad_norm": 0.058542776852846146,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 3.8022043704986572,
+ "step": 2284
+ },
+ {
+ "epoch": 31.739965095986037,
+ "grad_norm": 0.059078946709632874,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 3.79337739944458,
+ "step": 2285
+ },
+ {
+ "epoch": 31.75392670157068,
+ "grad_norm": 0.05854807421565056,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 3.769434928894043,
+ "step": 2286
+ },
+ {
+ "epoch": 31.767888307155324,
+ "grad_norm": 0.058214519172906876,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 3.778744697570801,
+ "step": 2287
+ },
+ {
+ "epoch": 31.781849912739965,
+ "grad_norm": 0.05981272831559181,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 3.8015875816345215,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79581151832461,
+ "grad_norm": 0.05833584442734718,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 3.7718634605407715,
+ "step": 2289
+ },
+ {
+ "epoch": 31.80977312390925,
+ "grad_norm": 0.058713216334581375,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 3.759026050567627,
+ "step": 2290
+ },
+ {
+ "epoch": 31.823734729493893,
+ "grad_norm": 0.06085464730858803,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 3.762263059616089,
+ "step": 2291
+ },
+ {
+ "epoch": 31.837696335078533,
+ "grad_norm": 0.05838872492313385,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 3.7770943641662598,
+ "step": 2292
+ },
+ {
+ "epoch": 31.851657940663177,
+ "grad_norm": 0.05941537022590637,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 3.8041505813598633,
+ "step": 2293
+ },
+ {
+ "epoch": 31.865619546247817,
+ "grad_norm": 0.058818284422159195,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 3.7724623680114746,
+ "step": 2294
+ },
+ {
+ "epoch": 31.87958115183246,
+ "grad_norm": 0.05925480276346207,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 3.7881040573120117,
+ "step": 2295
+ },
+ {
+ "epoch": 31.8935427574171,
+ "grad_norm": 0.058119483292102814,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 3.7992963790893555,
+ "step": 2296
+ },
+ {
+ "epoch": 31.907504363001745,
+ "grad_norm": 0.06139586120843887,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 3.800816297531128,
+ "step": 2297
+ },
+ {
+ "epoch": 31.921465968586386,
+ "grad_norm": 0.05812571942806244,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 3.7683157920837402,
+ "step": 2298
+ },
+ {
+ "epoch": 31.93542757417103,
+ "grad_norm": 0.05929575115442276,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 3.808281660079956,
+ "step": 2299
+ },
+ {
+ "epoch": 31.949389179755673,
+ "grad_norm": 0.057833533734083176,
+ "learning_rate": 0.000186516746349841,
+ "loss": 3.821108818054199,
+ "step": 2300
+ },
+ {
+ "epoch": 31.963350785340314,
+ "grad_norm": 0.05857503414154053,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 3.8082172870635986,
+ "step": 2301
+ },
+ {
+ "epoch": 31.977312390924958,
+ "grad_norm": 0.05893079936504364,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 3.802304267883301,
+ "step": 2302
+ },
+ {
+ "epoch": 31.991273996509598,
+ "grad_norm": 0.057888105511665344,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 3.7589073181152344,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.04306843504309654,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 2.359179735183716,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 0.6101394891738892,
+ "eval_runtime": 59.2767,
+ "eval_samples_per_second": 41.197,
+ "eval_steps_per_second": 0.658,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01396160558464,
+ "grad_norm": 0.059457872062921524,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 3.746081829071045,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02792321116929,
+ "grad_norm": 0.05826259404420853,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 3.7450389862060547,
+ "step": 2306
+ },
+ {
+ "epoch": 32.04188481675393,
+ "grad_norm": 0.05683758482336998,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 3.718315601348877,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05584642233857,
+ "grad_norm": 0.05709684267640114,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 3.7391161918640137,
+ "step": 2308
+ },
+ {
+ "epoch": 32.06980802792321,
+ "grad_norm": 0.0563763789832592,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 3.714268684387207,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083769633507856,
+ "grad_norm": 0.058482736349105835,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 3.7212839126586914,
+ "step": 2310
+ },
+ {
+ "epoch": 32.0977312390925,
+ "grad_norm": 0.05536908283829689,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 3.7573728561401367,
+ "step": 2311
+ },
+ {
+ "epoch": 32.11169284467714,
+ "grad_norm": 0.05950985103845596,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 3.715609312057495,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12565445026178,
+ "grad_norm": 0.05600661039352417,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 3.7144830226898193,
+ "step": 2313
+ },
+ {
+ "epoch": 32.139616055846425,
+ "grad_norm": 0.05814438313245773,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 3.7489242553710938,
+ "step": 2314
+ },
+ {
+ "epoch": 32.153577661431065,
+ "grad_norm": 0.0568280853331089,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 3.729698657989502,
+ "step": 2315
+ },
+ {
+ "epoch": 32.167539267015705,
+ "grad_norm": 0.05647209286689758,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 3.7484517097473145,
+ "step": 2316
+ },
+ {
+ "epoch": 32.181500872600346,
+ "grad_norm": 0.05811113119125366,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 3.7534449100494385,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19546247818499,
+ "grad_norm": 0.055132538080215454,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 3.695220470428467,
+ "step": 2318
+ },
+ {
+ "epoch": 32.20942408376963,
+ "grad_norm": 0.05722372978925705,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 3.7204041481018066,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223385689354274,
+ "grad_norm": 0.05533589795231819,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 3.7357606887817383,
+ "step": 2320
+ },
+ {
+ "epoch": 32.23734729493892,
+ "grad_norm": 0.056150201708078384,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 3.742290735244751,
+ "step": 2321
+ },
+ {
+ "epoch": 32.25130890052356,
+ "grad_norm": 0.05737832188606262,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 3.7484631538391113,
+ "step": 2322
+ },
+ {
+ "epoch": 32.2652705061082,
+ "grad_norm": 0.056017421185970306,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 3.744821071624756,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27923211169284,
+ "grad_norm": 0.05658293142914772,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 3.717287063598633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29319371727749,
+ "grad_norm": 0.05632270127534866,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 3.7172422409057617,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30715532286213,
+ "grad_norm": 0.05777734890580177,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 3.7551779747009277,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32111692844677,
+ "grad_norm": 0.05593368038535118,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 3.7148218154907227,
+ "step": 2327
+ },
+ {
+ "epoch": 32.33507853403141,
+ "grad_norm": 0.05844123661518097,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 3.7418324947357178,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34904013961606,
+ "grad_norm": 0.05616185814142227,
+ "learning_rate": 0.000179445406945268,
+ "loss": 3.7213454246520996,
+ "step": 2329
+ },
+ {
+ "epoch": 32.3630017452007,
+ "grad_norm": 0.058453164994716644,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 3.730037212371826,
+ "step": 2330
+ },
+ {
+ "epoch": 32.37696335078534,
+ "grad_norm": 0.05819960683584213,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 3.7571628093719482,
+ "step": 2331
+ },
+ {
+ "epoch": 32.390924956369986,
+ "grad_norm": 0.05889322608709335,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 3.7306394577026367,
+ "step": 2332
+ },
+ {
+ "epoch": 32.404886561954626,
+ "grad_norm": 0.05727435275912285,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 3.7304887771606445,
+ "step": 2333
+ },
+ {
+ "epoch": 32.41884816753927,
+ "grad_norm": 0.058260343968868256,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 3.714399814605713,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43280977312391,
+ "grad_norm": 0.06092117354273796,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 3.743901491165161,
+ "step": 2335
+ },
+ {
+ "epoch": 32.446771378708554,
+ "grad_norm": 0.05715249851346016,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 3.747819423675537,
+ "step": 2336
+ },
+ {
+ "epoch": 32.460732984293195,
+ "grad_norm": 0.060237716883420944,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 3.770688533782959,
+ "step": 2337
+ },
+ {
+ "epoch": 32.474694589877835,
+ "grad_norm": 0.059161923825740814,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 3.724484443664551,
+ "step": 2338
+ },
+ {
+ "epoch": 32.488656195462475,
+ "grad_norm": 0.05693944916129112,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 3.7228128910064697,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50261780104712,
+ "grad_norm": 0.05745118111371994,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 3.7756295204162598,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51657940663176,
+ "grad_norm": 0.05782153829932213,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 3.757032871246338,
+ "step": 2341
+ },
+ {
+ "epoch": 32.5305410122164,
+ "grad_norm": 0.058188699185848236,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 3.7462215423583984,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544502617801044,
+ "grad_norm": 0.05856512114405632,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 3.748447895050049,
+ "step": 2343
+ },
+ {
+ "epoch": 32.55846422338569,
+ "grad_norm": 0.05581355467438698,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 3.726031541824341,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57242582897033,
+ "grad_norm": 0.05866248160600662,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 3.7235984802246094,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58638743455497,
+ "grad_norm": 0.05732253938913345,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 3.7372875213623047,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60034904013962,
+ "grad_norm": 0.057102739810943604,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 3.768643617630005,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61431064572426,
+ "grad_norm": 0.05914342775940895,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 3.748710870742798,
+ "step": 2348
+ },
+ {
+ "epoch": 32.6282722513089,
+ "grad_norm": 0.05618954077363014,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 3.7640533447265625,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64223385689354,
+ "grad_norm": 0.05810703709721565,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 3.7230985164642334,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65619546247819,
+ "grad_norm": 0.05668202042579651,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 3.736083984375,
+ "step": 2351
+ },
+ {
+ "epoch": 32.67015706806283,
+ "grad_norm": 0.05779054015874863,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 3.7652668952941895,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68411867364747,
+ "grad_norm": 0.05945349857211113,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 3.75258207321167,
+ "step": 2353
+ },
+ {
+ "epoch": 32.69808027923211,
+ "grad_norm": 0.05697475001215935,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 3.7300591468811035,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712041884816756,
+ "grad_norm": 0.05820303037762642,
+ "learning_rate": 0.000173176617304673,
+ "loss": 3.7426252365112305,
+ "step": 2355
+ },
+ {
+ "epoch": 32.726003490401396,
+ "grad_norm": 0.05830353870987892,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 3.7716498374938965,
+ "step": 2356
+ },
+ {
+ "epoch": 32.73996509598604,
+ "grad_norm": 0.05690579116344452,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 3.7565040588378906,
+ "step": 2357
+ },
+ {
+ "epoch": 32.753926701570684,
+ "grad_norm": 0.059615254402160645,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 3.7484030723571777,
+ "step": 2358
+ },
+ {
+ "epoch": 32.767888307155324,
+ "grad_norm": 0.058043934404850006,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 3.76991868019104,
+ "step": 2359
+ },
+ {
+ "epoch": 32.781849912739965,
+ "grad_norm": 0.059192877262830734,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 3.7510385513305664,
+ "step": 2360
+ },
+ {
+ "epoch": 32.795811518324605,
+ "grad_norm": 0.05763391777873039,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 3.7479896545410156,
+ "step": 2361
+ },
+ {
+ "epoch": 32.80977312390925,
+ "grad_norm": 0.05946139991283417,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 3.7780566215515137,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82373472949389,
+ "grad_norm": 0.05837136134505272,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 3.7610299587249756,
+ "step": 2363
+ },
+ {
+ "epoch": 32.83769633507853,
+ "grad_norm": 0.05832294747233391,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 3.7216475009918213,
+ "step": 2364
+ },
+ {
+ "epoch": 32.85165794066317,
+ "grad_norm": 0.05887595936655998,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 3.767125129699707,
+ "step": 2365
+ },
+ {
+ "epoch": 32.86561954624782,
+ "grad_norm": 0.058594878762960434,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 3.751053810119629,
+ "step": 2366
+ },
+ {
+ "epoch": 32.87958115183246,
+ "grad_norm": 0.05895577743649483,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 3.7889444828033447,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8935427574171,
+ "grad_norm": 0.05831810459494591,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 3.7719945907592773,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90750436300174,
+ "grad_norm": 0.06090841069817543,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 3.7372164726257324,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92146596858639,
+ "grad_norm": 0.05728885903954506,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 3.782083511352539,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93542757417103,
+ "grad_norm": 0.061425331979990005,
+ "learning_rate": 0.00016935382741164,
+ "loss": 3.7716455459594727,
+ "step": 2371
+ },
+ {
+ "epoch": 32.94938917975567,
+ "grad_norm": 0.05963108688592911,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 3.7798497676849365,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96335078534032,
+ "grad_norm": 0.05818144232034683,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 3.746483325958252,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97731239092496,
+ "grad_norm": 0.05970250442624092,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 3.7578821182250977,
+ "step": 2374
+ },
+ {
+ "epoch": 32.9912739965096,
+ "grad_norm": 0.05898123234510422,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 3.762807846069336,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.042496208101511,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 2.353346347808838,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 0.611886203289032,
+ "eval_runtime": 58.794,
+ "eval_samples_per_second": 41.535,
+ "eval_steps_per_second": 0.663,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01396160558464,
+ "grad_norm": 0.05598964914679527,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 3.7107529640197754,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02792321116929,
+ "grad_norm": 0.057739973068237305,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 3.6825904846191406,
+ "step": 2378
+ },
+ {
+ "epoch": 33.04188481675393,
+ "grad_norm": 0.05335357040166855,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 3.684080123901367,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05584642233857,
+ "grad_norm": 0.05573783814907074,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 3.69419527053833,
+ "step": 2380
+ },
+ {
+ "epoch": 33.06980802792321,
+ "grad_norm": 0.055793728679418564,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 3.730604648590088,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083769633507856,
+ "grad_norm": 0.054736193269491196,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 3.6922354698181152,
+ "step": 2382
+ },
+ {
+ "epoch": 33.0977312390925,
+ "grad_norm": 0.05806516483426094,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 3.7343029975891113,
+ "step": 2383
+ },
+ {
+ "epoch": 33.11169284467714,
+ "grad_norm": 0.0541117899119854,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 3.7044200897216797,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12565445026178,
+ "grad_norm": 0.057081446051597595,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 3.6954641342163086,
+ "step": 2385
+ },
+ {
+ "epoch": 33.139616055846425,
+ "grad_norm": 0.05561508610844612,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 3.709800958633423,
+ "step": 2386
+ },
+ {
+ "epoch": 33.153577661431065,
+ "grad_norm": 0.05675433203577995,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 3.7057042121887207,
+ "step": 2387
+ },
+ {
+ "epoch": 33.167539267015705,
+ "grad_norm": 0.05598882585763931,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 3.7071051597595215,
+ "step": 2388
+ },
+ {
+ "epoch": 33.181500872600346,
+ "grad_norm": 0.05500704050064087,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 3.7088711261749268,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19546247818499,
+ "grad_norm": 0.057210177183151245,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 3.685708522796631,
+ "step": 2390
+ },
+ {
+ "epoch": 33.20942408376963,
+ "grad_norm": 0.05615527555346489,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 3.695154905319214,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223385689354274,
+ "grad_norm": 0.056796107441186905,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 3.7289557456970215,
+ "step": 2392
+ },
+ {
+ "epoch": 33.23734729493892,
+ "grad_norm": 0.056371185928583145,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 3.7177929878234863,
+ "step": 2393
+ },
+ {
+ "epoch": 33.25130890052356,
+ "grad_norm": 0.05646727234125137,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 3.7265491485595703,
+ "step": 2394
+ },
+ {
+ "epoch": 33.2652705061082,
+ "grad_norm": 0.05717485025525093,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 3.7198410034179688,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27923211169284,
+ "grad_norm": 0.056603338569402695,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 3.692781686782837,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29319371727749,
+ "grad_norm": 0.05612906441092491,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 3.7237164974212646,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30715532286213,
+ "grad_norm": 0.05704745277762413,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 3.698744058609009,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32111692844677,
+ "grad_norm": 0.055712319910526276,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 3.7240562438964844,
+ "step": 2399
+ },
+ {
+ "epoch": 33.33507853403141,
+ "grad_norm": 0.05763087794184685,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 3.709852695465088,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34904013961606,
+ "grad_norm": 0.05521798133850098,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 3.7102928161621094,
+ "step": 2401
+ },
+ {
+ "epoch": 33.3630017452007,
+ "grad_norm": 0.05802507698535919,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 3.697176933288574,
+ "step": 2402
+ },
+ {
+ "epoch": 33.37696335078534,
+ "grad_norm": 0.05547751113772392,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 3.7416112422943115,
+ "step": 2403
+ },
+ {
+ "epoch": 33.390924956369986,
+ "grad_norm": 0.05753115564584732,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 3.7048134803771973,
+ "step": 2404
+ },
+ {
+ "epoch": 33.404886561954626,
+ "grad_norm": 0.05642146244645119,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 3.690187931060791,
+ "step": 2405
+ },
+ {
+ "epoch": 33.41884816753927,
+ "grad_norm": 0.05780908465385437,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 3.7087855339050293,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43280977312391,
+ "grad_norm": 0.05772170051932335,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 3.7156143188476562,
+ "step": 2407
+ },
+ {
+ "epoch": 33.446771378708554,
+ "grad_norm": 0.05813417211174965,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 3.7155094146728516,
+ "step": 2408
+ },
+ {
+ "epoch": 33.460732984293195,
+ "grad_norm": 0.056307602673769,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 3.7018208503723145,
+ "step": 2409
+ },
+ {
+ "epoch": 33.474694589877835,
+ "grad_norm": 0.05811525136232376,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 3.7002341747283936,
+ "step": 2410
+ },
+ {
+ "epoch": 33.488656195462475,
+ "grad_norm": 0.056484196335077286,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 3.715003490447998,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50261780104712,
+ "grad_norm": 0.057813555002212524,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 3.744851589202881,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51657940663176,
+ "grad_norm": 0.05758954957127571,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 3.712819814682007,
+ "step": 2413
+ },
+ {
+ "epoch": 33.5305410122164,
+ "grad_norm": 0.05723990499973297,
+ "learning_rate": 0.000159218843594243,
+ "loss": 3.7235021591186523,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544502617801044,
+ "grad_norm": 0.058723341673612595,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 3.681633949279785,
+ "step": 2415
+ },
+ {
+ "epoch": 33.55846422338569,
+ "grad_norm": 0.05725013092160225,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 3.7118563652038574,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57242582897033,
+ "grad_norm": 0.05801474303007126,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 3.7071094512939453,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58638743455497,
+ "grad_norm": 0.05690481513738632,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 3.7366995811462402,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60034904013962,
+ "grad_norm": 0.05696716532111168,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 3.7125186920166016,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61431064572426,
+ "grad_norm": 0.05725441873073578,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 3.729522705078125,
+ "step": 2420
+ },
+ {
+ "epoch": 33.6282722513089,
+ "grad_norm": 0.05838564410805702,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 3.713085412979126,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64223385689354,
+ "grad_norm": 0.05807472765445709,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 3.722982883453369,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65619546247819,
+ "grad_norm": 0.05865970253944397,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 3.7254886627197266,
+ "step": 2423
+ },
+ {
+ "epoch": 33.67015706806283,
+ "grad_norm": 0.05879775062203407,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 3.7419567108154297,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68411867364747,
+ "grad_norm": 0.058425575494766235,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 3.7334160804748535,
+ "step": 2425
+ },
+ {
+ "epoch": 33.69808027923211,
+ "grad_norm": 0.05795733258128166,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 3.6824467182159424,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712041884816756,
+ "grad_norm": 0.05840831995010376,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 3.7226808071136475,
+ "step": 2427
+ },
+ {
+ "epoch": 33.726003490401396,
+ "grad_norm": 0.05915387347340584,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 3.7192320823669434,
+ "step": 2428
+ },
+ {
+ "epoch": 33.73996509598604,
+ "grad_norm": 0.05898639187216759,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 3.731309175491333,
+ "step": 2429
+ },
+ {
+ "epoch": 33.753926701570684,
+ "grad_norm": 0.058383479714393616,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 3.6850638389587402,
+ "step": 2430
+ },
+ {
+ "epoch": 33.767888307155324,
+ "grad_norm": 0.06078571453690529,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 3.7126986980438232,
+ "step": 2431
+ },
+ {
+ "epoch": 33.781849912739965,
+ "grad_norm": 0.05873129144310951,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 3.726377010345459,
+ "step": 2432
+ },
+ {
+ "epoch": 33.795811518324605,
+ "grad_norm": 0.060784872621297836,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 3.7208423614501953,
+ "step": 2433
+ },
+ {
+ "epoch": 33.80977312390925,
+ "grad_norm": 0.05992938578128815,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 3.7019124031066895,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82373472949389,
+ "grad_norm": 0.060495104640722275,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 3.7536425590515137,
+ "step": 2435
+ },
+ {
+ "epoch": 33.83769633507853,
+ "grad_norm": 0.059453774243593216,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 3.7357583045959473,
+ "step": 2436
+ },
+ {
+ "epoch": 33.85165794066317,
+ "grad_norm": 0.06152774766087532,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 3.7266006469726562,
+ "step": 2437
+ },
+ {
+ "epoch": 33.86561954624782,
+ "grad_norm": 0.05948188528418541,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 3.712332248687744,
+ "step": 2438
+ },
+ {
+ "epoch": 33.87958115183246,
+ "grad_norm": 0.05947373807430267,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 3.696099042892456,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8935427574171,
+ "grad_norm": 0.060062918812036514,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 3.7297117710113525,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90750436300174,
+ "grad_norm": 0.057695310562849045,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 3.7090835571289062,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92146596858639,
+ "grad_norm": 0.05904509872198105,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 3.763152599334717,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93542757417103,
+ "grad_norm": 0.05886756256222725,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 3.7213687896728516,
+ "step": 2443
+ },
+ {
+ "epoch": 33.94938917975567,
+ "grad_norm": 0.06159936264157295,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 3.7479395866394043,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96335078534032,
+ "grad_norm": 0.057995621114969254,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 3.7327473163604736,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97731239092496,
+ "grad_norm": 0.058990150690078735,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 3.699246883392334,
+ "step": 2446
+ },
+ {
+ "epoch": 33.9912739965096,
+ "grad_norm": 0.05924935266375542,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 3.730475425720215,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.04313633218407631,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 2.336765766143799,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 0.6134145855903625,
+ "eval_runtime": 59.4711,
+ "eval_samples_per_second": 41.062,
+ "eval_steps_per_second": 0.656,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01396160558464,
+ "grad_norm": 0.05632878839969635,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 3.667578935623169,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02792321116929,
+ "grad_norm": 0.05653175711631775,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 3.6973605155944824,
+ "step": 2450
+ },
+ {
+ "epoch": 34.04188481675393,
+ "grad_norm": 0.05482367426156998,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 3.675959587097168,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05584642233857,
+ "grad_norm": 0.05636861175298691,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 3.673849582672119,
+ "step": 2452
+ },
+ {
+ "epoch": 34.06980802792321,
+ "grad_norm": 0.05500229448080063,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 3.6991372108459473,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083769633507856,
+ "grad_norm": 0.056783828884363174,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 3.6838951110839844,
+ "step": 2454
+ },
+ {
+ "epoch": 34.0977312390925,
+ "grad_norm": 0.05543818324804306,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 3.6698226928710938,
+ "step": 2455
+ },
+ {
+ "epoch": 34.11169284467714,
+ "grad_norm": 0.05481055751442909,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 3.678802967071533,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12565445026178,
+ "grad_norm": 0.055152181535959244,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 3.706040620803833,
+ "step": 2457
+ },
+ {
+ "epoch": 34.139616055846425,
+ "grad_norm": 0.05652368441224098,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 3.669009208679199,
+ "step": 2458
+ },
+ {
+ "epoch": 34.153577661431065,
+ "grad_norm": 0.05562841519713402,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 3.6889967918395996,
+ "step": 2459
+ },
+ {
+ "epoch": 34.167539267015705,
+ "grad_norm": 0.05579658970236778,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 3.68369197845459,
+ "step": 2460
+ },
+ {
+ "epoch": 34.181500872600346,
+ "grad_norm": 0.055683478713035583,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 3.670243740081787,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19546247818499,
+ "grad_norm": 0.05624355748295784,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 3.633751392364502,
+ "step": 2462
+ },
+ {
+ "epoch": 34.20942408376963,
+ "grad_norm": 0.05518807843327522,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 3.689868927001953,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223385689354274,
+ "grad_norm": 0.05650133267045021,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 3.694890022277832,
+ "step": 2464
+ },
+ {
+ "epoch": 34.23734729493892,
+ "grad_norm": 0.0555979385972023,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 3.6806082725524902,
+ "step": 2465
+ },
+ {
+ "epoch": 34.25130890052356,
+ "grad_norm": 0.0558185949921608,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 3.6637582778930664,
+ "step": 2466
+ },
+ {
+ "epoch": 34.2652705061082,
+ "grad_norm": 0.056465502828359604,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 3.693030834197998,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27923211169284,
+ "grad_norm": 0.0570668987929821,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 3.6703829765319824,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29319371727749,
+ "grad_norm": 0.05592618137598038,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 3.6874935626983643,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30715532286213,
+ "grad_norm": 0.055560462176799774,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 3.656379222869873,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32111692844677,
+ "grad_norm": 0.056786514818668365,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 3.685967445373535,
+ "step": 2471
+ },
+ {
+ "epoch": 34.33507853403141,
+ "grad_norm": 0.055693406611680984,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 3.693864345550537,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34904013961606,
+ "grad_norm": 0.05816012993454933,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 3.7092270851135254,
+ "step": 2473
+ },
+ {
+ "epoch": 34.3630017452007,
+ "grad_norm": 0.05769353732466698,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 3.6529183387756348,
+ "step": 2474
+ },
+ {
+ "epoch": 34.37696335078534,
+ "grad_norm": 0.058298323303461075,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 3.7092885971069336,
+ "step": 2475
+ },
+ {
+ "epoch": 34.390924956369986,
+ "grad_norm": 0.055963121354579926,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 3.6982855796813965,
+ "step": 2476
+ },
+ {
+ "epoch": 34.404886561954626,
+ "grad_norm": 0.057205602526664734,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 3.690000534057617,
+ "step": 2477
+ },
+ {
+ "epoch": 34.41884816753927,
+ "grad_norm": 0.05652588978409767,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 3.6800036430358887,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43280977312391,
+ "grad_norm": 0.059187956154346466,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 3.677968978881836,
+ "step": 2479
+ },
+ {
+ "epoch": 34.446771378708554,
+ "grad_norm": 0.05786960572004318,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 3.7025253772735596,
+ "step": 2480
+ },
+ {
+ "epoch": 34.460732984293195,
+ "grad_norm": 0.06170958653092384,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 3.6797211170196533,
+ "step": 2481
+ },
+ {
+ "epoch": 34.474694589877835,
+ "grad_norm": 0.0579230934381485,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 3.6861648559570312,
+ "step": 2482
+ },
+ {
+ "epoch": 34.488656195462475,
+ "grad_norm": 0.059319235384464264,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 3.6565022468566895,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50261780104712,
+ "grad_norm": 0.05783943459391594,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 3.722566604614258,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51657940663176,
+ "grad_norm": 0.062192801386117935,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 3.6969285011291504,
+ "step": 2485
+ },
+ {
+ "epoch": 34.5305410122164,
+ "grad_norm": 0.05853543430566788,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 3.6851119995117188,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544502617801044,
+ "grad_norm": 0.05946554243564606,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 3.6720314025878906,
+ "step": 2487
+ },
+ {
+ "epoch": 34.55846422338569,
+ "grad_norm": 0.05766003578901291,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 3.665830135345459,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57242582897033,
+ "grad_norm": 0.06046074256300926,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 3.6955373287200928,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58638743455497,
+ "grad_norm": 0.057485196739435196,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 3.7058653831481934,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60034904013962,
+ "grad_norm": 0.0580572746694088,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 3.6663808822631836,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61431064572426,
+ "grad_norm": 0.05744212865829468,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 3.7024950981140137,
+ "step": 2492
+ },
+ {
+ "epoch": 34.6282722513089,
+ "grad_norm": 0.058456238359212875,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 3.6953721046447754,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64223385689354,
+ "grad_norm": 0.05971676856279373,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 3.6731884479522705,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65619546247819,
+ "grad_norm": 0.058337483555078506,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 3.7169854640960693,
+ "step": 2495
+ },
+ {
+ "epoch": 34.67015706806283,
+ "grad_norm": 0.059567082673311234,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 3.6867635250091553,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68411867364747,
+ "grad_norm": 0.05710845813155174,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 3.6897926330566406,
+ "step": 2497
+ },
+ {
+ "epoch": 34.69808027923211,
+ "grad_norm": 0.05975678935647011,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 3.699439764022827,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712041884816756,
+ "grad_norm": 0.05828933045268059,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 3.677755355834961,
+ "step": 2499
+ },
+ {
+ "epoch": 34.726003490401396,
+ "grad_norm": 0.05896460637450218,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 3.687589645385742,
+ "step": 2500
+ },
+ {
+ "epoch": 34.73996509598604,
+ "grad_norm": 0.05770917236804962,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 3.7188401222229004,
+ "step": 2501
+ },
+ {
+ "epoch": 34.753926701570684,
+ "grad_norm": 0.05971298739314079,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 3.6803629398345947,
+ "step": 2502
+ },
+ {
+ "epoch": 34.767888307155324,
+ "grad_norm": 0.05826874077320099,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 3.7006959915161133,
+ "step": 2503
+ },
+ {
+ "epoch": 34.781849912739965,
+ "grad_norm": 0.05863497033715248,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 3.680962562561035,
+ "step": 2504
+ },
+ {
+ "epoch": 34.795811518324605,
+ "grad_norm": 0.058764465153217316,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 3.7000298500061035,
+ "step": 2505
+ },
+ {
+ "epoch": 34.80977312390925,
+ "grad_norm": 0.05830385908484459,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 3.6838178634643555,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82373472949389,
+ "grad_norm": 0.05855685845017433,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 3.709963321685791,
+ "step": 2507
+ },
+ {
+ "epoch": 34.83769633507853,
+ "grad_norm": 0.05872069671750069,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 3.702852725982666,
+ "step": 2508
+ },
+ {
+ "epoch": 34.85165794066317,
+ "grad_norm": 0.05955277755856514,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 3.699869155883789,
+ "step": 2509
+ },
+ {
+ "epoch": 34.86561954624782,
+ "grad_norm": 0.061080168932676315,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 3.7028279304504395,
+ "step": 2510
+ },
+ {
+ "epoch": 34.87958115183246,
+ "grad_norm": 0.05777806416153908,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 3.6681065559387207,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8935427574171,
+ "grad_norm": 0.0598529689013958,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 3.7238996028900146,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90750436300174,
+ "grad_norm": 0.05878746509552002,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 3.696312427520752,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92146596858639,
+ "grad_norm": 0.06028176471590996,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 3.6749677658081055,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93542757417103,
+ "grad_norm": 0.05837864801287651,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 3.675783634185791,
+ "step": 2515
+ },
+ {
+ "epoch": 34.94938917975567,
+ "grad_norm": 0.060354143381118774,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 3.6795129776000977,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96335078534032,
+ "grad_norm": 0.05837022140622139,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 3.711968183517456,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97731239092496,
+ "grad_norm": 0.06035105511546135,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 3.6994686126708984,
+ "step": 2518
+ },
+ {
+ "epoch": 34.9912739965096,
+ "grad_norm": 0.05695690959692001,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 3.6764185428619385,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.04363907873630524,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 2.2967724800109863,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 0.614627480506897,
+ "eval_runtime": 59.3089,
+ "eval_samples_per_second": 41.174,
+ "eval_steps_per_second": 0.658,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01396160558464,
+ "grad_norm": 0.05742518976330757,
+ "learning_rate": 0.0001349604788260823,
+ "loss": 3.6706724166870117,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02792321116929,
+ "grad_norm": 0.05394700914621353,
+ "learning_rate": 0.00013474073268599152,
+ "loss": 3.6419858932495117,
+ "step": 2522
+ },
+ {
+ "epoch": 35.04188481675393,
+ "grad_norm": 0.05702907219529152,
+ "learning_rate": 0.00013452112210935339,
+ "loss": 3.6621429920196533,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05584642233857,
+ "grad_norm": 0.05404812470078468,
+ "learning_rate": 0.00013430164727310399,
+ "loss": 3.6589713096618652,
+ "step": 2524
+ },
+ {
+ "epoch": 35.06980802792321,
+ "grad_norm": 0.05607686936855316,
+ "learning_rate": 0.00013408230835407027,
+ "loss": 3.634030342102051,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083769633507856,
+ "grad_norm": 0.05493795499205589,
+ "learning_rate": 0.00013386310552896977,
+ "loss": 3.678600788116455,
+ "step": 2526
+ },
+ {
+ "epoch": 35.0977312390925,
+ "grad_norm": 0.0575743205845356,
+ "learning_rate": 0.00013364403897441016,
+ "loss": 3.6530723571777344,
+ "step": 2527
+ },
+ {
+ "epoch": 35.11169284467714,
+ "grad_norm": 0.05405096337199211,
+ "learning_rate": 0.00013342510886688944,
+ "loss": 3.633359432220459,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12565445026178,
+ "grad_norm": 0.055748239159584045,
+ "learning_rate": 0.00013320631538279548,
+ "loss": 3.647953987121582,
+ "step": 2529
+ },
+ {
+ "epoch": 35.139616055846425,
+ "grad_norm": 0.05661927908658981,
+ "learning_rate": 0.00013298765869840658,
+ "loss": 3.6305713653564453,
+ "step": 2530
+ },
+ {
+ "epoch": 35.153577661431065,
+ "grad_norm": 0.05812063813209534,
+ "learning_rate": 0.00013276913898989013,
+ "loss": 3.656938076019287,
+ "step": 2531
+ },
+ {
+ "epoch": 35.167539267015705,
+ "grad_norm": 0.057856228202581406,
+ "learning_rate": 0.00013255075643330369,
+ "loss": 3.657390594482422,
+ "step": 2532
+ },
+ {
+ "epoch": 35.181500872600346,
+ "grad_norm": 0.057047586888074875,
+ "learning_rate": 0.00013233251120459407,
+ "loss": 3.6266355514526367,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19546247818499,
+ "grad_norm": 0.05664423853158951,
+ "learning_rate": 0.0001321144034795976,
+ "loss": 3.6741762161254883,
+ "step": 2534
+ },
+ {
+ "epoch": 35.20942408376963,
+ "grad_norm": 0.056682784110307693,
+ "learning_rate": 0.00013189643343403977,
+ "loss": 3.6106696128845215,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223385689354274,
+ "grad_norm": 0.05825677514076233,
+ "learning_rate": 0.00013167860124353476,
+ "loss": 3.6535234451293945,
+ "step": 2536
+ },
+ {
+ "epoch": 35.23734729493892,
+ "grad_norm": 0.05540613457560539,
+ "learning_rate": 0.00013146090708358657,
+ "loss": 3.6676511764526367,
+ "step": 2537
+ },
+ {
+ "epoch": 35.25130890052356,
+ "grad_norm": 0.05705772340297699,
+ "learning_rate": 0.00013124335112958704,
+ "loss": 3.630685567855835,
+ "step": 2538
+ },
+ {
+ "epoch": 35.2652705061082,
+ "grad_norm": 0.056012462824583054,
+ "learning_rate": 0.0001310259335568172,
+ "loss": 3.6648166179656982,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27923211169284,
+ "grad_norm": 0.05761443451046944,
+ "learning_rate": 0.00013080865454044647,
+ "loss": 3.6521716117858887,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29319371727749,
+ "grad_norm": 0.05589524284005165,
+ "learning_rate": 0.00013059151425553263,
+ "loss": 3.663506031036377,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30715532286213,
+ "grad_norm": 0.056549277156591415,
+ "learning_rate": 0.00013037451287702183,
+ "loss": 3.653916835784912,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32111692844677,
+ "grad_norm": 0.056388892233371735,
+ "learning_rate": 0.0001301576505797479,
+ "loss": 3.6622371673583984,
+ "step": 2543
+ },
+ {
+ "epoch": 35.33507853403141,
+ "grad_norm": 0.05646223947405815,
+ "learning_rate": 0.00012994092753843293,
+ "loss": 3.697643756866455,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34904013961606,
+ "grad_norm": 0.055597055703401566,
+ "learning_rate": 0.00012972434392768687,
+ "loss": 3.657986640930176,
+ "step": 2545
+ },
+ {
+ "epoch": 35.3630017452007,
+ "grad_norm": 0.055099040269851685,
+ "learning_rate": 0.00012950789992200714,
+ "loss": 3.6413917541503906,
+ "step": 2546
+ },
+ {
+ "epoch": 35.37696335078534,
+ "grad_norm": 0.0568387471139431,
+ "learning_rate": 0.00012929159569577886,
+ "loss": 3.640183210372925,
+ "step": 2547
+ },
+ {
+ "epoch": 35.390924956369986,
+ "grad_norm": 0.05552329868078232,
+ "learning_rate": 0.0001290754314232743,
+ "loss": 3.6469502449035645,
+ "step": 2548
+ },
+ {
+ "epoch": 35.404886561954626,
+ "grad_norm": 0.05681881681084633,
+ "learning_rate": 0.00012885940727865334,
+ "loss": 3.6672420501708984,
+ "step": 2549
+ },
+ {
+ "epoch": 35.41884816753927,
+ "grad_norm": 0.05603218451142311,
+ "learning_rate": 0.00012864352343596247,
+ "loss": 3.6844213008880615,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43280977312391,
+ "grad_norm": 0.05711217224597931,
+ "learning_rate": 0.0001284277800691355,
+ "loss": 3.6611440181732178,
+ "step": 2551
+ },
+ {
+ "epoch": 35.446771378708554,
+ "grad_norm": 0.05441901832818985,
+ "learning_rate": 0.00012821217735199298,
+ "loss": 3.6248745918273926,
+ "step": 2552
+ },
+ {
+ "epoch": 35.460732984293195,
+ "grad_norm": 0.05754980444908142,
+ "learning_rate": 0.00012799671545824212,
+ "loss": 3.660444736480713,
+ "step": 2553
+ },
+ {
+ "epoch": 35.474694589877835,
+ "grad_norm": 0.055383119732141495,
+ "learning_rate": 0.0001277813945614768,
+ "loss": 3.6592612266540527,
+ "step": 2554
+ },
+ {
+ "epoch": 35.488656195462475,
+ "grad_norm": 0.05781600996851921,
+ "learning_rate": 0.00012756621483517682,
+ "loss": 3.65926456451416,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50261780104712,
+ "grad_norm": 0.05716165155172348,
+ "learning_rate": 0.00012735117645270905,
+ "loss": 3.6506943702697754,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51657940663176,
+ "grad_norm": 0.05843847990036011,
+ "learning_rate": 0.00012713627958732567,
+ "loss": 3.6542019844055176,
+ "step": 2557
+ },
+ {
+ "epoch": 35.5305410122164,
+ "grad_norm": 0.057625964283943176,
+ "learning_rate": 0.00012692152441216539,
+ "loss": 3.6501665115356445,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544502617801044,
+ "grad_norm": 0.0576319545507431,
+ "learning_rate": 0.0001267069111002525,
+ "loss": 3.647266387939453,
+ "step": 2559
+ },
+ {
+ "epoch": 35.55846422338569,
+ "grad_norm": 0.05821964517235756,
+ "learning_rate": 0.00012649243982449718,
+ "loss": 3.6559696197509766,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57242582897033,
+ "grad_norm": 0.05659124627709389,
+ "learning_rate": 0.00012627811075769502,
+ "loss": 3.6653411388397217,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58638743455497,
+ "grad_norm": 0.05749163031578064,
+ "learning_rate": 0.00012606392407252687,
+ "loss": 3.6690258979797363,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60034904013962,
+ "grad_norm": 0.056859105825424194,
+ "learning_rate": 0.00012584987994155943,
+ "loss": 3.639690637588501,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61431064572426,
+ "grad_norm": 0.0584794282913208,
+ "learning_rate": 0.00012563597853724388,
+ "loss": 3.6673288345336914,
+ "step": 2564
+ },
+ {
+ "epoch": 35.6282722513089,
+ "grad_norm": 0.057991158217191696,
+ "learning_rate": 0.000125422220031917,
+ "loss": 3.6846871376037598,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64223385689354,
+ "grad_norm": 0.05805192142724991,
+ "learning_rate": 0.00012520860459779972,
+ "loss": 3.639286518096924,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65619546247819,
+ "grad_norm": 0.05739975348114967,
+ "learning_rate": 0.0001249951324069986,
+ "loss": 3.6255669593811035,
+ "step": 2567
+ },
+ {
+ "epoch": 35.67015706806283,
+ "grad_norm": 0.057306256145238876,
+ "learning_rate": 0.0001247818036315042,
+ "loss": 3.6530256271362305,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68411867364747,
+ "grad_norm": 0.056031618267297745,
+ "learning_rate": 0.00012456861844319155,
+ "loss": 3.6340036392211914,
+ "step": 2569
+ },
+ {
+ "epoch": 35.69808027923211,
+ "grad_norm": 0.05747546628117561,
+ "learning_rate": 0.00012435557701382012,
+ "loss": 3.670499086380005,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712041884816756,
+ "grad_norm": 0.056736692786216736,
+ "learning_rate": 0.0001241426795150336,
+ "loss": 3.6485583782196045,
+ "step": 2571
+ },
+ {
+ "epoch": 35.726003490401396,
+ "grad_norm": 0.05779654160141945,
+ "learning_rate": 0.00012392992611835973,
+ "loss": 3.6740036010742188,
+ "step": 2572
+ },
+ {
+ "epoch": 35.73996509598604,
+ "grad_norm": 0.05827609449625015,
+ "learning_rate": 0.0001237173169952098,
+ "loss": 3.687256336212158,
+ "step": 2573
+ },
+ {
+ "epoch": 35.753926701570684,
+ "grad_norm": 0.0587199442088604,
+ "learning_rate": 0.00012350485231687954,
+ "loss": 3.660586357116699,
+ "step": 2574
+ },
+ {
+ "epoch": 35.767888307155324,
+ "grad_norm": 0.05669032037258148,
+ "learning_rate": 0.0001232925322545476,
+ "loss": 3.6783642768859863,
+ "step": 2575
+ },
+ {
+ "epoch": 35.781849912739965,
+ "grad_norm": 0.059275999665260315,
+ "learning_rate": 0.0001230803569792765,
+ "loss": 3.660548448562622,
+ "step": 2576
+ },
+ {
+ "epoch": 35.795811518324605,
+ "grad_norm": 0.056993551552295685,
+ "learning_rate": 0.0001228683266620121,
+ "loss": 3.6559674739837646,
+ "step": 2577
+ },
+ {
+ "epoch": 35.80977312390925,
+ "grad_norm": 0.058626480400562286,
+ "learning_rate": 0.00012265644147358326,
+ "loss": 3.679635763168335,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82373472949389,
+ "grad_norm": 0.05863586813211441,
+ "learning_rate": 0.00012244470158470226,
+ "loss": 3.657258987426758,
+ "step": 2579
+ },
+ {
+ "epoch": 35.83769633507853,
+ "grad_norm": 0.05809088423848152,
+ "learning_rate": 0.00012223310716596387,
+ "loss": 3.672903299331665,
+ "step": 2580
+ },
+ {
+ "epoch": 35.85165794066317,
+ "grad_norm": 0.05710308998823166,
+ "learning_rate": 0.00012202165838784601,
+ "loss": 3.679863929748535,
+ "step": 2581
+ },
+ {
+ "epoch": 35.86561954624782,
+ "grad_norm": 0.05867332965135574,
+ "learning_rate": 0.00012181035542070913,
+ "loss": 3.689939498901367,
+ "step": 2582
+ },
+ {
+ "epoch": 35.87958115183246,
+ "grad_norm": 0.057553596794605255,
+ "learning_rate": 0.00012159919843479617,
+ "loss": 3.6570992469787598,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8935427574171,
+ "grad_norm": 0.05948316305875778,
+ "learning_rate": 0.00012138818760023257,
+ "loss": 3.6697707176208496,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90750436300174,
+ "grad_norm": 0.058080170303583145,
+ "learning_rate": 0.00012117732308702592,
+ "loss": 3.667323589324951,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92146596858639,
+ "grad_norm": 0.05842271074652672,
+ "learning_rate": 0.00012096660506506605,
+ "loss": 3.688253164291382,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93542757417103,
+ "grad_norm": 0.058192040771245956,
+ "learning_rate": 0.00012075603370412443,
+ "loss": 3.6728200912475586,
+ "step": 2587
+ },
+ {
+ "epoch": 35.94938917975567,
+ "grad_norm": 0.0602075457572937,
+ "learning_rate": 0.00012054560917385476,
+ "loss": 3.6644749641418457,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96335078534032,
+ "grad_norm": 0.05764734745025635,
+ "learning_rate": 0.00012033533164379225,
+ "loss": 3.711341381072998,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97731239092496,
+ "grad_norm": 0.05787983164191246,
+ "learning_rate": 0.0001201252012833537,
+ "loss": 3.7075908184051514,
+ "step": 2590
+ },
+ {
+ "epoch": 35.9912739965096,
+ "grad_norm": 0.05705810338258743,
+ "learning_rate": 0.00011991521826183747,
+ "loss": 3.664104461669922,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.04269756004214287,
+ "learning_rate": 0.00011970538274842282,
+ "loss": 2.2993788719177246,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 0.6160264611244202,
+ "eval_runtime": 59.3623,
+ "eval_samples_per_second": 41.137,
+ "eval_steps_per_second": 0.657,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01396160558464,
+ "grad_norm": 0.05477261170744896,
+ "learning_rate": 0.00011949569491217073,
+ "loss": 3.6223397254943848,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02792321116929,
+ "grad_norm": 0.05240058898925781,
+ "learning_rate": 0.00011928615492202269,
+ "loss": 3.6275014877319336,
+ "step": 2594
+ },
+ {
+ "epoch": 36.04188481675393,
+ "grad_norm": 0.05488358810544014,
+ "learning_rate": 0.0001190767629468014,
+ "loss": 3.6421685218811035,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05584642233857,
+ "grad_norm": 0.05403005704283714,
+ "learning_rate": 0.00011886751915521007,
+ "loss": 3.6416306495666504,
+ "step": 2596
+ },
+ {
+ "epoch": 36.06980802792321,
+ "grad_norm": 0.05383783578872681,
+ "learning_rate": 0.00011865842371583278,
+ "loss": 3.6026291847229004,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083769633507856,
+ "grad_norm": 0.05379350110888481,
+ "learning_rate": 0.00011844947679713396,
+ "loss": 3.6012558937072754,
+ "step": 2598
+ },
+ {
+ "epoch": 36.0977312390925,
+ "grad_norm": 0.054601337760686874,
+ "learning_rate": 0.00011824067856745812,
+ "loss": 3.6351709365844727,
+ "step": 2599
+ },
+ {
+ "epoch": 36.11169284467714,
+ "grad_norm": 0.0523032620549202,
+ "learning_rate": 0.00011803202919503064,
+ "loss": 3.6155266761779785,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12565445026178,
+ "grad_norm": 0.05588298290967941,
+ "learning_rate": 0.00011782352884795615,
+ "loss": 3.6355879306793213,
+ "step": 2601
+ },
+ {
+ "epoch": 36.139616055846425,
+ "grad_norm": 0.05497356131672859,
+ "learning_rate": 0.00011761517769421983,
+ "loss": 3.5998620986938477,
+ "step": 2602
+ },
+ {
+ "epoch": 36.153577661431065,
+ "grad_norm": 0.055099356919527054,
+ "learning_rate": 0.00011740697590168635,
+ "loss": 3.6135497093200684,
+ "step": 2603
+ },
+ {
+ "epoch": 36.167539267015705,
+ "grad_norm": 0.05480733513832092,
+ "learning_rate": 0.00011719892363810018,
+ "loss": 3.631089687347412,
+ "step": 2604
+ },
+ {
+ "epoch": 36.181500872600346,
+ "grad_norm": 0.05424918979406357,
+ "learning_rate": 0.00011699102107108539,
+ "loss": 3.610616683959961,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19546247818499,
+ "grad_norm": 0.05589015781879425,
+ "learning_rate": 0.00011678326836814507,
+ "loss": 3.6014404296875,
+ "step": 2606
+ },
+ {
+ "epoch": 36.20942408376963,
+ "grad_norm": 0.05365624278783798,
+ "learning_rate": 0.00011657566569666198,
+ "loss": 3.6203207969665527,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223385689354274,
+ "grad_norm": 0.05549495667219162,
+ "learning_rate": 0.00011636821322389777,
+ "loss": 3.639944076538086,
+ "step": 2608
+ },
+ {
+ "epoch": 36.23734729493892,
+ "grad_norm": 0.05381424352526665,
+ "learning_rate": 0.00011616091111699333,
+ "loss": 3.6630518436431885,
+ "step": 2609
+ },
+ {
+ "epoch": 36.25130890052356,
+ "grad_norm": 0.05610237643122673,
+ "learning_rate": 0.00011595375954296786,
+ "loss": 3.634977102279663,
+ "step": 2610
+ },
+ {
+ "epoch": 36.2652705061082,
+ "grad_norm": 0.055497750639915466,
+ "learning_rate": 0.00011574675866871997,
+ "loss": 3.64717960357666,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27923211169284,
+ "grad_norm": 0.0543544627726078,
+ "learning_rate": 0.00011553990866102651,
+ "loss": 3.6224818229675293,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29319371727749,
+ "grad_norm": 0.056236982345581055,
+ "learning_rate": 0.00011533320968654265,
+ "loss": 3.6230814456939697,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30715532286213,
+ "grad_norm": 0.053707852959632874,
+ "learning_rate": 0.00011512666191180204,
+ "loss": 3.6355342864990234,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32111692844677,
+ "grad_norm": 0.057036347687244415,
+ "learning_rate": 0.00011492026550321653,
+ "loss": 3.6099886894226074,
+ "step": 2615
+ },
+ {
+ "epoch": 36.33507853403141,
+ "grad_norm": 0.05581443011760712,
+ "learning_rate": 0.00011471402062707607,
+ "loss": 3.6411590576171875,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34904013961606,
+ "grad_norm": 0.05495850369334221,
+ "learning_rate": 0.00011450792744954827,
+ "loss": 3.6358461380004883,
+ "step": 2617
+ },
+ {
+ "epoch": 36.3630017452007,
+ "grad_norm": 0.0549640990793705,
+ "learning_rate": 0.00011430198613667861,
+ "loss": 3.633671283721924,
+ "step": 2618
+ },
+ {
+ "epoch": 36.37696335078534,
+ "grad_norm": 0.05502061918377876,
+ "learning_rate": 0.00011409619685439064,
+ "loss": 3.632413864135742,
+ "step": 2619
+ },
+ {
+ "epoch": 36.390924956369986,
+ "grad_norm": 0.05518985912203789,
+ "learning_rate": 0.00011389055976848477,
+ "loss": 3.646486759185791,
+ "step": 2620
+ },
+ {
+ "epoch": 36.404886561954626,
+ "grad_norm": 0.053396400064229965,
+ "learning_rate": 0.00011368507504463914,
+ "loss": 3.6610465049743652,
+ "step": 2621
+ },
+ {
+ "epoch": 36.41884816753927,
+ "grad_norm": 0.05592766031622887,
+ "learning_rate": 0.00011347974284840914,
+ "loss": 3.6322021484375,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43280977312391,
+ "grad_norm": 0.054665833711624146,
+ "learning_rate": 0.00011327456334522725,
+ "loss": 3.6100504398345947,
+ "step": 2623
+ },
+ {
+ "epoch": 36.446771378708554,
+ "grad_norm": 0.05556786060333252,
+ "learning_rate": 0.00011306953670040275,
+ "loss": 3.647045135498047,
+ "step": 2624
+ },
+ {
+ "epoch": 36.460732984293195,
+ "grad_norm": 0.05619419366121292,
+ "learning_rate": 0.00011286466307912184,
+ "loss": 3.6307947635650635,
+ "step": 2625
+ },
+ {
+ "epoch": 36.474694589877835,
+ "grad_norm": 0.0551043376326561,
+ "learning_rate": 0.00011265994264644783,
+ "loss": 3.6331896781921387,
+ "step": 2626
+ },
+ {
+ "epoch": 36.488656195462475,
+ "grad_norm": 0.0548650398850441,
+ "learning_rate": 0.00011245537556732001,
+ "loss": 3.633991241455078,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50261780104712,
+ "grad_norm": 0.056781429797410965,
+ "learning_rate": 0.00011225096200655442,
+ "loss": 3.653165340423584,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51657940663176,
+ "grad_norm": 0.05489041283726692,
+ "learning_rate": 0.00011204670212884342,
+ "loss": 3.6232171058654785,
+ "step": 2629
+ },
+ {
+ "epoch": 36.5305410122164,
+ "grad_norm": 0.055730190128088,
+ "learning_rate": 0.00011184259609875557,
+ "loss": 3.644775390625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544502617801044,
+ "grad_norm": 0.058052800595760345,
+ "learning_rate": 0.00011163864408073522,
+ "loss": 3.6141133308410645,
+ "step": 2631
+ },
+ {
+ "epoch": 36.55846422338569,
+ "grad_norm": 0.05508120357990265,
+ "learning_rate": 0.00011143484623910293,
+ "loss": 3.6275734901428223,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57242582897033,
+ "grad_norm": 0.0572693794965744,
+ "learning_rate": 0.00011123120273805496,
+ "loss": 3.631286859512329,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58638743455497,
+ "grad_norm": 0.05745089426636696,
+ "learning_rate": 0.0001110277137416632,
+ "loss": 3.6396098136901855,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60034904013962,
+ "grad_norm": 0.05624939128756523,
+ "learning_rate": 0.00011082437941387512,
+ "loss": 3.655447006225586,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61431064572426,
+ "grad_norm": 0.057037804275751114,
+ "learning_rate": 0.00011062119991851322,
+ "loss": 3.6295433044433594,
+ "step": 2636
+ },
+ {
+ "epoch": 36.6282722513089,
+ "grad_norm": 0.05740547552704811,
+ "learning_rate": 0.000110418175419276,
+ "loss": 3.6350302696228027,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64223385689354,
+ "grad_norm": 0.05586782470345497,
+ "learning_rate": 0.00011021530607973626,
+ "loss": 3.6298985481262207,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65619546247819,
+ "grad_norm": 0.05766303837299347,
+ "learning_rate": 0.00011001259206334235,
+ "loss": 3.646199941635132,
+ "step": 2639
+ },
+ {
+ "epoch": 36.67015706806283,
+ "grad_norm": 0.05768761411309242,
+ "learning_rate": 0.00010981003353341721,
+ "loss": 3.638676881790161,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68411867364747,
+ "grad_norm": 0.05806223303079605,
+ "learning_rate": 0.00010960763065315864,
+ "loss": 3.61555814743042,
+ "step": 2641
+ },
+ {
+ "epoch": 36.69808027923211,
+ "grad_norm": 0.05654353275895119,
+ "learning_rate": 0.00010940538358563906,
+ "loss": 3.6435892581939697,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712041884816756,
+ "grad_norm": 0.0572567842900753,
+ "learning_rate": 0.00010920329249380509,
+ "loss": 3.638545513153076,
+ "step": 2643
+ },
+ {
+ "epoch": 36.726003490401396,
+ "grad_norm": 0.05685669183731079,
+ "learning_rate": 0.00010900135754047799,
+ "loss": 3.6248223781585693,
+ "step": 2644
+ },
+ {
+ "epoch": 36.73996509598604,
+ "grad_norm": 0.05836780369281769,
+ "learning_rate": 0.00010879957888835307,
+ "loss": 3.6728572845458984,
+ "step": 2645
+ },
+ {
+ "epoch": 36.753926701570684,
+ "grad_norm": 0.05652812868356705,
+ "learning_rate": 0.00010859795669999981,
+ "loss": 3.6360411643981934,
+ "step": 2646
+ },
+ {
+ "epoch": 36.767888307155324,
+ "grad_norm": 0.05731404572725296,
+ "learning_rate": 0.00010839649113786137,
+ "loss": 3.624966621398926,
+ "step": 2647
+ },
+ {
+ "epoch": 36.781849912739965,
+ "grad_norm": 0.05664033442735672,
+ "learning_rate": 0.00010819518236425514,
+ "loss": 3.6407511234283447,
+ "step": 2648
+ },
+ {
+ "epoch": 36.795811518324605,
+ "grad_norm": 0.056350719183683395,
+ "learning_rate": 0.00010799403054137197,
+ "loss": 3.6436259746551514,
+ "step": 2649
+ },
+ {
+ "epoch": 36.80977312390925,
+ "grad_norm": 0.05680667608976364,
+ "learning_rate": 0.00010779303583127609,
+ "loss": 3.6506614685058594,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82373472949389,
+ "grad_norm": 0.05701072886586189,
+ "learning_rate": 0.00010759219839590535,
+ "loss": 3.643155574798584,
+ "step": 2651
+ },
+ {
+ "epoch": 36.83769633507853,
+ "grad_norm": 0.05728844553232193,
+ "learning_rate": 0.00010739151839707089,
+ "loss": 3.6734180450439453,
+ "step": 2652
+ },
+ {
+ "epoch": 36.85165794066317,
+ "grad_norm": 0.05713675543665886,
+ "learning_rate": 0.00010719099599645706,
+ "loss": 3.665465831756592,
+ "step": 2653
+ },
+ {
+ "epoch": 36.86561954624782,
+ "grad_norm": 0.05626962333917618,
+ "learning_rate": 0.0001069906313556208,
+ "loss": 3.6559338569641113,
+ "step": 2654
+ },
+ {
+ "epoch": 36.87958115183246,
+ "grad_norm": 0.05827031657099724,
+ "learning_rate": 0.00010679042463599263,
+ "loss": 3.64331316947937,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8935427574171,
+ "grad_norm": 0.05828213319182396,
+ "learning_rate": 0.00010659037599887545,
+ "loss": 3.631319046020508,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90750436300174,
+ "grad_norm": 0.057474542409181595,
+ "learning_rate": 0.0001063904856054446,
+ "loss": 3.6365761756896973,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92146596858639,
+ "grad_norm": 0.056101586669683456,
+ "learning_rate": 0.00010619075361674836,
+ "loss": 3.62568998336792,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93542757417103,
+ "grad_norm": 0.058322884142398834,
+ "learning_rate": 0.0001059911801937071,
+ "loss": 3.6452713012695312,
+ "step": 2659
+ },
+ {
+ "epoch": 36.94938917975567,
+ "grad_norm": 0.05735790356993675,
+ "learning_rate": 0.00010579176549711365,
+ "loss": 3.644145965576172,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96335078534032,
+ "grad_norm": 0.05761740356683731,
+ "learning_rate": 0.00010559250968763265,
+ "loss": 3.652322292327881,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97731239092496,
+ "grad_norm": 0.05765356495976448,
+ "learning_rate": 0.00010539341292580086,
+ "loss": 3.6571314334869385,
+ "step": 2662
+ },
+ {
+ "epoch": 36.9912739965096,
+ "grad_norm": 0.05606544017791748,
+ "learning_rate": 0.00010519447537202729,
+ "loss": 3.6158785820007324,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.042550574988126755,
+ "learning_rate": 0.00010499569718659201,
+ "loss": 2.262657403945923,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 0.6172080039978027,
+ "eval_runtime": 59.1561,
+ "eval_samples_per_second": 41.281,
+ "eval_steps_per_second": 0.659,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01396160558464,
+ "grad_norm": 0.054308392107486725,
+ "learning_rate": 0.00010479707852964713,
+ "loss": 3.6250295639038086,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02792321116929,
+ "grad_norm": 0.05218822881579399,
+ "learning_rate": 0.00010459861956121611,
+ "loss": 3.598968982696533,
+ "step": 2666
+ },
+ {
+ "epoch": 37.04188481675393,
+ "grad_norm": 0.05366450920701027,
+ "learning_rate": 0.00010440032044119383,
+ "loss": 3.5767178535461426,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05584642233857,
+ "grad_norm": 0.05169176682829857,
+ "learning_rate": 0.00010420218132934614,
+ "loss": 3.5973262786865234,
+ "step": 2668
+ },
+ {
+ "epoch": 37.06980802792321,
+ "grad_norm": 0.05438115820288658,
+ "learning_rate": 0.00010400420238531023,
+ "loss": 3.584498167037964,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083769633507856,
+ "grad_norm": 0.05290330946445465,
+ "learning_rate": 0.00010380638376859416,
+ "loss": 3.614943504333496,
+ "step": 2670
+ },
+ {
+ "epoch": 37.0977312390925,
+ "grad_norm": 0.052781153470277786,
+ "learning_rate": 0.00010360872563857682,
+ "loss": 3.5990896224975586,
+ "step": 2671
+ },
+ {
+ "epoch": 37.11169284467714,
+ "grad_norm": 0.05197159945964813,
+ "learning_rate": 0.0001034112281545079,
+ "loss": 3.579005241394043,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12565445026178,
+ "grad_norm": 0.05242403224110603,
+ "learning_rate": 0.00010321389147550725,
+ "loss": 3.608119010925293,
+ "step": 2673
+ },
+ {
+ "epoch": 37.139616055846425,
+ "grad_norm": 0.05325670540332794,
+ "learning_rate": 0.00010301671576056588,
+ "loss": 3.6093316078186035,
+ "step": 2674
+ },
+ {
+ "epoch": 37.153577661431065,
+ "grad_norm": 0.05325893685221672,
+ "learning_rate": 0.00010281970116854436,
+ "loss": 3.6037325859069824,
+ "step": 2675
+ },
+ {
+ "epoch": 37.167539267015705,
+ "grad_norm": 0.05180941894650459,
+ "learning_rate": 0.00010262284785817392,
+ "loss": 3.584935188293457,
+ "step": 2676
+ },
+ {
+ "epoch": 37.181500872600346,
+ "grad_norm": 0.05396177992224693,
+ "learning_rate": 0.00010242615598805574,
+ "loss": 3.623363494873047,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19546247818499,
+ "grad_norm": 0.05336499959230423,
+ "learning_rate": 0.00010222962571666088,
+ "loss": 3.6231229305267334,
+ "step": 2678
+ },
+ {
+ "epoch": 37.20942408376963,
+ "grad_norm": 0.052409425377845764,
+ "learning_rate": 0.00010203325720233032,
+ "loss": 3.599355697631836,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223385689354274,
+ "grad_norm": 0.0531868040561676,
+ "learning_rate": 0.00010183705060327433,
+ "loss": 3.6197028160095215,
+ "step": 2680
+ },
+ {
+ "epoch": 37.23734729493892,
+ "grad_norm": 0.053192511200904846,
+ "learning_rate": 0.00010164100607757346,
+ "loss": 3.612718105316162,
+ "step": 2681
+ },
+ {
+ "epoch": 37.25130890052356,
+ "grad_norm": 0.052938785403966904,
+ "learning_rate": 0.00010144512378317687,
+ "loss": 3.6197361946105957,
+ "step": 2682
+ },
+ {
+ "epoch": 37.2652705061082,
+ "grad_norm": 0.0538516603410244,
+ "learning_rate": 0.00010124940387790354,
+ "loss": 3.6179161071777344,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27923211169284,
+ "grad_norm": 0.05364573001861572,
+ "learning_rate": 0.00010105384651944148,
+ "loss": 3.5888404846191406,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29319371727749,
+ "grad_norm": 0.05395280569791794,
+ "learning_rate": 0.00010085845186534769,
+ "loss": 3.6174027919769287,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30715532286213,
+ "grad_norm": 0.05424579605460167,
+ "learning_rate": 0.00010066322007304819,
+ "loss": 3.6061956882476807,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32111692844677,
+ "grad_norm": 0.05412362515926361,
+ "learning_rate": 0.00010046815129983757,
+ "loss": 3.596550464630127,
+ "step": 2687
+ },
+ {
+ "epoch": 37.33507853403141,
+ "grad_norm": 0.05387300252914429,
+ "learning_rate": 0.00010027324570287925,
+ "loss": 3.627476215362549,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34904013961606,
+ "grad_norm": 0.05500417947769165,
+ "learning_rate": 0.00010007850343920519,
+ "loss": 3.601841926574707,
+ "step": 2689
+ },
+ {
+ "epoch": 37.3630017452007,
+ "grad_norm": 0.05346129834651947,
+ "learning_rate": 9.988392466571572e-05,
+ "loss": 3.6150965690612793,
+ "step": 2690
+ },
+ {
+ "epoch": 37.37696335078534,
+ "grad_norm": 0.054463278502225876,
+ "learning_rate": 9.968950953917945e-05,
+ "loss": 3.5890097618103027,
+ "step": 2691
+ },
+ {
+ "epoch": 37.390924956369986,
+ "grad_norm": 0.05312243849039078,
+ "learning_rate": 9.949525821623309e-05,
+ "loss": 3.597362518310547,
+ "step": 2692
+ },
+ {
+ "epoch": 37.404886561954626,
+ "grad_norm": 0.05424235016107559,
+ "learning_rate": 9.930117085338156e-05,
+ "loss": 3.5707716941833496,
+ "step": 2693
+ },
+ {
+ "epoch": 37.41884816753927,
+ "grad_norm": 0.054021209478378296,
+ "learning_rate": 9.910724760699731e-05,
+ "loss": 3.6366126537323,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43280977312391,
+ "grad_norm": 0.05502825230360031,
+ "learning_rate": 9.891348863332093e-05,
+ "loss": 3.6250295639038086,
+ "step": 2695
+ },
+ {
+ "epoch": 37.446771378708554,
+ "grad_norm": 0.05483090132474899,
+ "learning_rate": 9.871989408846051e-05,
+ "loss": 3.611407518386841,
+ "step": 2696
+ },
+ {
+ "epoch": 37.460732984293195,
+ "grad_norm": 0.05381562188267708,
+ "learning_rate": 9.852646412839178e-05,
+ "loss": 3.6136293411254883,
+ "step": 2697
+ },
+ {
+ "epoch": 37.474694589877835,
+ "grad_norm": 0.054287511855363846,
+ "learning_rate": 9.833319890895756e-05,
+ "loss": 3.6080033779144287,
+ "step": 2698
+ },
+ {
+ "epoch": 37.488656195462475,
+ "grad_norm": 0.05399102717638016,
+ "learning_rate": 9.814009858586817e-05,
+ "loss": 3.6026833057403564,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50261780104712,
+ "grad_norm": 0.053677961230278015,
+ "learning_rate": 9.794716331470131e-05,
+ "loss": 3.6080117225646973,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51657940663176,
+ "grad_norm": 0.055184055119752884,
+ "learning_rate": 9.77543932509012e-05,
+ "loss": 3.6028759479522705,
+ "step": 2701
+ },
+ {
+ "epoch": 37.5305410122164,
+ "grad_norm": 0.0554618164896965,
+ "learning_rate": 9.756178854977925e-05,
+ "loss": 3.6170661449432373,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544502617801044,
+ "grad_norm": 0.05491705983877182,
+ "learning_rate": 9.736934936651362e-05,
+ "loss": 3.6152079105377197,
+ "step": 2703
+ },
+ {
+ "epoch": 37.55846422338569,
+ "grad_norm": 0.055647511035203934,
+ "learning_rate": 9.717707585614916e-05,
+ "loss": 3.610424757003784,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57242582897033,
+ "grad_norm": 0.0561630018055439,
+ "learning_rate": 9.698496817359698e-05,
+ "loss": 3.6318728923797607,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58638743455497,
+ "grad_norm": 0.05447172373533249,
+ "learning_rate": 9.679302647363476e-05,
+ "loss": 3.6186347007751465,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60034904013962,
+ "grad_norm": 0.05509922280907631,
+ "learning_rate": 9.660125091090675e-05,
+ "loss": 3.6077895164489746,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61431064572426,
+ "grad_norm": 0.055178701877593994,
+ "learning_rate": 9.64096416399228e-05,
+ "loss": 3.637204647064209,
+ "step": 2708
+ },
+ {
+ "epoch": 37.6282722513089,
+ "grad_norm": 0.05637604370713234,
+ "learning_rate": 9.621819881505918e-05,
+ "loss": 3.6154603958129883,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64223385689354,
+ "grad_norm": 0.056297753006219864,
+ "learning_rate": 9.602692259055767e-05,
+ "loss": 3.6120991706848145,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65619546247819,
+ "grad_norm": 0.054915741086006165,
+ "learning_rate": 9.583581312052646e-05,
+ "loss": 3.626859188079834,
+ "step": 2711
+ },
+ {
+ "epoch": 37.67015706806283,
+ "grad_norm": 0.05664214491844177,
+ "learning_rate": 9.564487055893867e-05,
+ "loss": 3.6137616634368896,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68411867364747,
+ "grad_norm": 0.05572838336229324,
+ "learning_rate": 9.545409505963338e-05,
+ "loss": 3.622990131378174,
+ "step": 2713
+ },
+ {
+ "epoch": 37.69808027923211,
+ "grad_norm": 0.056440044194459915,
+ "learning_rate": 9.526348677631499e-05,
+ "loss": 3.6483852863311768,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712041884816756,
+ "grad_norm": 0.0582021027803421,
+ "learning_rate": 9.507304586255311e-05,
+ "loss": 3.6085047721862793,
+ "step": 2715
+ },
+ {
+ "epoch": 37.726003490401396,
+ "grad_norm": 0.055895350873470306,
+ "learning_rate": 9.488277247178267e-05,
+ "loss": 3.625001907348633,
+ "step": 2716
+ },
+ {
+ "epoch": 37.73996509598604,
+ "grad_norm": 0.055818382650613785,
+ "learning_rate": 9.469266675730319e-05,
+ "loss": 3.627051830291748,
+ "step": 2717
+ },
+ {
+ "epoch": 37.753926701570684,
+ "grad_norm": 0.05565737560391426,
+ "learning_rate": 9.450272887227983e-05,
+ "loss": 3.607292890548706,
+ "step": 2718
+ },
+ {
+ "epoch": 37.767888307155324,
+ "grad_norm": 0.05473797023296356,
+ "learning_rate": 9.431295896974182e-05,
+ "loss": 3.613346815109253,
+ "step": 2719
+ },
+ {
+ "epoch": 37.781849912739965,
+ "grad_norm": 0.056405726820230484,
+ "learning_rate": 9.412335720258341e-05,
+ "loss": 3.6148681640625,
+ "step": 2720
+ },
+ {
+ "epoch": 37.795811518324605,
+ "grad_norm": 0.0563163161277771,
+ "learning_rate": 9.393392372356335e-05,
+ "loss": 3.6225497722625732,
+ "step": 2721
+ },
+ {
+ "epoch": 37.80977312390925,
+ "grad_norm": 0.05682912468910217,
+ "learning_rate": 9.374465868530477e-05,
+ "loss": 3.620290756225586,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82373472949389,
+ "grad_norm": 0.05696539953351021,
+ "learning_rate": 9.355556224029515e-05,
+ "loss": 3.617171287536621,
+ "step": 2723
+ },
+ {
+ "epoch": 37.83769633507853,
+ "grad_norm": 0.05691574513912201,
+ "learning_rate": 9.336663454088593e-05,
+ "loss": 3.6293022632598877,
+ "step": 2724
+ },
+ {
+ "epoch": 37.85165794066317,
+ "grad_norm": 0.05566944554448128,
+ "learning_rate": 9.317787573929282e-05,
+ "loss": 3.6245670318603516,
+ "step": 2725
+ },
+ {
+ "epoch": 37.86561954624782,
+ "grad_norm": 0.05736027657985687,
+ "learning_rate": 9.298928598759541e-05,
+ "loss": 3.625857353210449,
+ "step": 2726
+ },
+ {
+ "epoch": 37.87958115183246,
+ "grad_norm": 0.055642999708652496,
+ "learning_rate": 9.280086543773698e-05,
+ "loss": 3.610138416290283,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8935427574171,
+ "grad_norm": 0.057330258190631866,
+ "learning_rate": 9.26126142415247e-05,
+ "loss": 3.593010902404785,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90750436300174,
+ "grad_norm": 0.05633636936545372,
+ "learning_rate": 9.2424532550629e-05,
+ "loss": 3.6144258975982666,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92146596858639,
+ "grad_norm": 0.05787842720746994,
+ "learning_rate": 9.223662051658408e-05,
+ "loss": 3.609016180038452,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93542757417103,
+ "grad_norm": 0.05679284781217575,
+ "learning_rate": 9.204887829078709e-05,
+ "loss": 3.5973129272460938,
+ "step": 2731
+ },
+ {
+ "epoch": 37.94938917975567,
+ "grad_norm": 0.056735891848802567,
+ "learning_rate": 9.186130602449861e-05,
+ "loss": 3.6157774925231934,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96335078534032,
+ "grad_norm": 0.05625693500041962,
+ "learning_rate": 9.167390386884224e-05,
+ "loss": 3.617924928665161,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97731239092496,
+ "grad_norm": 0.05621540546417236,
+ "learning_rate": 9.148667197480455e-05,
+ "loss": 3.597618341445923,
+ "step": 2734
+ },
+ {
+ "epoch": 37.9912739965096,
+ "grad_norm": 0.0570329986512661,
+ "learning_rate": 9.129961049323494e-05,
+ "loss": 3.6188273429870605,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.04125213250517845,
+ "learning_rate": 9.111271957484519e-05,
+ "loss": 2.2358860969543457,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 0.6186589002609253,
+ "eval_runtime": 58.8494,
+ "eval_samples_per_second": 41.496,
+ "eval_steps_per_second": 0.663,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01396160558464,
+ "grad_norm": 0.05371171608567238,
+ "learning_rate": 9.092599937021031e-05,
+ "loss": 3.5614218711853027,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02792321116929,
+ "grad_norm": 0.05166265740990639,
+ "learning_rate": 9.073945002976715e-05,
+ "loss": 3.5690364837646484,
+ "step": 2738
+ },
+ {
+ "epoch": 38.04188481675393,
+ "grad_norm": 0.053279295563697815,
+ "learning_rate": 9.055307170381522e-05,
+ "loss": 3.575930595397949,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05584642233857,
+ "grad_norm": 0.051418207585811615,
+ "learning_rate": 9.036686454251613e-05,
+ "loss": 3.568121910095215,
+ "step": 2740
+ },
+ {
+ "epoch": 38.06980802792321,
+ "grad_norm": 0.051363058388233185,
+ "learning_rate": 9.018082869589369e-05,
+ "loss": 3.5916590690612793,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083769633507856,
+ "grad_norm": 0.05149349942803383,
+ "learning_rate": 8.999496431383368e-05,
+ "loss": 3.5861239433288574,
+ "step": 2742
+ },
+ {
+ "epoch": 38.0977312390925,
+ "grad_norm": 0.05174670368432999,
+ "learning_rate": 8.980927154608341e-05,
+ "loss": 3.5812082290649414,
+ "step": 2743
+ },
+ {
+ "epoch": 38.11169284467714,
+ "grad_norm": 0.05128023028373718,
+ "learning_rate": 8.962375054225255e-05,
+ "loss": 3.5962631702423096,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12565445026178,
+ "grad_norm": 0.05153957009315491,
+ "learning_rate": 8.943840145181177e-05,
+ "loss": 3.5739307403564453,
+ "step": 2745
+ },
+ {
+ "epoch": 38.139616055846425,
+ "grad_norm": 0.0521809458732605,
+ "learning_rate": 8.925322442409361e-05,
+ "loss": 3.5964818000793457,
+ "step": 2746
+ },
+ {
+ "epoch": 38.153577661431065,
+ "grad_norm": 0.05092332884669304,
+ "learning_rate": 8.906821960829163e-05,
+ "loss": 3.5737056732177734,
+ "step": 2747
+ },
+ {
+ "epoch": 38.167539267015705,
+ "grad_norm": 0.052281491458415985,
+ "learning_rate": 8.888338715346113e-05,
+ "loss": 3.5943212509155273,
+ "step": 2748
+ },
+ {
+ "epoch": 38.181500872600346,
+ "grad_norm": 0.05170704424381256,
+ "learning_rate": 8.869872720851831e-05,
+ "loss": 3.585753917694092,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19546247818499,
+ "grad_norm": 0.05175631120800972,
+ "learning_rate": 8.851423992224012e-05,
+ "loss": 3.606010913848877,
+ "step": 2750
+ },
+ {
+ "epoch": 38.20942408376963,
+ "grad_norm": 0.0525258332490921,
+ "learning_rate": 8.832992544326473e-05,
+ "loss": 3.591367244720459,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223385689354274,
+ "grad_norm": 0.051923174411058426,
+ "learning_rate": 8.814578392009104e-05,
+ "loss": 3.563871383666992,
+ "step": 2752
+ },
+ {
+ "epoch": 38.23734729493892,
+ "grad_norm": 0.053746938705444336,
+ "learning_rate": 8.796181550107857e-05,
+ "loss": 3.576528787612915,
+ "step": 2753
+ },
+ {
+ "epoch": 38.25130890052356,
+ "grad_norm": 0.05283728986978531,
+ "learning_rate": 8.777802033444712e-05,
+ "loss": 3.600759983062744,
+ "step": 2754
+ },
+ {
+ "epoch": 38.2652705061082,
+ "grad_norm": 0.05214889720082283,
+ "learning_rate": 8.75943985682774e-05,
+ "loss": 3.582427501678467,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27923211169284,
+ "grad_norm": 0.05283285677433014,
+ "learning_rate": 8.741095035051017e-05,
+ "loss": 3.579399585723877,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29319371727749,
+ "grad_norm": 0.053677286952733994,
+ "learning_rate": 8.722767582894613e-05,
+ "loss": 3.5718324184417725,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30715532286213,
+ "grad_norm": 0.05384226143360138,
+ "learning_rate": 8.704457515124636e-05,
+ "loss": 3.593702793121338,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32111692844677,
+ "grad_norm": 0.05237982049584389,
+ "learning_rate": 8.686164846493176e-05,
+ "loss": 3.580988883972168,
+ "step": 2759
+ },
+ {
+ "epoch": 38.33507853403141,
+ "grad_norm": 0.05386205390095711,
+ "learning_rate": 8.667889591738317e-05,
+ "loss": 3.5706114768981934,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34904013961606,
+ "grad_norm": 0.05325917899608612,
+ "learning_rate": 8.649631765584083e-05,
+ "loss": 3.5707662105560303,
+ "step": 2761
+ },
+ {
+ "epoch": 38.3630017452007,
+ "grad_norm": 0.053492557257413864,
+ "learning_rate": 8.631391382740482e-05,
+ "loss": 3.618515968322754,
+ "step": 2762
+ },
+ {
+ "epoch": 38.37696335078534,
+ "grad_norm": 0.05370897799730301,
+ "learning_rate": 8.613168457903459e-05,
+ "loss": 3.5731120109558105,
+ "step": 2763
+ },
+ {
+ "epoch": 38.390924956369986,
+ "grad_norm": 0.05325158312916756,
+ "learning_rate": 8.594963005754901e-05,
+ "loss": 3.591339588165283,
+ "step": 2764
+ },
+ {
+ "epoch": 38.404886561954626,
+ "grad_norm": 0.05411994829773903,
+ "learning_rate": 8.57677504096261e-05,
+ "loss": 3.5969743728637695,
+ "step": 2765
+ },
+ {
+ "epoch": 38.41884816753927,
+ "grad_norm": 0.05363132059574127,
+ "learning_rate": 8.558604578180301e-05,
+ "loss": 3.620718479156494,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43280977312391,
+ "grad_norm": 0.0541871078312397,
+ "learning_rate": 8.540451632047593e-05,
+ "loss": 3.589287281036377,
+ "step": 2767
+ },
+ {
+ "epoch": 38.446771378708554,
+ "grad_norm": 0.05317665636539459,
+ "learning_rate": 8.522316217189972e-05,
+ "loss": 3.5677599906921387,
+ "step": 2768
+ },
+ {
+ "epoch": 38.460732984293195,
+ "grad_norm": 0.05288826301693916,
+ "learning_rate": 8.504198348218821e-05,
+ "loss": 3.5985312461853027,
+ "step": 2769
+ },
+ {
+ "epoch": 38.474694589877835,
+ "grad_norm": 0.05443241819739342,
+ "learning_rate": 8.486098039731384e-05,
+ "loss": 3.5577774047851562,
+ "step": 2770
+ },
+ {
+ "epoch": 38.488656195462475,
+ "grad_norm": 0.054173294454813004,
+ "learning_rate": 8.46801530631075e-05,
+ "loss": 3.5860495567321777,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50261780104712,
+ "grad_norm": 0.05407063290476799,
+ "learning_rate": 8.449950162525859e-05,
+ "loss": 3.604748249053955,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51657940663176,
+ "grad_norm": 0.05461187660694122,
+ "learning_rate": 8.431902622931443e-05,
+ "loss": 3.563236713409424,
+ "step": 2773
+ },
+ {
+ "epoch": 38.5305410122164,
+ "grad_norm": 0.05529208108782768,
+ "learning_rate": 8.413872702068119e-05,
+ "loss": 3.586937427520752,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544502617801044,
+ "grad_norm": 0.05481037497520447,
+ "learning_rate": 8.395860414462238e-05,
+ "loss": 3.6134698390960693,
+ "step": 2775
+ },
+ {
+ "epoch": 38.55846422338569,
+ "grad_norm": 0.054424822330474854,
+ "learning_rate": 8.377865774625985e-05,
+ "loss": 3.575363874435425,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57242582897033,
+ "grad_norm": 0.0553407222032547,
+ "learning_rate": 8.359888797057316e-05,
+ "loss": 3.5970382690429688,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58638743455497,
+ "grad_norm": 0.05625150352716446,
+ "learning_rate": 8.341929496239958e-05,
+ "loss": 3.6032121181488037,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60034904013962,
+ "grad_norm": 0.05586005747318268,
+ "learning_rate": 8.323987886643402e-05,
+ "loss": 3.602691888809204,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61431064572426,
+ "grad_norm": 0.054208606481552124,
+ "learning_rate": 8.306063982722855e-05,
+ "loss": 3.601468563079834,
+ "step": 2780
+ },
+ {
+ "epoch": 38.6282722513089,
+ "grad_norm": 0.056410644203424454,
+ "learning_rate": 8.288157798919318e-05,
+ "loss": 3.585646867752075,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64223385689354,
+ "grad_norm": 0.054060712456703186,
+ "learning_rate": 8.27026934965945e-05,
+ "loss": 3.566913604736328,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65619546247819,
+ "grad_norm": 0.0546812042593956,
+ "learning_rate": 8.25239864935566e-05,
+ "loss": 3.581869602203369,
+ "step": 2783
+ },
+ {
+ "epoch": 38.67015706806283,
+ "grad_norm": 0.05502549931406975,
+ "learning_rate": 8.23454571240605e-05,
+ "loss": 3.584887981414795,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68411867364747,
+ "grad_norm": 0.056419193744659424,
+ "learning_rate": 8.216710553194416e-05,
+ "loss": 3.5797278881073,
+ "step": 2785
+ },
+ {
+ "epoch": 38.69808027923211,
+ "grad_norm": 0.05669495090842247,
+ "learning_rate": 8.198893186090222e-05,
+ "loss": 3.604696273803711,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712041884816756,
+ "grad_norm": 0.05554523691534996,
+ "learning_rate": 8.181093625448585e-05,
+ "loss": 3.591275215148926,
+ "step": 2787
+ },
+ {
+ "epoch": 38.726003490401396,
+ "grad_norm": 0.05579648166894913,
+ "learning_rate": 8.163311885610299e-05,
+ "loss": 3.5911900997161865,
+ "step": 2788
+ },
+ {
+ "epoch": 38.73996509598604,
+ "grad_norm": 0.05564593896269798,
+ "learning_rate": 8.145547980901791e-05,
+ "loss": 3.6048314571380615,
+ "step": 2789
+ },
+ {
+ "epoch": 38.753926701570684,
+ "grad_norm": 0.05567439645528793,
+ "learning_rate": 8.127801925635126e-05,
+ "loss": 3.564915180206299,
+ "step": 2790
+ },
+ {
+ "epoch": 38.767888307155324,
+ "grad_norm": 0.055503327399492264,
+ "learning_rate": 8.110073734107954e-05,
+ "loss": 3.574024200439453,
+ "step": 2791
+ },
+ {
+ "epoch": 38.781849912739965,
+ "grad_norm": 0.05515526980161667,
+ "learning_rate": 8.092363420603584e-05,
+ "loss": 3.6017959117889404,
+ "step": 2792
+ },
+ {
+ "epoch": 38.795811518324605,
+ "grad_norm": 0.05573352053761482,
+ "learning_rate": 8.074670999390896e-05,
+ "loss": 3.5869383811950684,
+ "step": 2793
+ },
+ {
+ "epoch": 38.80977312390925,
+ "grad_norm": 0.05688846483826637,
+ "learning_rate": 8.056996484724342e-05,
+ "loss": 3.612272024154663,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82373472949389,
+ "grad_norm": 0.05551573261618614,
+ "learning_rate": 8.039339890843958e-05,
+ "loss": 3.5818700790405273,
+ "step": 2795
+ },
+ {
+ "epoch": 38.83769633507853,
+ "grad_norm": 0.05531308427453041,
+ "learning_rate": 8.02170123197535e-05,
+ "loss": 3.558943748474121,
+ "step": 2796
+ },
+ {
+ "epoch": 38.85165794066317,
+ "grad_norm": 0.05705123022198677,
+ "learning_rate": 8.004080522329674e-05,
+ "loss": 3.603858709335327,
+ "step": 2797
+ },
+ {
+ "epoch": 38.86561954624782,
+ "grad_norm": 0.055374935269355774,
+ "learning_rate": 7.986477776103601e-05,
+ "loss": 3.5933685302734375,
+ "step": 2798
+ },
+ {
+ "epoch": 38.87958115183246,
+ "grad_norm": 0.056353647261857986,
+ "learning_rate": 7.968893007479343e-05,
+ "loss": 3.598839282989502,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8935427574171,
+ "grad_norm": 0.05722062662243843,
+ "learning_rate": 7.951326230624658e-05,
+ "loss": 3.645443916320801,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90750436300174,
+ "grad_norm": 0.055990997701883316,
+ "learning_rate": 7.933777459692756e-05,
+ "loss": 3.5926828384399414,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92146596858639,
+ "grad_norm": 0.0561661571264267,
+ "learning_rate": 7.916246708822373e-05,
+ "loss": 3.6137099266052246,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93542757417103,
+ "grad_norm": 0.05624176934361458,
+ "learning_rate": 7.898733992137715e-05,
+ "loss": 3.597254991531372,
+ "step": 2803
+ },
+ {
+ "epoch": 38.94938917975567,
+ "grad_norm": 0.05730769410729408,
+ "learning_rate": 7.881239323748475e-05,
+ "loss": 3.595649242401123,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96335078534032,
+ "grad_norm": 0.05613584816455841,
+ "learning_rate": 7.863762717749771e-05,
+ "loss": 3.5786027908325195,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97731239092496,
+ "grad_norm": 0.05605073273181915,
+ "learning_rate": 7.846304188222189e-05,
+ "loss": 3.5973892211914062,
+ "step": 2806
+ },
+ {
+ "epoch": 38.9912739965096,
+ "grad_norm": 0.05703037977218628,
+ "learning_rate": 7.828863749231777e-05,
+ "loss": 3.614419460296631,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.042099807411432266,
+ "learning_rate": 7.811441414829958e-05,
+ "loss": 2.2424778938293457,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 0.620037853717804,
+ "eval_runtime": 59.1566,
+ "eval_samples_per_second": 41.28,
+ "eval_steps_per_second": 0.659,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01396160558464,
+ "grad_norm": 0.05103587359189987,
+ "learning_rate": 7.794037199053598e-05,
+ "loss": 3.556558132171631,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02792321116929,
+ "grad_norm": 0.049501027911901474,
+ "learning_rate": 7.776651115924959e-05,
+ "loss": 3.5505282878875732,
+ "step": 2810
+ },
+ {
+ "epoch": 39.04188481675393,
+ "grad_norm": 0.05127207562327385,
+ "learning_rate": 7.759283179451704e-05,
+ "loss": 3.5475456714630127,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05584642233857,
+ "grad_norm": 0.05074001103639603,
+ "learning_rate": 7.741933403626848e-05,
+ "loss": 3.5623722076416016,
+ "step": 2812
+ },
+ {
+ "epoch": 39.06980802792321,
+ "grad_norm": 0.05134766176342964,
+ "learning_rate": 7.724601802428805e-05,
+ "loss": 3.540435791015625,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083769633507856,
+ "grad_norm": 0.05144895985722542,
+ "learning_rate": 7.707288389821329e-05,
+ "loss": 3.6154632568359375,
+ "step": 2814
+ },
+ {
+ "epoch": 39.0977312390925,
+ "grad_norm": 0.0506402924656868,
+ "learning_rate": 7.689993179753519e-05,
+ "loss": 3.559809684753418,
+ "step": 2815
+ },
+ {
+ "epoch": 39.11169284467714,
+ "grad_norm": 0.05063097923994064,
+ "learning_rate": 7.672716186159829e-05,
+ "loss": 3.562814474105835,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12565445026178,
+ "grad_norm": 0.051375292241573334,
+ "learning_rate": 7.655457422959993e-05,
+ "loss": 3.5684101581573486,
+ "step": 2817
+ },
+ {
+ "epoch": 39.139616055846425,
+ "grad_norm": 0.0503469854593277,
+ "learning_rate": 7.638216904059122e-05,
+ "loss": 3.556791305541992,
+ "step": 2818
+ },
+ {
+ "epoch": 39.153577661431065,
+ "grad_norm": 0.050908755511045456,
+ "learning_rate": 7.620994643347559e-05,
+ "loss": 3.575982093811035,
+ "step": 2819
+ },
+ {
+ "epoch": 39.167539267015705,
+ "grad_norm": 0.049920931458473206,
+ "learning_rate": 7.60379065470098e-05,
+ "loss": 3.5394392013549805,
+ "step": 2820
+ },
+ {
+ "epoch": 39.181500872600346,
+ "grad_norm": 0.052037835121154785,
+ "learning_rate": 7.586604951980326e-05,
+ "loss": 3.5330114364624023,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19546247818499,
+ "grad_norm": 0.05024134740233421,
+ "learning_rate": 7.569437549031813e-05,
+ "loss": 3.555091142654419,
+ "step": 2822
+ },
+ {
+ "epoch": 39.20942408376963,
+ "grad_norm": 0.05144693702459335,
+ "learning_rate": 7.55228845968691e-05,
+ "loss": 3.5827114582061768,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223385689354274,
+ "grad_norm": 0.052199963480234146,
+ "learning_rate": 7.535157697762313e-05,
+ "loss": 3.559880256652832,
+ "step": 2824
+ },
+ {
+ "epoch": 39.23734729493892,
+ "grad_norm": 0.05210355296730995,
+ "learning_rate": 7.518045277059979e-05,
+ "loss": 3.5808815956115723,
+ "step": 2825
+ },
+ {
+ "epoch": 39.25130890052356,
+ "grad_norm": 0.05106557160615921,
+ "learning_rate": 7.500951211367068e-05,
+ "loss": 3.5711770057678223,
+ "step": 2826
+ },
+ {
+ "epoch": 39.2652705061082,
+ "grad_norm": 0.051606904715299606,
+ "learning_rate": 7.483875514455979e-05,
+ "loss": 3.5610320568084717,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27923211169284,
+ "grad_norm": 0.05146080255508423,
+ "learning_rate": 7.466818200084264e-05,
+ "loss": 3.558434247970581,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29319371727749,
+ "grad_norm": 0.05288682132959366,
+ "learning_rate": 7.449779281994712e-05,
+ "loss": 3.6013455390930176,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30715532286213,
+ "grad_norm": 0.05138479545712471,
+ "learning_rate": 7.43275877391528e-05,
+ "loss": 3.5789742469787598,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32111692844677,
+ "grad_norm": 0.052338190376758575,
+ "learning_rate": 7.415756689559061e-05,
+ "loss": 3.5699806213378906,
+ "step": 2831
+ },
+ {
+ "epoch": 39.33507853403141,
+ "grad_norm": 0.051882997155189514,
+ "learning_rate": 7.398773042624341e-05,
+ "loss": 3.54583740234375,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34904013961606,
+ "grad_norm": 0.05100737139582634,
+ "learning_rate": 7.381807846794532e-05,
+ "loss": 3.5269007682800293,
+ "step": 2833
+ },
+ {
+ "epoch": 39.3630017452007,
+ "grad_norm": 0.05167330801486969,
+ "learning_rate": 7.3648611157382e-05,
+ "loss": 3.547883987426758,
+ "step": 2834
+ },
+ {
+ "epoch": 39.37696335078534,
+ "grad_norm": 0.052119553089141846,
+ "learning_rate": 7.34793286310899e-05,
+ "loss": 3.570747137069702,
+ "step": 2835
+ },
+ {
+ "epoch": 39.390924956369986,
+ "grad_norm": 0.05193249508738518,
+ "learning_rate": 7.331023102545716e-05,
+ "loss": 3.5619454383850098,
+ "step": 2836
+ },
+ {
+ "epoch": 39.404886561954626,
+ "grad_norm": 0.05142345651984215,
+ "learning_rate": 7.314131847672268e-05,
+ "loss": 3.568817138671875,
+ "step": 2837
+ },
+ {
+ "epoch": 39.41884816753927,
+ "grad_norm": 0.05171925202012062,
+ "learning_rate": 7.297259112097608e-05,
+ "loss": 3.5891952514648438,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43280977312391,
+ "grad_norm": 0.051284465938806534,
+ "learning_rate": 7.280404909415801e-05,
+ "loss": 3.56705379486084,
+ "step": 2839
+ },
+ {
+ "epoch": 39.446771378708554,
+ "grad_norm": 0.05284653604030609,
+ "learning_rate": 7.263569253205973e-05,
+ "loss": 3.5296974182128906,
+ "step": 2840
+ },
+ {
+ "epoch": 39.460732984293195,
+ "grad_norm": 0.05274311453104019,
+ "learning_rate": 7.24675215703232e-05,
+ "loss": 3.5653305053710938,
+ "step": 2841
+ },
+ {
+ "epoch": 39.474694589877835,
+ "grad_norm": 0.05247433856129646,
+ "learning_rate": 7.229953634444055e-05,
+ "loss": 3.584923267364502,
+ "step": 2842
+ },
+ {
+ "epoch": 39.488656195462475,
+ "grad_norm": 0.05246812105178833,
+ "learning_rate": 7.213173698975444e-05,
+ "loss": 3.604335069656372,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50261780104712,
+ "grad_norm": 0.052823301404714584,
+ "learning_rate": 7.1964123641458e-05,
+ "loss": 3.5415382385253906,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51657940663176,
+ "grad_norm": 0.05378150939941406,
+ "learning_rate": 7.179669643459405e-05,
+ "loss": 3.5816025733947754,
+ "step": 2845
+ },
+ {
+ "epoch": 39.5305410122164,
+ "grad_norm": 0.053023822605609894,
+ "learning_rate": 7.162945550405575e-05,
+ "loss": 3.542984962463379,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544502617801044,
+ "grad_norm": 0.0532502755522728,
+ "learning_rate": 7.146240098458613e-05,
+ "loss": 3.5629453659057617,
+ "step": 2847
+ },
+ {
+ "epoch": 39.55846422338569,
+ "grad_norm": 0.05303101986646652,
+ "learning_rate": 7.129553301077801e-05,
+ "loss": 3.5725502967834473,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57242582897033,
+ "grad_norm": 0.052930813282728195,
+ "learning_rate": 7.112885171707378e-05,
+ "loss": 3.5748348236083984,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58638743455497,
+ "grad_norm": 0.0558759830892086,
+ "learning_rate": 7.096235723776562e-05,
+ "loss": 3.569337844848633,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60034904013962,
+ "grad_norm": 0.05266093835234642,
+ "learning_rate": 7.079604970699514e-05,
+ "loss": 3.579071044921875,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61431064572426,
+ "grad_norm": 0.0530242957174778,
+ "learning_rate": 7.062992925875332e-05,
+ "loss": 3.569580078125,
+ "step": 2852
+ },
+ {
+ "epoch": 39.6282722513089,
+ "grad_norm": 0.05434603616595268,
+ "learning_rate": 7.046399602688045e-05,
+ "loss": 3.590569496154785,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64223385689354,
+ "grad_norm": 0.05442731827497482,
+ "learning_rate": 7.029825014506572e-05,
+ "loss": 3.584219455718994,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65619546247819,
+ "grad_norm": 0.05413390323519707,
+ "learning_rate": 7.013269174684795e-05,
+ "loss": 3.5826475620269775,
+ "step": 2855
+ },
+ {
+ "epoch": 39.67015706806283,
+ "grad_norm": 0.053916264325380325,
+ "learning_rate": 6.996732096561422e-05,
+ "loss": 3.590533971786499,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68411867364747,
+ "grad_norm": 0.05374522507190704,
+ "learning_rate": 6.980213793460092e-05,
+ "loss": 3.5486674308776855,
+ "step": 2857
+ },
+ {
+ "epoch": 39.69808027923211,
+ "grad_norm": 0.05496302619576454,
+ "learning_rate": 6.963714278689304e-05,
+ "loss": 3.585171699523926,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712041884816756,
+ "grad_norm": 0.05460911989212036,
+ "learning_rate": 6.947233565542417e-05,
+ "loss": 3.545654296875,
+ "step": 2859
+ },
+ {
+ "epoch": 39.726003490401396,
+ "grad_norm": 0.055129971355199814,
+ "learning_rate": 6.930771667297653e-05,
+ "loss": 3.5461995601654053,
+ "step": 2860
+ },
+ {
+ "epoch": 39.73996509598604,
+ "grad_norm": 0.05326374992728233,
+ "learning_rate": 6.914328597218042e-05,
+ "loss": 3.5750136375427246,
+ "step": 2861
+ },
+ {
+ "epoch": 39.753926701570684,
+ "grad_norm": 0.05515550076961517,
+ "learning_rate": 6.8979043685515e-05,
+ "loss": 3.6057279109954834,
+ "step": 2862
+ },
+ {
+ "epoch": 39.767888307155324,
+ "grad_norm": 0.05500713735818863,
+ "learning_rate": 6.881498994530708e-05,
+ "loss": 3.5784642696380615,
+ "step": 2863
+ },
+ {
+ "epoch": 39.781849912739965,
+ "grad_norm": 0.05474477633833885,
+ "learning_rate": 6.865112488373186e-05,
+ "loss": 3.562279224395752,
+ "step": 2864
+ },
+ {
+ "epoch": 39.795811518324605,
+ "grad_norm": 0.05467807129025459,
+ "learning_rate": 6.84874486328125e-05,
+ "loss": 3.582568645477295,
+ "step": 2865
+ },
+ {
+ "epoch": 39.80977312390925,
+ "grad_norm": 0.05313905328512192,
+ "learning_rate": 6.832396132441993e-05,
+ "loss": 3.5913379192352295,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82373472949389,
+ "grad_norm": 0.05476069450378418,
+ "learning_rate": 6.816066309027309e-05,
+ "loss": 3.588045835494995,
+ "step": 2867
+ },
+ {
+ "epoch": 39.83769633507853,
+ "grad_norm": 0.05372064188122749,
+ "learning_rate": 6.799755406193815e-05,
+ "loss": 3.5555458068847656,
+ "step": 2868
+ },
+ {
+ "epoch": 39.85165794066317,
+ "grad_norm": 0.05371057987213135,
+ "learning_rate": 6.783463437082929e-05,
+ "loss": 3.5642948150634766,
+ "step": 2869
+ },
+ {
+ "epoch": 39.86561954624782,
+ "grad_norm": 0.05320272222161293,
+ "learning_rate": 6.767190414820792e-05,
+ "loss": 3.5728759765625,
+ "step": 2870
+ },
+ {
+ "epoch": 39.87958115183246,
+ "grad_norm": 0.054637160152196884,
+ "learning_rate": 6.750936352518284e-05,
+ "loss": 3.5528564453125,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8935427574171,
+ "grad_norm": 0.05465191975235939,
+ "learning_rate": 6.734701263271011e-05,
+ "loss": 3.5658421516418457,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90750436300174,
+ "grad_norm": 0.05452556163072586,
+ "learning_rate": 6.718485160159289e-05,
+ "loss": 3.5749800205230713,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92146596858639,
+ "grad_norm": 0.05434390529990196,
+ "learning_rate": 6.702288056248145e-05,
+ "loss": 3.5784506797790527,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93542757417103,
+ "grad_norm": 0.05458179861307144,
+ "learning_rate": 6.686109964587285e-05,
+ "loss": 3.5742459297180176,
+ "step": 2875
+ },
+ {
+ "epoch": 39.94938917975567,
+ "grad_norm": 0.05434086546301842,
+ "learning_rate": 6.669950898211106e-05,
+ "loss": 3.5479142665863037,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96335078534032,
+ "grad_norm": 0.054262660443782806,
+ "learning_rate": 6.653810870138676e-05,
+ "loss": 3.5872840881347656,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97731239092496,
+ "grad_norm": 0.05524756386876106,
+ "learning_rate": 6.637689893373729e-05,
+ "loss": 3.569241523742676,
+ "step": 2878
+ },
+ {
+ "epoch": 39.9912739965096,
+ "grad_norm": 0.05461305379867554,
+ "learning_rate": 6.621587980904651e-05,
+ "loss": 3.579705238342285,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.04095758497714996,
+ "learning_rate": 6.605505145704437e-05,
+ "loss": 2.2304811477661133,
+ "step": 2880
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.2294538755519283e+18,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-2880/training_args.bin b/runs/l2r90-ssl/checkpoint-2880/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..103048067ef8897d0386013ed89ce6764bad5521
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-2880/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f320fb44ea3a4a93198f1a6c2405447428148ce142cd94ee283a3c077a327c2d
+size 4792
diff --git a/runs/l2r90-ssl/checkpoint-3240/chat_template.jinja b/runs/l2r90-ssl/checkpoint-3240/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3240/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-ssl/checkpoint-3240/config.json b/runs/l2r90-ssl/checkpoint-3240/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3240/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-ssl/checkpoint-3240/generation_config.json b/runs/l2r90-ssl/checkpoint-3240/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3240/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-ssl/checkpoint-3240/model.safetensors b/runs/l2r90-ssl/checkpoint-3240/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..0f5de3415e3dbd2047733dea6da37a994ad581c2
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3240/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:20f79ca6b80f3eeb5e96d121776e636d760a93830548c8ae5b852199c4284805
+size 583356232
diff --git a/runs/l2r90-ssl/checkpoint-3240/optimizer.pt b/runs/l2r90-ssl/checkpoint-3240/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..880c886371b978688fe72032ae66206163228baf
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3240/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:cce2bc9bb855f4c7bfae15f51c3eb2cb4fb91036b3f73366faf457f24d3241c6
+size 1166827898
diff --git a/runs/l2r90-ssl/checkpoint-3240/rng_state_0.pth b/runs/l2r90-ssl/checkpoint-3240/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..6fb901f11f49c59b39a622f7767ba7f3f47a8e68
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3240/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c07d15af951fd54259f185b472ffc81cf9402e4a8ec70b56c9edd912c69850c4
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-3240/rng_state_1.pth b/runs/l2r90-ssl/checkpoint-3240/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..7c6be540f2709b4585034dcadb7428c21ac6b6e5
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3240/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9b6d8e0cbda51cdd8640c4f6ac246f9e2d927fedbdb6bbb9ef1e60d78520899d
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-3240/scheduler.pt b/runs/l2r90-ssl/checkpoint-3240/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..84479f3138d6041f59e136173e080a0ddbdf81b1
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3240/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1045dafc67457b916f466f6a7399234fef3fce01fc96958c0e146dea7784e901
+size 1064
diff --git a/runs/l2r90-ssl/checkpoint-3240/tokenizer.json b/runs/l2r90-ssl/checkpoint-3240/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3240/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-ssl/checkpoint-3240/tokenizer_config.json b/runs/l2r90-ssl/checkpoint-3240/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3240/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-3240/trainer_state.json b/runs/l2r90-ssl/checkpoint-3240/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..dce95c7d67274817cbef1741b06f782413effb0c
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3240/trainer_state.json
@@ -0,0 +1,23058 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 45.0,
+ "eval_steps": 500,
+ "global_step": 3240,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.2588357925415039,
+ "learning_rate": 0.0,
+ "loss": 3.9982247352600098,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.25455471873283386,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.998683452606201,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.2555367052555084,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.9821574687957764,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.2383691370487213,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.9531917572021484,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2023179680109024,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.9148876667022705,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.17691479623317719,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.875638961791992,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.14372293651103973,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.836796283721924,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1074419692158699,
+ "learning_rate": 4.2e-05,
+ "loss": 3.8036489486694336,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.08204353600740433,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.774808406829834,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.0632256343960762,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.7503671646118164,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.047985248267650604,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.7345473766326904,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.03675895184278488,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.721525192260742,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.02908993512392044,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.713454246520996,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.02443614974617958,
+ "learning_rate": 7.8e-05,
+ "loss": 3.7078704833984375,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02109951712191105,
+ "learning_rate": 8.4e-05,
+ "loss": 3.7034401893615723,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.018616363406181335,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.7009854316711426,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.016399752348661423,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.697768211364746,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.014256286434829235,
+ "learning_rate": 0.000102,
+ "loss": 3.6974306106567383,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.013020666316151619,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.694406509399414,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.011879238300025463,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.691807270050049,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.010290366597473621,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.6937897205352783,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.009482803754508495,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.6902384757995605,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.008425266481935978,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.686815023422241,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.007676936220377684,
+ "learning_rate": 0.000138,
+ "loss": 3.685192823410034,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.007491032127290964,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.6839442253112793,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.007428477052599192,
+ "learning_rate": 0.00015,
+ "loss": 3.6783409118652344,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.006965892389416695,
+ "learning_rate": 0.000156,
+ "loss": 3.6813433170318604,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.008106694556772709,
+ "learning_rate": 0.000162,
+ "loss": 3.6741726398468018,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.0076780058443546295,
+ "learning_rate": 0.000168,
+ "loss": 3.67673921585083,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.007904314436018467,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.6724467277526855,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.008422375656664371,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.673013925552368,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00811714306473732,
+ "learning_rate": 0.000186,
+ "loss": 3.6698389053344727,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.009063836187124252,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.668990135192871,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.01878332532942295,
+ "learning_rate": 0.000198,
+ "loss": 3.6658849716186523,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.01812782697379589,
+ "learning_rate": 0.000204,
+ "loss": 3.6697864532470703,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.020587686449289322,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.662757158279419,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.01515690516680479,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.6558732986450195,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.02510508894920349,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.6495113372802734,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.020487826317548752,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.638803243637085,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.017446046695113182,
+ "learning_rate": 0.000234,
+ "loss": 3.633307456970215,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.030667584389448166,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.6277506351470947,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.023789944127202034,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.6189804077148438,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.02660694345831871,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.610522747039795,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.023814303800463676,
+ "learning_rate": 0.000258,
+ "loss": 3.615910291671753,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.021863561123609543,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.612008571624756,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03705216199159622,
+ "learning_rate": 0.00027,
+ "loss": 3.6066887378692627,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.024143919348716736,
+ "learning_rate": 0.000276,
+ "loss": 3.5987331867218018,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.03251276910305023,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.595367431640625,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06619632989168167,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.59018611907959,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.042714253067970276,
+ "learning_rate": 0.000294,
+ "loss": 3.58981990814209,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.018893254920840263,
+ "learning_rate": 0.0003,
+ "loss": 3.578472375869751,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0749685987830162,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.5828654766082764,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.12142857164144516,
+ "learning_rate": 0.000312,
+ "loss": 3.5840892791748047,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.06944665312767029,
+ "learning_rate": 0.000318,
+ "loss": 3.580961227416992,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.04545000568032265,
+ "learning_rate": 0.000324,
+ "loss": 3.5674569606781006,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.061073049902915955,
+ "learning_rate": 0.00033,
+ "loss": 3.569991111755371,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.04133498668670654,
+ "learning_rate": 0.000336,
+ "loss": 3.5657706260681152,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.055982090532779694,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.559979200363159,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.05259208381175995,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.558039426803589,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.029976001009345055,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.5491275787353516,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.04998327046632767,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.5406670570373535,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.03230690956115723,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.5391530990600586,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.06102178618311882,
+ "learning_rate": 0.000372,
+ "loss": 3.5311684608459473,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.031715717166662216,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.518899440765381,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.06841139495372772,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.514200210571289,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.10323675721883774,
+ "learning_rate": 0.00039,
+ "loss": 3.513646125793457,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.1448228508234024,
+ "learning_rate": 0.000396,
+ "loss": 3.5110087394714355,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.12129563093185425,
+ "learning_rate": 0.000402,
+ "loss": 3.501497507095337,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.030224842950701714,
+ "learning_rate": 0.000408,
+ "loss": 3.4870352745056152,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.07492101192474365,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.482234001159668,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.03942760080099106,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.4746525287628174,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.04967200756072998,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.604069232940674,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8671882152557373,
+ "eval_runtime": 42.6602,
+ "eval_samples_per_second": 57.243,
+ "eval_steps_per_second": 0.469,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.05385325849056244,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.462052822113037,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.04870929941534996,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.459104537963867,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.08701689541339874,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.4533209800720215,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.12219952791929245,
+ "learning_rate": 0.00045,
+ "loss": 3.4558796882629395,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.0823654904961586,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.443049192428589,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.12765446305274963,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.4373486042022705,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.14215047657489777,
+ "learning_rate": 0.000468,
+ "loss": 3.449429988861084,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.03649887070059776,
+ "learning_rate": 0.000474,
+ "loss": 3.42877197265625,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09551949054002762,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.4352874755859375,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.030483486130833626,
+ "learning_rate": 0.000486,
+ "loss": 3.4125874042510986,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.0745147094130516,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.4153835773468018,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.06620422005653381,
+ "learning_rate": 0.000498,
+ "loss": 3.4053854942321777,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.06098725646734238,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.404346466064453,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.047781601548194885,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.3991761207580566,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.04409855231642723,
+ "learning_rate": 0.000516,
+ "loss": 3.396144390106201,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.04355994611978531,
+ "learning_rate": 0.000522,
+ "loss": 3.3928911685943604,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.0670008510351181,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.3831443786621094,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.0812622532248497,
+ "learning_rate": 0.000534,
+ "loss": 3.382413148880005,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.0632157102227211,
+ "learning_rate": 0.00054,
+ "loss": 3.3784892559051514,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.028686825186014175,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.3673207759857178,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.034818850457668304,
+ "learning_rate": 0.000552,
+ "loss": 3.3648972511291504,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.04509204626083374,
+ "learning_rate": 0.000558,
+ "loss": 3.35561203956604,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.054211195558309555,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.3612167835235596,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.03752049803733826,
+ "learning_rate": 0.00057,
+ "loss": 3.357103109359741,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.03620241954922676,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.3469700813293457,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.02454194985330105,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.3429036140441895,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.041030462831258774,
+ "learning_rate": 0.000588,
+ "loss": 3.338247299194336,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.03266424685716629,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.332688331604004,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.04730091616511345,
+ "learning_rate": 0.0006,
+ "loss": 3.335902452468872,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.065738245844841,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.3289852142333984,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.09062394499778748,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.3278567790985107,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.1482805460691452,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.332685947418213,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.20066532492637634,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.363180637359619,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.11500677466392517,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.3308753967285156,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.11754138022661209,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.3383703231811523,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.09672239422798157,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.3267624378204346,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.07190799713134766,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.317737579345703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.0815449133515358,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.3198161125183105,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09612327069044113,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.315814971923828,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.14795289933681488,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.317605495452881,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06667589396238327,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.3127357959747314,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.058537885546684265,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.319807767868042,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.06428155303001404,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.297541618347168,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.03881775587797165,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.297539710998535,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.061801470816135406,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.290297508239746,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.07642851769924164,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.284428596496582,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08123870939016342,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.2820234298706055,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07174684852361679,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.2697296142578125,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.04030010476708412,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.264328956604004,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.05662153288722038,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.2699337005615234,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.03761111572384834,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.255704879760742,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.043027132749557495,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.2579030990600586,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.044409338384866714,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.248547077178955,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.0239730142056942,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.249307632446289,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03628705441951752,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.2451741695404053,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.029564054682850838,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.238445997238159,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.032036975026130676,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.2320892810821533,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.030085794627666473,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.231468915939331,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.025232303887605667,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.232199192047119,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.027053389698266983,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.210745334625244,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.028464114293456078,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.210076332092285,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04616783559322357,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.2103490829467773,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.039081938564777374,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.202662467956543,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.03428936004638672,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.1997475624084473,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.04351917654275894,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.19844913482666,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.0840945914387703,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.2013347148895264,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.36558300256729126,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.26310396194458,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.09432770311832428,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.2881736755371094,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.061970267444849014,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.2840280532836914,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.05112256482243538,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.2632040977478027,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06797203421592712,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.254606246948242,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.051026541739702225,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.429918050765991,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.8117228746414185,
+ "eval_runtime": 42.8572,
+ "eval_samples_per_second": 56.98,
+ "eval_steps_per_second": 0.467,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.0608622208237648,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.2348179817199707,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.05581629276275635,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.22896146774292,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.09162136912345886,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.2343664169311523,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.13750384747982025,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.2507717609405518,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.0872517079114914,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.2440056800842285,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.07161762565374374,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.229494571685791,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04082738608121872,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.2183239459991455,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.04972497373819351,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.218075752258301,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.049254074692726135,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.213379383087158,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.056671254336833954,
+ "learning_rate": 0.000599663984012462,
+ "loss": 3.209568977355957,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.05541909113526344,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.207760810852051,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.03819117695093155,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 3.199732780456543,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.0529443584382534,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.1945419311523438,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.043010979890823364,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.185936212539673,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.07311078161001205,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.189894914627075,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06110258400440216,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 3.1911773681640625,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.06880149245262146,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 3.187437057495117,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.051326874643564224,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 3.1826558113098145,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.03363822400569916,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 3.175527572631836,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.10268989205360413,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 3.172367572784424,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.09723010659217834,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 3.1873984336853027,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.04129317030310631,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 3.165799617767334,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.05292229354381561,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 3.171780824661255,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.03873143345117569,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 3.1680874824523926,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.047018859535455704,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 3.161402940750122,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.03942955285310745,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 3.1530532836914062,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.039801858365535736,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 3.1465344429016113,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.03795155510306358,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 3.1364030838012695,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.0342850424349308,
+ "learning_rate": 0.000599379982511273,
+ "loss": 3.1391525268554688,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.03243280574679375,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 3.1331253051757812,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.029679352417588234,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 3.1346120834350586,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.030230293050408363,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 3.132615804672241,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.02845083735883236,
+ "learning_rate": 0.000599309205504528,
+ "loss": 3.1242547035217285,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.03208369016647339,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 3.119110584259033,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.040268607437610626,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 3.113833427429199,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.06951889395713806,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 3.1186914443969727,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.08879823237657547,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 3.1140987873077393,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07374681532382965,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 3.1147828102111816,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.038190409541130066,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 3.1161253452301025,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.03297388181090355,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 3.1006083488464355,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.05905177444219589,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 3.106564521789551,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.037567321211099625,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 3.0884456634521484,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.02725658379495144,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 3.09000825881958,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.04272424429655075,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 3.090528964996338,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.027587950229644775,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 3.085131883621216,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.03282274678349495,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 3.0880112648010254,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.04064026474952698,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 3.079259157180786,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.023374097421765327,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 3.0747299194335938,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.025003870949149132,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 3.0721259117126465,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.030445197597146034,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 3.0733323097229004,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.025935636833310127,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 3.070709228515625,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.021337421610951424,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 3.0627551078796387,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.040481697767972946,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 3.0529394149780273,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.047608256340026855,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 3.041020631790161,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.03356955572962761,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 3.0532307624816895,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.02101682312786579,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 3.0492985248565674,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.030098581686615944,
+ "learning_rate": 0.000598804365316586,
+ "loss": 3.0527801513671875,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.040166422724723816,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 3.0505762100219727,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.051579318940639496,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 3.0422110557556152,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.06485334783792496,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 3.0435101985931396,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.055590901523828506,
+ "learning_rate": 0.000598706872384039,
+ "loss": 3.0448739528656006,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.03183475881814957,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 3.0416858196258545,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.03995629772543907,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 3.0255980491638184,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.03907827287912369,
+ "learning_rate": 0.000598631251036868,
+ "loss": 3.032339096069336,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.027681387960910797,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 3.025439739227295,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.03470184653997421,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 3.02883243560791,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.031407181173563004,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 3.0289883613586426,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.0371522456407547,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 3.013608932495117,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.03651345893740654,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 3.0120253562927246,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.06004595756530762,
+ "learning_rate": 0.000598473577921366,
+ "loss": 3.017958402633667,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.09979518502950668,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 3.023041248321533,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.10138462483882904,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.275729179382324,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7571707367897034,
+ "eval_runtime": 42.8283,
+ "eval_samples_per_second": 57.018,
+ "eval_steps_per_second": 0.467,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.04982660710811615,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 3.016209602355957,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.06532539427280426,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 3.015043020248413,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.04369623214006424,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 3.0096778869628906,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.05180002748966217,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 3.0121543407440186,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.0440233014523983,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.9937305450439453,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.05645671859383583,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 3.0023865699768066,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.03990716487169266,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 3.0003232955932617,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.03818608075380325,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.994137763977051,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.035388972610235214,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.993170738220215,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.03162284567952156,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.9922614097595215,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.029497336596250534,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.9852986335754395,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.02586926706135273,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.9798672199249268,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.027355050668120384,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.985158920288086,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.0341978520154953,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.982581615447998,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.030727429315447807,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.9701411724090576,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.04429870471358299,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.9789254665374756,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.05775132030248642,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.9758143424987793,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.05635962635278702,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.981480121612549,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.03639192506670952,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.968752861022949,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.028748342767357826,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.9584972858428955,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.035653006285429,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.967545747756958,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.03697885945439339,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.9578099250793457,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.03518654406070709,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.9680328369140625,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.034762222319841385,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.9670262336730957,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03813016042113304,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.957522392272949,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.04372904822230339,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.9540514945983887,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.06646446138620377,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.951028823852539,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.07831092923879623,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.95898699760437,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.05779588967561722,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.965884208679199,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.04414296895265579,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.9510560035705566,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.04797815904021263,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.955803394317627,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.04482145980000496,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.954059362411499,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.037104446440935135,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.94792103767395,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.03616701066493988,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.9500932693481445,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.03882797807455063,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.9558305740356445,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.04356204345822334,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.9237029552459717,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.04535257816314697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.925853729248047,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.04519587755203247,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.924984931945801,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.055922407656908035,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.932561159133911,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.06330762803554535,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.9388322830200195,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.0517115443944931,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.9272942543029785,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.057369694113731384,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.9201431274414062,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.05687880516052246,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.922886371612549,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.055525410920381546,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.9211201667785645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04503652825951576,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.9164583683013916,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.03385302051901817,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.9145395755767822,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03564384952187538,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.916003704071045,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.04188712686300278,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.915839433670044,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03657370060682297,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.903120279312134,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.03126626834273338,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.895824432373047,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.04196290299296379,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.9093422889709473,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.1706087589263916,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.9183316230773926,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.05810421332716942,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.911245107650757,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.061330314725637436,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.9159231185913086,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.06062281131744385,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.9211535453796387,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.05590800568461418,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.9140496253967285,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.04496239125728607,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.9040920734405518,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.03835335746407509,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.9026548862457275,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.05137286335229874,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.889634132385254,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.04091101884841919,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.886838674545288,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.04076292738318443,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.892918109893799,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.045437734574079514,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.893305778503418,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.041114673018455505,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.88517165184021,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.03600773215293884,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.8779795169830322,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.04311220720410347,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.87247896194458,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.04818756878376007,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.889103412628174,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.053460877388715744,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.8899683952331543,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.06763825565576553,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.8823537826538086,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.07304801046848297,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.877340316772461,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.061337269842624664,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.8820111751556396,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.04660911113023758,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.868886709213257,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.053929202258586884,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.159337043762207,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.721762478351593,
+ "eval_runtime": 42.7977,
+ "eval_samples_per_second": 57.059,
+ "eval_steps_per_second": 0.467,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.04259369149804115,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.859579086303711,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.044425517320632935,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.8681139945983887,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.04301869496703148,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.853911876678467,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.04574296995997429,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.8724427223205566,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.050247322767972946,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.8528172969818115,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.046683840453624725,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.853405714035034,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.05966169387102127,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.8605127334594727,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.08935574442148209,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.862339973449707,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.08019179105758667,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.8655967712402344,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.04753892868757248,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.8657939434051514,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06868117302656174,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.861976146697998,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.07040470093488693,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.8562135696411133,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.05683179199695587,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.8576629161834717,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.04918001592159271,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.861943244934082,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.04921915754675865,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.8495378494262695,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.037651967257261276,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.8428618907928467,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.03522057458758354,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.8372769355773926,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.03111186809837818,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.8377151489257812,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.03906773030757904,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.846196174621582,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.02546556107699871,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.8288016319274902,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.03119821846485138,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.8334617614746094,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.030583377927541733,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.8209362030029297,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.029693076387047768,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.8267486095428467,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.028978517279028893,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.819244861602783,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.027871523052453995,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.8289647102355957,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.029287064447999,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.8235747814178467,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.024990517646074295,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.819016218185425,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.030980689451098442,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.828155040740967,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03721801936626434,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.813418388366699,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03738148882985115,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.808328866958618,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.03580411896109581,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.812070846557617,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.039036594331264496,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.807159423828125,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.044666968286037445,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.799992561340332,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.055589817464351654,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.804142951965332,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.06139020994305611,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.8182730674743652,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.07769318670034409,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.8166518211364746,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.06521332263946533,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.8142249584198,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.09083619713783264,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.8166933059692383,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.09737467020750046,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.8111555576324463,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.07148188352584839,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.8168413639068604,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.055516090244054794,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.8183553218841553,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.05338391661643982,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.8127028942108154,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.040785472840070724,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.8124775886535645,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.037135496735572815,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.8065357208251953,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.032455578446388245,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.798459053039551,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.03465662896633148,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.798649549484253,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.032026033848524094,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.7906997203826904,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.026019956916570663,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.791969060897827,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.030298635363578796,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.784593105316162,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.028147678822278976,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.782454013824463,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.023933131247758865,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.787851333618164,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.026789527386426926,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.788067579269409,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.024554884061217308,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.777033567428589,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.028679484501481056,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.786342144012451,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.026410218328237534,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.7790610790252686,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.027413051575422287,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.778944969177246,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.02908242493867874,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.763850212097168,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.029632994905114174,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.7672982215881348,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.027458716183900833,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.7702178955078125,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.03263958916068077,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.7581803798675537,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04427146911621094,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.768141508102417,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.07142394036054611,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.75819730758667,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.08833584934473038,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.782716751098633,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.08168833702802658,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.7763240337371826,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07253874838352203,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.7717690467834473,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.08186902105808258,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.767303466796875,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.05854064226150513,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.761544704437256,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.04602917283773422,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.7669177055358887,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.03920970857143402,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.766085624694824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.04551195353269577,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.7623450756073,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.040861159563064575,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.7471840381622314,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.026907291263341904,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 2.0710244178771973,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6934371590614319,
+ "eval_runtime": 42.6015,
+ "eval_samples_per_second": 57.322,
+ "eval_steps_per_second": 0.469,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.03801899403333664,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.739086627960205,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.03707238286733627,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.741638422012329,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.0387551411986351,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.740777015686035,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.028010359033942223,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.7353010177612305,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.026474518701434135,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.73842453956604,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.02920856513082981,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.736295461654663,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.026074763387441635,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.7295751571655273,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.026862047612667084,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.723040819168091,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.026300370693206787,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.7199485301971436,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.026030950248241425,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.7276217937469482,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.028342435136437416,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.7193045616149902,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.032135091722011566,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.7143118381500244,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.03068603202700615,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.7191050052642822,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.035828713327646255,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.718085765838623,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.03343002125620842,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.707131862640381,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.035051796585321426,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.7091879844665527,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.03395913913846016,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6980862617492676,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.034198079258203506,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.699704647064209,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.0322839580476284,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.702791213989258,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.041531339287757874,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.7018051147460938,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.040806882083415985,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.7096598148345947,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.04156940430402756,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6956775188446045,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.05608797073364258,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.70127272605896,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0729481428861618,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.708817720413208,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.0905836969614029,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.718085289001465,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.08345075696706772,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.7235910892486572,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.08599896728992462,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.7178783416748047,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.07301691174507141,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.714029312133789,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.08231015503406525,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.7196826934814453,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.08286412060260773,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.7176082134246826,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.08590883016586304,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.7242164611816406,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.07531397044658661,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.7225024700164795,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.04711190238595009,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.695887327194214,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.05076327919960022,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.7029781341552734,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.04495669901371002,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.7057900428771973,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.04312257468700409,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.697537899017334,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.043580975383520126,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.67315673828125,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.03360886126756668,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.6883413791656494,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.030360296368598938,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.700927257537842,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.0310810673981905,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.6917803287506104,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.02703232690691948,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.6667580604553223,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.02756045199930668,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.6702821254730225,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.02611083723604679,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.669790744781494,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.02615324780344963,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.6672821044921875,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.026000119745731354,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.6671881675720215,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.02882794849574566,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.67586088180542,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03327213600277901,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.6618690490722656,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.032920945435762405,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.665444850921631,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.02812962979078293,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.6680707931518555,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.025886602699756622,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.652898073196411,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.025222789496183395,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.6560399532318115,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.023131893947720528,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.66127347946167,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.02513662725687027,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.6463732719421387,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.027243830263614655,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.6507630348205566,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.025437965989112854,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.6485800743103027,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.02770340070128441,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.657241106033325,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.03594067320227623,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.6281137466430664,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04486451297998428,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.6340909004211426,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.05102410912513733,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.6463394165039062,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04514491558074951,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.636554002761841,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04469453915953636,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.6453680992126465,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.05421523004770279,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.646010398864746,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04779462888836861,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.6315622329711914,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.06697724014520645,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.6416730880737305,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.07520416378974915,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.641632318496704,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.055692676454782486,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.6289877891540527,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.04539895057678223,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.6363251209259033,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.04829065501689911,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.641608476638794,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.041314542293548584,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.6269643306732178,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.04100198298692703,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.6243174076080322,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.04105962812900543,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.626585006713867,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.03172122687101364,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9550899267196655,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6614077091217041,
+ "eval_runtime": 42.9009,
+ "eval_samples_per_second": 56.922,
+ "eval_steps_per_second": 0.466,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.03258531913161278,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.612126588821411,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.040250636637210846,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.6025328636169434,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.034234337508678436,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.6085972785949707,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.03635786101222038,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.6126222610473633,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.039404451847076416,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.6035327911376953,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.038504838943481445,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5985798835754395,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.03744236007332802,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5873115062713623,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.034184787422418594,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.5899314880371094,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.03766429424285889,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.602734327316284,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.03741588443517685,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.592207193374634,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.03187853842973709,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5850372314453125,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.038447409868240356,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.592654228210449,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.04046263173222542,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.5881710052490234,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.03872602432966232,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5881471633911133,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.03960775211453438,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5826945304870605,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.033132150769233704,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.584038734436035,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.029504230245947838,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.584357261657715,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.03665253147482872,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5677168369293213,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.037208545953035355,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.560795783996582,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.038244519382715225,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.591458559036255,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.043307431042194366,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.5642588138580322,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.055119726806879044,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.57377290725708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.053323689848184586,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5767312049865723,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.0502864345908165,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.562190055847168,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.05568842589855194,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.565556049346924,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.061149828135967255,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5883288383483887,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.059933699667453766,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.5710184574127197,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.06041933596134186,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.572603225708008,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.05482695251703262,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.5667271614074707,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.04712968319654465,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.554041624069214,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.04842432215809822,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.5660672187805176,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.04273047298192978,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.554839849472046,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.05356749892234802,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.5521483421325684,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.06062676012516022,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.5551886558532715,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.05431462824344635,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5654542446136475,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.0505850687623024,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5648036003112793,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.04696856811642647,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.5559377670288086,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.042585279792547226,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.5660319328308105,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.04026484116911888,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.551206588745117,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03382421284914017,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.556361198425293,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.031177924945950508,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.5413894653320312,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.028217196464538574,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.5600497722625732,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.027062954381108284,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5447568893432617,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.03001212701201439,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.540614604949951,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.028869854286313057,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.544583797454834,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.02869918756186962,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.5471596717834473,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03359716013073921,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5416982173919678,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.04000182822346687,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.534921407699585,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.043030042201280594,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5368564128875732,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.04329848289489746,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.539764881134033,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.03740967437624931,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.5246665477752686,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.03733935207128525,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.529362916946411,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.039609868079423904,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.542544364929199,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.03937261551618576,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5366649627685547,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.03261927142739296,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.5250535011291504,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.03904294595122337,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.520721435546875,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.03913239762187004,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.5122013092041016,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.03610851988196373,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.5212767124176025,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.03944548964500427,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.537186622619629,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04762452840805054,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.5204620361328125,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04427091404795647,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5286149978637695,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04366641491651535,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.5398361682891846,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.04644155502319336,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.519331932067871,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.04372382536530495,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.531022787094116,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.04404324293136597,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.521758794784546,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.042115990072488785,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.5146074295043945,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.03796013817191124,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.5040054321289062,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.04327977076172829,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.5209312438964844,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04150385782122612,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.515871286392212,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.04465346783399582,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.5247764587402344,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.05690346285700798,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.501911163330078,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04907930642366409,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8968371152877808,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6358620524406433,
+ "eval_runtime": 42.9991,
+ "eval_samples_per_second": 56.792,
+ "eval_steps_per_second": 0.465,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.03755255416035652,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.486576795578003,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.040234558284282684,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.501087188720703,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.041416723281145096,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4942233562469482,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.039655014872550964,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.503453254699707,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.041707683354616165,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.490978717803955,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.04545537009835243,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4757118225097656,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.04328389838337898,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.504516839981079,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.043312299996614456,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.4920215606689453,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.046466976404190063,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.494760513305664,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.041762981563806534,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4875552654266357,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.03713781014084816,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.495652675628662,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.040603458881378174,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4821407794952393,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.037969645112752914,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.484981060028076,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.03652074560523033,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4736409187316895,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.03783264383673668,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.483346700668335,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.03404324874281883,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.48856782913208,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.03294725716114044,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.483706474304199,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.0347001850605011,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.474888324737549,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.0359107106924057,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.478990077972412,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.03738103806972504,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.4709420204162598,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.03722188249230385,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.465726852416992,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.034907322376966476,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.482231855392456,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.035695310682058334,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4785828590393066,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.04009179025888443,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.4795875549316406,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.038996778428554535,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.478696823120117,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04082407429814339,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.465200901031494,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04856136813759804,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4605681896209717,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.05671004578471184,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.47277569770813,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.0481606163084507,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.476151466369629,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.037689510732889175,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.466142177581787,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.05005146935582161,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.4798035621643066,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.04745829105377197,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.46163272857666,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.04638158157467842,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4728071689605713,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.04530330374836922,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.4621877670288086,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.03971488028764725,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.474224090576172,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.04521876201033592,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.46122670173645,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03923020884394646,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.4603679180145264,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03986494988203049,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.4589664936065674,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.04280020296573639,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.4624319076538086,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.03930043429136276,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.47080135345459,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.0314876064658165,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.458728790283203,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.03189660236239433,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4467148780822754,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03264275938272476,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4613699913024902,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03021484799683094,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.4535279273986816,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.030740290880203247,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.4621875286102295,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03096810355782509,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.458500385284424,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03105205111205578,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.461958646774292,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.03330068290233612,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.449777364730835,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.032831475138664246,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.4616191387176514,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.0387776717543602,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.445256471633911,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.041803423315286636,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.4545960426330566,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.04148292914032936,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.433408737182617,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.04648579657077789,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.449134349822998,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.04186409339308739,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.4449565410614014,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.038569554686546326,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.443675994873047,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03908300772309303,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4626526832580566,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.038897737860679626,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.438591957092285,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.0360480472445488,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.436077117919922,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0429830439388752,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4384970664978027,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.04465422406792641,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.458029270172119,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.037096526473760605,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.436816930770874,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.047823466360569,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.44439959526062,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.05929310992360115,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.4599215984344482,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.053512319922447205,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4387903213500977,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.042625587433576584,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4359588623046875,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.056042976677417755,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.4441447257995605,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05813189223408699,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4399144649505615,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.06313478201627731,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.453447103500366,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.05281798169016838,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.4428963661193848,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.03991910442709923,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4359219074249268,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.047981832176446915,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4506607055664062,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.03646523132920265,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8225739002227783,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6196661591529846,
+ "eval_runtime": 43.8106,
+ "eval_samples_per_second": 55.74,
+ "eval_steps_per_second": 0.457,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.03673036769032478,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.4227209091186523,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.044547371566295624,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.4073169231414795,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.04089285805821419,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4279446601867676,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.03538106009364128,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.4234392642974854,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.03152978792786598,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4214606285095215,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.03324202075600624,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.4085447788238525,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.03352442383766174,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.417776107788086,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.03582802042365074,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.411144256591797,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.0351114459335804,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.405872344970703,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.038898173719644547,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4029338359832764,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.03926656022667885,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4100089073181152,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.04285634309053421,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.410360336303711,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.04179581627249718,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4093637466430664,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.03356701880693436,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4118056297302246,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.038140635937452316,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.4154677391052246,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.04468780383467674,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416846513748169,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.062421441078186035,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4157638549804688,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06317313015460968,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.418593406677246,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.049498945474624634,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4204845428466797,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.046322405338287354,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.403489589691162,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.053081002086400986,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.4148740768432617,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05222945287823677,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.416757583618164,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04893574118614197,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4241795539855957,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04951956123113632,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.401386260986328,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04019852355122566,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.4168622493743896,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.03625848889350891,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4026846885681152,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.03464345261454582,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4021215438842773,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.03750818222761154,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4113571643829346,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.0360100194811821,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.3977622985839844,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.03464623540639877,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.413273334503174,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.03193575143814087,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.423659324645996,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.02710454724729061,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.396085262298584,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.02830432914197445,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.420255422592163,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.029717663303017616,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.3958120346069336,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.030005933716893196,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3908543586730957,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.029054716229438782,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.3873138427734375,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.031556472182273865,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3892760276794434,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03518872708082199,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4113478660583496,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.03471249341964722,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.384857654571533,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03223155066370964,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3840765953063965,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.030012644827365875,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.388247489929199,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.03162285313010216,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.3987512588500977,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.03700442612171173,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.3993430137634277,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.03710822016000748,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.387359380722046,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.028411181643605232,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3925249576568604,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031246855854988098,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.38845157623291,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.034664880484342575,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.392436981201172,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.03354701027274132,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.390890121459961,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.03527764603495598,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.3954997062683105,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03325049206614494,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3847177028656006,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.028021875768899918,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.385910987854004,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.02964954823255539,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3787484169006348,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.03068520873785019,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.400784969329834,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03416873514652252,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.391033411026001,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.03534875437617302,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.3796474933624268,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.03732474148273468,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3831934928894043,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.03648678958415985,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3785600662231445,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03801294043660164,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.387587070465088,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.04186025261878967,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.3893089294433594,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.04416101053357124,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3852195739746094,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.04137178510427475,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.391308307647705,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.05010323226451874,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.3906192779541016,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.049587685614824295,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.3846192359924316,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.04125971347093582,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.384507179260254,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.039676666259765625,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.38167142868042,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.038775283843278885,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.3789167404174805,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.03634221479296684,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.383060932159424,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03464275971055031,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.388552665710449,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.031778786331415176,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.3903298377990723,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.027904309332370758,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.377493381500244,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.028964119032025337,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.381959915161133,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.0233962032943964,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7810559272766113,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6070130467414856,
+ "eval_runtime": 50.871,
+ "eval_samples_per_second": 48.004,
+ "eval_steps_per_second": 0.393,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.030209315940737724,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.3569841384887695,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.04112662002444267,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.376680374145508,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.055858466774225235,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.3621644973754883,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.06746986508369446,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3607563972473145,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.06736797094345093,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3548707962036133,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.0562627948820591,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.364654541015625,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.051833927631378174,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.3748884201049805,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.04824110120534897,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.3502087593078613,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.05518793687224388,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3687515258789062,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.06564056873321533,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.367131233215332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.0605909489095211,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.3589868545532227,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.05744732916355133,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.370117664337158,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.06430432945489883,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.3703484535217285,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.06186915934085846,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3720293045043945,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.05977998673915863,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.3512699604034424,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.06413097679615021,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.377257823944092,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.053935080766677856,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.350717544555664,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.049110863357782364,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.3688902854919434,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.04313219338655472,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3374459743499756,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.039556123316287994,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3482556343078613,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.03796547278761864,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3547937870025635,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.03646104037761688,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.344686269760132,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.03355357423424721,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.3588008880615234,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.03004707582294941,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.352043628692627,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.030186176300048828,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3500709533691406,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.029748762026429176,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.3523058891296387,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.03238696977496147,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3635129928588867,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.032599564641714096,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.346269130706787,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.028626957908272743,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.357151746749878,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.02909751422703266,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3557677268981934,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.029951099306344986,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.34753680229187,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.029768310487270355,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3408985137939453,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.03005973994731903,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.362542152404785,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03360462561249733,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.352529287338257,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.034002047032117844,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.3334460258483887,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.03292316198348999,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3446991443634033,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.0337117575109005,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.350605010986328,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03428212180733681,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.357408046722412,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.03169601410627365,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.349088191986084,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.03389411047101021,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.361576795578003,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.036497049033641815,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.35420560836792,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03574785590171814,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3678526878356934,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.030757993459701538,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3520517349243164,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.0289121363312006,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.337864637374878,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.028012607246637344,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.353999614715576,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.027558492496609688,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.342132568359375,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.02871510572731495,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3425655364990234,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03070886805653572,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3401665687561035,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.03056657873094082,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.345277786254883,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.030907969921827316,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.346494674682617,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0318334735929966,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.3390543460845947,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.029700733721256256,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.3398728370666504,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.029687359929084778,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.3474364280700684,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.027234600856900215,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3457484245300293,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.02823043428361416,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.337700366973877,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03002646192908287,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.335954189300537,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.027246717363595963,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.330207347869873,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.029703574255108833,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.328280448913574,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.028796236962080002,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3448386192321777,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.027944380417466164,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.320234537124634,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.030626650899648666,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.334698438644409,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.029164662584662437,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3450326919555664,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03034045360982418,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.3420588970184326,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.02856968529522419,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3396072387695312,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.02892562374472618,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.340027332305908,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.02938242256641388,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.33550763130188,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.029261089861392975,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3246421813964844,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.030340535566210747,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.3465514183044434,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.033335570245981216,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.3467206954956055,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.03452218323945999,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.343702793121338,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.03326255828142166,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.337631940841675,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025433633476495743,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7402067184448242,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03175730258226395,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.62623405456543,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.04780422896146774,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.623239517211914,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.0504758283495903,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.6369194984436035,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.04481920227408409,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.57640266418457,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0440552793443203,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.630992889404297,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0503336526453495,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.627910137176514,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.053885314613580704,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.637897491455078,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.05481956899166107,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.610715866088867,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.050510745495557785,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.639884948730469,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.04579266160726547,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.590188980102539,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.04817410930991173,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.636295318603516,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.058862220495939255,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.614317417144775,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.06328638643026352,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.655928611755371,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.058752477169036865,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.652004241943359,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.06049015372991562,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.6513800621032715,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.058941151946783066,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.657688140869141,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.05542639642953873,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.637376308441162,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06024057790637016,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.63592529296875,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.06040216609835625,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.6275787353515625,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.059600673615932465,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.652995586395264,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.05379730463027954,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.597358226776123,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05355260521173477,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.6106367111206055,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.06649646162986755,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.654443740844727,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.058309994637966156,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.6869659423828125,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.05403570458292961,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.654750823974609,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04678157716989517,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.653955459594727,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04297948628664017,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.644915580749512,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.04235142841935158,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.662282943725586,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.0405130498111248,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.64569091796875,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.03351007401943207,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.653165817260742,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.03182631731033325,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.606291770935059,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.031552448868751526,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.642643928527832,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027941036969423294,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.623664379119873,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.027744563296437263,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.605886459350586,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02962736040353775,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.606762886047363,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.027251560240983963,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.620492935180664,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.027894044294953346,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.613903999328613,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.02778882533311844,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.634457588195801,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.028082096949219704,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.610980033874512,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.026210976764559746,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.612090110778809,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.026537755504250526,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.581171035766602,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.028709523379802704,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.628799915313721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.028233813121914864,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.586182117462158,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.02966046892106533,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.62320613861084,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.03198636695742607,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.6018524169921875,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.033404890447854996,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.610121726989746,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.029217196628451347,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.601312160491943,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.029335757717490196,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.620820999145508,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.031007781624794006,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.626503944396973,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.028453458100557327,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.589323997497559,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02526312880218029,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.598350524902344,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.027317512780427933,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.62388277053833,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02710605598986149,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.606523513793945,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.027491388842463493,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.586644172668457,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.027158567681908607,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.606039047241211,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.0298276599496603,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.593776702880859,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.026622479781508446,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.59510612487793,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.028238043189048767,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.5899505615234375,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.029481008648872375,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.6092071533203125,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.030138323083519936,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.600069046020508,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.02951825223863125,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.600125789642334,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.029242202639579773,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.590723037719727,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.02941613830626011,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.623342514038086,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.029764151200652122,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.631185531616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.031503792852163315,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.605094909667969,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.03184918314218521,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.592767715454102,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.03223618119955063,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.5930585861206055,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.0340263694524765,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.609347343444824,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.03245621174573898,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.598440170288086,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.03204867243766785,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.619702339172363,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.030866602435708046,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.609165191650391,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.01960863545536995,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8655829429626465,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.593852162361145,
+ "eval_runtime": 60.0971,
+ "eval_samples_per_second": 40.634,
+ "eval_steps_per_second": 0.649,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03552858531475067,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.539240837097168,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.0536109134554863,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.52609395980835,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.061411239206790924,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.536109447479248,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.06352879852056503,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.546850204467773,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.05683520436286926,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.563828468322754,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.05522637069225311,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.550929069519043,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.058990392833948135,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.554411888122559,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.057443976402282715,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.545921325683594,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.05383576825261116,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.5481414794921875,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04797440022230148,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.55556583404541,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04923664405941963,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.545513153076172,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.056191347539424896,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.536705017089844,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.06088989973068237,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.570624351501465,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.04802761599421501,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.526930809020996,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.04018320515751839,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.562432289123535,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.04280981421470642,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.559737205505371,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.04258878529071808,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.561749458312988,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.04411563277244568,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.562010765075684,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.041077401489019394,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.573984622955322,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.04125386103987694,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.550564289093018,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.041974026709795,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.552739143371582,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03689907118678093,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.540107727050781,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03586242347955704,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.554999351501465,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.036602675914764404,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.546337127685547,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.03707926720380783,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.543377876281738,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.03322125971317291,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.574325084686279,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03359243646264076,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.5613603591918945,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.0314965583384037,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.530994415283203,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.031806789338588715,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.56138801574707,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.034988485276699066,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.553149700164795,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.037627559155225754,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.555401802062988,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03723530471324921,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.52452278137207,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.036220453679561615,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.54799747467041,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03228041157126427,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.574814319610596,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.030872676521539688,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.539666652679443,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.033935241401195526,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.579252243041992,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.034338828176259995,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.538324356079102,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.03441951423883438,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.540358543395996,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03328141197562218,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.520340919494629,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03224223107099533,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.5646209716796875,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.02999890223145485,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.504915714263916,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.03014362044632435,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.516615867614746,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.029469173401594162,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.560475826263428,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.027887821197509766,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.586333751678467,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.029970254749059677,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.530396461486816,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.031648170202970505,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.554404258728027,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.0325581431388855,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.544102668762207,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.03555625304579735,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.543701171875,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.033053912222385406,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.508375644683838,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.03059076890349388,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.545750617980957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.029640182852745056,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.54715633392334,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030139951035380363,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.533708572387695,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029753828421235085,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.543603897094727,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.029481034725904465,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.534200668334961,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.028991524130105972,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.5683674812316895,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.028145765885710716,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.543053150177002,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.02999844029545784,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.5482354164123535,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03134117275476456,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.5360107421875,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.029277140274643898,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.5729827880859375,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028516866266727448,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.543548583984375,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028387080878019333,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.552616596221924,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.028977863490581512,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.5456695556640625,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.028010815382003784,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.534921646118164,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.030222605913877487,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.561526298522949,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.03140007704496384,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.500124931335449,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.029269184917211533,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.536653995513916,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.02935507707297802,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.521398544311523,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02801917865872383,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.527442455291748,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.027995768934488297,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.556139945983887,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.028441084548830986,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.555442810058594,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.029100807383656502,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.532083511352539,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020385215058922768,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.8488006591796875,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5902590751647949,
+ "eval_runtime": 60.1424,
+ "eval_samples_per_second": 40.604,
+ "eval_steps_per_second": 0.648,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03513244539499283,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.486500263214111,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.052679311484098434,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.489691734313965,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.06171489134430885,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.471555709838867,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.06164781004190445,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.486891746520996,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.06410388648509979,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.494929313659668,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06399523466825485,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.5334320068359375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.0636502280831337,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.5039381980896,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.06139931082725525,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.512765884399414,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.07460282742977142,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.477229118347168,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.075187548995018,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.484212398529053,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.07584068924188614,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.495598793029785,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.07192158699035645,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.507598400115967,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.05366869643330574,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.514532089233398,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.06397061794996262,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.512920379638672,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.06391509622335434,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.465702056884766,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.059490568935871124,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.5351881980896,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.057501111179590225,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.510270595550537,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.056947533041238785,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.509241104125977,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.050393324345350266,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.50428581237793,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.057608287781476974,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.474919319152832,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.054470717906951904,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.521528244018555,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.054405540227890015,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.510211944580078,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.051328495144844055,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.480720520019531,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.045930180698633194,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.503024101257324,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.04298274591565132,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.49046516418457,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.0406254380941391,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.544064521789551,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.03557496517896652,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.496966361999512,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.0364493802189827,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.5047502517700195,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.034521300345659256,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.485365867614746,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.03305799141526222,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.484884738922119,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.033053625375032425,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.478784561157227,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03426017984747887,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.482428073883057,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.03638783469796181,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.433219909667969,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.03597250208258629,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.522584915161133,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.03086097538471222,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.500803470611572,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.029642771929502487,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.444099426269531,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.03190716728568077,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.493034362792969,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.0312873050570488,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.487887382507324,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03201943263411522,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.479177474975586,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.033088769763708115,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.5020246505737305,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.02975946106016636,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.455426216125488,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03213953226804733,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.501615047454834,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.03262908384203911,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.485121250152588,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.031102091073989868,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.488893508911133,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.02945449948310852,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.490121841430664,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.033680472522974014,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.51833438873291,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.03277525305747986,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.480165481567383,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.03303041681647301,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.501155376434326,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.034927863627672195,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.513175010681152,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.03098093718290329,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.503798484802246,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.029513906687498093,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.49006986618042,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029630785807967186,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.522514343261719,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.030888330191373825,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.4895734786987305,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.032392196357250214,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.471953868865967,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.03215661272406578,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.48297643661499,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.03346327692270279,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.518160820007324,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.031311362981796265,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.484991550445557,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.03056488186120987,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.525470733642578,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.03043230250477791,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.510650634765625,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.02916652522981167,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.496840953826904,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.030673664063215256,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.494868278503418,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030033906921744347,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.479182720184326,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.028828797861933708,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.471983909606934,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028316056355834007,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.5049147605896,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.03158092871308327,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.449270248413086,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.03041706420481205,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.508289813995361,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.02967553399503231,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.465579032897949,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030721619725227356,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.471003532409668,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03066832385957241,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.488987445831299,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.032297998666763306,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.491018295288086,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03221398591995239,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.471237659454346,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.021120507270097733,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.8080053329467773,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5888046026229858,
+ "eval_runtime": 59.1638,
+ "eval_samples_per_second": 41.275,
+ "eval_steps_per_second": 0.659,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.03799329698085785,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.401787757873535,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.05785121023654938,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.408820629119873,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.05964238941669464,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.427263259887695,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05466582626104355,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.403415679931641,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.05613672360777855,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.413750648498535,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.05396881327033043,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.41916561126709,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.05689029023051262,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.395590782165527,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.056090328842401505,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.421253204345703,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.0473211295902729,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.385158061981201,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.04464765638113022,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.431746959686279,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.046421997249126434,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.428816795349121,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.05039866268634796,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.452417850494385,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.05491664633154869,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.447483062744141,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05866196006536484,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.434992790222168,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.05532094091176987,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.450716018676758,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.051083534955978394,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.402632713317871,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.03815016150474548,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.433640956878662,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04203395918011665,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.415888786315918,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.043729111552238464,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.423260688781738,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.039426498115062714,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.428949356079102,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.03837858885526657,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.434070587158203,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.038315922021865845,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.471246719360352,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.03963236138224602,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.4434990882873535,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04027022793889046,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.436992645263672,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.0409243181347847,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.426936149597168,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03404631093144417,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.43462610244751,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.034393735229969025,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.437409400939941,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03635180741548538,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.451286315917969,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03349733352661133,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.419209003448486,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.03238677605986595,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.456454277038574,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.034473925828933716,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.3988728523254395,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.03455805033445358,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.434884071350098,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.037486057728528976,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.466741561889648,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.03617125004529953,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.429305076599121,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.033646389842033386,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.407780170440674,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03401118144392967,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.42977237701416,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.036534134298563004,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.451094627380371,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.0384049266576767,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.414777755737305,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03460627421736717,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.420644760131836,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.033746346831321716,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.461175918579102,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.037153810262680054,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.444699287414551,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03283539041876793,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.431022644042969,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03248969465494156,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.447571277618408,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03327852487564087,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.42236328125,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.031114818528294563,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.418244361877441,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.03231207653880119,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.455004692077637,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.030702251940965652,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.483610153198242,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031027289107441902,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.4470109939575195,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.03129073604941368,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.452799320220947,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.03634301945567131,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4661664962768555,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03718913346529007,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.412006855010986,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.034018341451883316,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.436755180358887,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032502319663763046,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.452866554260254,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.032902952283620834,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.449224472045898,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.0340626984834671,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.445425510406494,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.033442508429288864,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.432218551635742,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.03489428386092186,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.4205827713012695,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.031868621706962585,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.414196968078613,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.032986633479595184,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.46116304397583,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.03348355367779732,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.424936771392822,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03174544870853424,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.434438705444336,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.03293656185269356,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.440502166748047,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.035124607384204865,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.440943717956543,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03438971936702728,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.431408882141113,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.03537093102931976,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.43867301940918,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.03267325833439827,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.462226867675781,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.030995232984423637,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.433740615844727,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.03384929150342941,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.508179664611816,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.032025814056396484,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.407378196716309,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03288578242063522,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.455531120300293,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.032061848789453506,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.452212333679199,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022320276126265526,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7567386627197266,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5878658294677734,
+ "eval_runtime": 59.5054,
+ "eval_samples_per_second": 41.038,
+ "eval_steps_per_second": 0.655,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.039114587008953094,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.347412586212158,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.05164389684796333,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.391279220581055,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055882278829813004,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.355602264404297,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.05625041201710701,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3934855461120605,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05861108750104904,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.364226341247559,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.052431970834732056,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.363334655761719,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.06147193908691406,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.384820938110352,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.06701470166444778,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.373311996459961,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.06714873015880585,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.32745361328125,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.06684907525777817,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.373660087585449,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.05738082900643349,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.362325668334961,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05306725949048996,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.351685047149658,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.05831029266119003,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.374813079833984,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05547972396016121,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.361842632293701,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.048057056963443756,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.383205413818359,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.04682513326406479,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.365846157073975,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.05001377314329147,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.362151145935059,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.05140263959765434,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.376064300537109,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.04640384018421173,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.363801002502441,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04175020009279251,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.3477983474731445,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.04134776070713997,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.337469100952148,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.042178574949502945,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.351490020751953,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.03989115729928017,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.40725040435791,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04086652770638466,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.341119766235352,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.03931201249361038,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.3786468505859375,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.03855905309319496,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.3701605796813965,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.03803952783346176,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.401841163635254,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.036908265203237534,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.410647392272949,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.0398186556994915,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.380431175231934,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03899035602807999,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.374293804168701,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.03946344554424286,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.396918773651123,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.03738604858517647,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.404016017913818,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.03579043596982956,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.396914958953857,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.037139154970645905,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.407739639282227,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.03581099584698677,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.411419868469238,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.03548453748226166,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.332690238952637,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.037561237812042236,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.381314277648926,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.035827480256557465,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.371536731719971,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03538314625620842,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.395467758178711,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.036672890186309814,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.420590400695801,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.038981080055236816,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.395296096801758,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.038234200328588486,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.415799140930176,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.038830097764730453,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.413394927978516,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03990919142961502,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.39515495300293,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.036424681544303894,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.39730978012085,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03488154709339142,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.403696060180664,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.0363851860165596,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.368661403656006,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.035461440682411194,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.396518707275391,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03647598251700401,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.440667152404785,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.03667188808321953,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.3781352043151855,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034516457468271255,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.414007186889648,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.03536032512784004,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.42198371887207,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.037454646080732346,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.409174919128418,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03632275015115738,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3547682762146,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.038922134786844254,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.39993143081665,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.03562692925333977,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.409564971923828,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.033203717321157455,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.421782493591309,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03830202296376228,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.430092811584473,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.03835884481668472,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.379505634307861,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.037440430372953415,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.403937339782715,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03732272610068321,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.440683364868164,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03325957432389259,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.3949294090271,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.035377148538827896,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.400777816772461,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.03725374862551689,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.422494411468506,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03601110726594925,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.4008941650390625,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03193027526140213,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.400361061096191,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.034382354468107224,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.408535003662109,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.03747732564806938,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.412189483642578,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03641800582408905,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.381453037261963,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.0315779373049736,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.4169769287109375,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03583509847521782,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.385746955871582,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02574268914759159,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7456769943237305,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.588273823261261,
+ "eval_runtime": 59.9356,
+ "eval_samples_per_second": 40.744,
+ "eval_steps_per_second": 0.651,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013961605584642,
+ "grad_norm": 0.037516895681619644,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 4.33476448059082,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027923211169284,
+ "grad_norm": 0.048450272530317307,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 4.296454429626465,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041884816753926,
+ "grad_norm": 0.05368480458855629,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 4.314815998077393,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055846422338568,
+ "grad_norm": 0.05520009621977806,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 4.330798625946045,
+ "step": 1084
+ },
+ {
+ "epoch": 15.06980802792321,
+ "grad_norm": 0.05928800627589226,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 4.317663192749023,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083769633507853,
+ "grad_norm": 0.06315027177333832,
+ "learning_rate": 0.000491285979372924,
+ "loss": 4.329221725463867,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097731239092496,
+ "grad_norm": 0.06342213600873947,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 4.32546329498291,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111692844677139,
+ "grad_norm": 0.0667002946138382,
+ "learning_rate": 0.000490873338088198,
+ "loss": 4.309828758239746,
+ "step": 1088
+ },
+ {
+ "epoch": 15.12565445026178,
+ "grad_norm": 0.05264135077595711,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 4.307073593139648,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139616055846423,
+ "grad_norm": 0.051895879209041595,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 4.333526611328125,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153577661431065,
+ "grad_norm": 0.05508396774530411,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 4.330171585083008,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167539267015707,
+ "grad_norm": 0.057272691279649734,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 4.329268455505371,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181500872600349,
+ "grad_norm": 0.058956146240234375,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 4.341983795166016,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195462478184991,
+ "grad_norm": 0.0557059645652771,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 4.348274230957031,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209424083769633,
+ "grad_norm": 0.05661974847316742,
+ "learning_rate": 0.000489424334303338,
+ "loss": 4.347631454467773,
+ "step": 1095
+ },
+ {
+ "epoch": 15.223385689354275,
+ "grad_norm": 0.05338657647371292,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 4.334639072418213,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237347294938917,
+ "grad_norm": 0.05091148987412453,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 4.355465412139893,
+ "step": 1097
+ },
+ {
+ "epoch": 15.25130890052356,
+ "grad_norm": 0.04996965080499649,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 4.346457004547119,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265270506108202,
+ "grad_norm": 0.05573750659823418,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 4.346520900726318,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279232111692846,
+ "grad_norm": 0.05776917189359665,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 4.359007835388184,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293193717277488,
+ "grad_norm": 0.050820041447877884,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 4.331710338592529,
+ "step": 1101
+ },
+ {
+ "epoch": 15.30715532286213,
+ "grad_norm": 0.04441820830106735,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 4.3126420974731445,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321116928446772,
+ "grad_norm": 0.04582618921995163,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 4.312925338745117,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335078534031414,
+ "grad_norm": 0.043592359870672226,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 4.345674514770508,
+ "step": 1104
+ },
+ {
+ "epoch": 15.349040139616056,
+ "grad_norm": 0.04069037735462189,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 4.334713935852051,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363001745200698,
+ "grad_norm": 0.04083041846752167,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 4.358171463012695,
+ "step": 1106
+ },
+ {
+ "epoch": 15.37696335078534,
+ "grad_norm": 0.04119594022631645,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 4.317891597747803,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390924956369982,
+ "grad_norm": 0.042060643434524536,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 4.355810165405273,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404886561954624,
+ "grad_norm": 0.04721749201416969,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 4.353546619415283,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418848167539267,
+ "grad_norm": 0.04657730832695961,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 4.341683387756348,
+ "step": 1110
+ },
+ {
+ "epoch": 15.432809773123909,
+ "grad_norm": 0.04096715524792671,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 4.357192039489746,
+ "step": 1111
+ },
+ {
+ "epoch": 15.44677137870855,
+ "grad_norm": 0.03904155269265175,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 4.354775428771973,
+ "step": 1112
+ },
+ {
+ "epoch": 15.460732984293193,
+ "grad_norm": 0.03828521445393562,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 4.356499671936035,
+ "step": 1113
+ },
+ {
+ "epoch": 15.474694589877837,
+ "grad_norm": 0.03688685595989227,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 4.339485168457031,
+ "step": 1114
+ },
+ {
+ "epoch": 15.488656195462479,
+ "grad_norm": 0.03945618495345116,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 4.363500118255615,
+ "step": 1115
+ },
+ {
+ "epoch": 15.502617801047121,
+ "grad_norm": 0.03821462020277977,
+ "learning_rate": 0.000485033394934934,
+ "loss": 4.353529930114746,
+ "step": 1116
+ },
+ {
+ "epoch": 15.516579406631763,
+ "grad_norm": 0.036265499889850616,
+ "learning_rate": 0.000484822676912974,
+ "loss": 4.337663173675537,
+ "step": 1117
+ },
+ {
+ "epoch": 15.530541012216405,
+ "grad_norm": 0.03803930804133415,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 4.34660530090332,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544502617801047,
+ "grad_norm": 0.038425102829933167,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 4.373246192932129,
+ "step": 1119
+ },
+ {
+ "epoch": 15.55846422338569,
+ "grad_norm": 0.036852262914180756,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 4.331226825714111,
+ "step": 1120
+ },
+ {
+ "epoch": 15.572425828970331,
+ "grad_norm": 0.03700809180736542,
+ "learning_rate": 0.000483978341612154,
+ "loss": 4.375818729400635,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586387434554974,
+ "grad_norm": 0.03628664091229439,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 4.32705020904541,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600349040139616,
+ "grad_norm": 0.036831192672252655,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 4.337127208709717,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614310645724258,
+ "grad_norm": 0.03736541047692299,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 4.36502742767334,
+ "step": 1124
+ },
+ {
+ "epoch": 15.6282722513089,
+ "grad_norm": 0.03658579662442207,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 4.359844207763672,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642233856893542,
+ "grad_norm": 0.03688633441925049,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 4.3355793952941895,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656195462478184,
+ "grad_norm": 0.03477008268237114,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 4.333405494689941,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670157068062828,
+ "grad_norm": 0.03655015677213669,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 4.338570594787598,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68411867364747,
+ "grad_norm": 0.03664765506982803,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 4.352496147155762,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698080279232112,
+ "grad_norm": 0.037019114941358566,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 4.335735321044922,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712041884816754,
+ "grad_norm": 0.03949085995554924,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 4.342585563659668,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726003490401396,
+ "grad_norm": 0.04084350913763046,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 4.329622268676758,
+ "step": 1132
+ },
+ {
+ "epoch": 15.739965095986038,
+ "grad_norm": 0.038439832627773285,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 4.360318660736084,
+ "step": 1133
+ },
+ {
+ "epoch": 15.75392670157068,
+ "grad_norm": 0.03932987526059151,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 4.343446254730225,
+ "step": 1134
+ },
+ {
+ "epoch": 15.767888307155323,
+ "grad_norm": 0.041595373302698135,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 4.351495742797852,
+ "step": 1135
+ },
+ {
+ "epoch": 15.781849912739965,
+ "grad_norm": 0.03917490318417549,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 4.380537033081055,
+ "step": 1136
+ },
+ {
+ "epoch": 15.795811518324607,
+ "grad_norm": 0.038597021251916885,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 4.3527116775512695,
+ "step": 1137
+ },
+ {
+ "epoch": 15.809773123909249,
+ "grad_norm": 0.03651711344718933,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 4.369256496429443,
+ "step": 1138
+ },
+ {
+ "epoch": 15.823734729493891,
+ "grad_norm": 0.035118963569402695,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 4.350616455078125,
+ "step": 1139
+ },
+ {
+ "epoch": 15.837696335078533,
+ "grad_norm": 0.034945692867040634,
+ "learning_rate": 0.000479936075927473,
+ "loss": 4.365446090698242,
+ "step": 1140
+ },
+ {
+ "epoch": 15.851657940663177,
+ "grad_norm": 0.03519820421934128,
+ "learning_rate": 0.000479721889242305,
+ "loss": 4.357390403747559,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86561954624782,
+ "grad_norm": 0.03407098725438118,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 4.350573539733887,
+ "step": 1142
+ },
+ {
+ "epoch": 15.879581151832461,
+ "grad_norm": 0.036511123180389404,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 4.328500270843506,
+ "step": 1143
+ },
+ {
+ "epoch": 15.893542757417103,
+ "grad_norm": 0.039194293320178986,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 4.346471786499023,
+ "step": 1144
+ },
+ {
+ "epoch": 15.907504363001745,
+ "grad_norm": 0.03777911514043808,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 4.358243942260742,
+ "step": 1145
+ },
+ {
+ "epoch": 15.921465968586388,
+ "grad_norm": 0.0354655385017395,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 4.368768692016602,
+ "step": 1146
+ },
+ {
+ "epoch": 15.93542757417103,
+ "grad_norm": 0.03710418567061424,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 4.3482184410095215,
+ "step": 1147
+ },
+ {
+ "epoch": 15.949389179755672,
+ "grad_norm": 0.03545419126749039,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 4.360692501068115,
+ "step": 1148
+ },
+ {
+ "epoch": 15.963350785340314,
+ "grad_norm": 0.035963475704193115,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 4.3592119216918945,
+ "step": 1149
+ },
+ {
+ "epoch": 15.977312390924956,
+ "grad_norm": 0.03425297513604164,
+ "learning_rate": 0.000477787822648007,
+ "loss": 4.329504013061523,
+ "step": 1150
+ },
+ {
+ "epoch": 15.991273996509598,
+ "grad_norm": 0.03453255817294121,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 4.298142433166504,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.025766054168343544,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 2.7026405334472656,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.5888971090316772,
+ "eval_runtime": 58.9721,
+ "eval_samples_per_second": 41.409,
+ "eval_steps_per_second": 0.661,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013961605584644,
+ "grad_norm": 0.04088004305958748,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 4.268601417541504,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027923211169284,
+ "grad_norm": 0.05194641277194023,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 4.267483711242676,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041884816753928,
+ "grad_norm": 0.05488700792193413,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 4.2527570724487305,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05584642233857,
+ "grad_norm": 0.059292279183864594,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 4.283864974975586,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069808027923212,
+ "grad_norm": 0.0567227303981781,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 4.291148662567139,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083769633507853,
+ "grad_norm": 0.05582835525274277,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 4.250524520874023,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097731239092496,
+ "grad_norm": 0.055709999054670334,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 4.259716033935547,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111692844677137,
+ "grad_norm": 0.053324732929468155,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 4.247231483459473,
+ "step": 1160
+ },
+ {
+ "epoch": 16.12565445026178,
+ "grad_norm": 0.05367910861968994,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 4.263905048370361,
+ "step": 1161
+ },
+ {
+ "epoch": 16.13961605584642,
+ "grad_norm": 0.057094886898994446,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 4.266510009765625,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153577661431065,
+ "grad_norm": 0.05921962857246399,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 4.278281211853027,
+ "step": 1163
+ },
+ {
+ "epoch": 16.167539267015705,
+ "grad_norm": 0.060246821492910385,
+ "learning_rate": 0.000474756648870413,
+ "loss": 4.2848052978515625,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18150087260035,
+ "grad_norm": 0.05686050280928612,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 4.286434173583984,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195462478184993,
+ "grad_norm": 0.05207265168428421,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 4.246121406555176,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209424083769633,
+ "grad_norm": 0.0477963425219059,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 4.287411689758301,
+ "step": 1167
+ },
+ {
+ "epoch": 16.223385689354277,
+ "grad_norm": 0.05586806684732437,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 4.2721333503723145,
+ "step": 1168
+ },
+ {
+ "epoch": 16.237347294938917,
+ "grad_norm": 0.05427060276269913,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 4.2751641273498535,
+ "step": 1169
+ },
+ {
+ "epoch": 16.25130890052356,
+ "grad_norm": 0.0518609918653965,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 4.287444114685059,
+ "step": 1170
+ },
+ {
+ "epoch": 16.2652705061082,
+ "grad_norm": 0.053606316447257996,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 4.287125587463379,
+ "step": 1171
+ },
+ {
+ "epoch": 16.279232111692846,
+ "grad_norm": 0.05118813365697861,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 4.291893005371094,
+ "step": 1172
+ },
+ {
+ "epoch": 16.293193717277486,
+ "grad_norm": 0.04774491861462593,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 4.325250148773193,
+ "step": 1173
+ },
+ {
+ "epoch": 16.30715532286213,
+ "grad_norm": 0.04634870961308479,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 4.311764717102051,
+ "step": 1174
+ },
+ {
+ "epoch": 16.32111692844677,
+ "grad_norm": 0.04411927983164787,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 4.277068138122559,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335078534031414,
+ "grad_norm": 0.043819356709718704,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 4.309638977050781,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349040139616054,
+ "grad_norm": 0.0418904684484005,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 4.306234836578369,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363001745200698,
+ "grad_norm": 0.04327298328280449,
+ "learning_rate": 0.000471698352726896,
+ "loss": 4.296874046325684,
+ "step": 1178
+ },
+ {
+ "epoch": 16.376963350785342,
+ "grad_norm": 0.045296184718608856,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 4.288957595825195,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390924956369982,
+ "grad_norm": 0.04447923228144646,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 4.313834190368652,
+ "step": 1180
+ },
+ {
+ "epoch": 16.404886561954626,
+ "grad_norm": 0.04548390209674835,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 4.284562587738037,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418848167539267,
+ "grad_norm": 0.04567817226052284,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 4.295994758605957,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43280977312391,
+ "grad_norm": 0.04277439042925835,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 4.262106895446777,
+ "step": 1183
+ },
+ {
+ "epoch": 16.44677137870855,
+ "grad_norm": 0.043597545474767685,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 4.294536590576172,
+ "step": 1184
+ },
+ {
+ "epoch": 16.460732984293195,
+ "grad_norm": 0.04424033313989639,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 4.297497749328613,
+ "step": 1185
+ },
+ {
+ "epoch": 16.474694589877835,
+ "grad_norm": 0.04354134574532509,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 4.321508884429932,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48865619546248,
+ "grad_norm": 0.04241178184747696,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 4.322726249694824,
+ "step": 1187
+ },
+ {
+ "epoch": 16.50261780104712,
+ "grad_norm": 0.039317622780799866,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 4.311155319213867,
+ "step": 1188
+ },
+ {
+ "epoch": 16.516579406631763,
+ "grad_norm": 0.04126536473631859,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 4.306636810302734,
+ "step": 1189
+ },
+ {
+ "epoch": 16.530541012216403,
+ "grad_norm": 0.040116794407367706,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 4.314671516418457,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544502617801047,
+ "grad_norm": 0.04045999422669411,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 4.301012992858887,
+ "step": 1191
+ },
+ {
+ "epoch": 16.55846422338569,
+ "grad_norm": 0.04013422504067421,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 4.29913330078125,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57242582897033,
+ "grad_norm": 0.03878502920269966,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 4.351600646972656,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586387434554975,
+ "grad_norm": 0.038935888558626175,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 4.27667236328125,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600349040139616,
+ "grad_norm": 0.04055892676115036,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 4.29019021987915,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61431064572426,
+ "grad_norm": 0.03959144279360771,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 4.327725887298584,
+ "step": 1196
+ },
+ {
+ "epoch": 16.6282722513089,
+ "grad_norm": 0.039878539741039276,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 4.297952651977539,
+ "step": 1197
+ },
+ {
+ "epoch": 16.642233856893544,
+ "grad_norm": 0.041209928691387177,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 4.313030242919922,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656195462478184,
+ "grad_norm": 0.04346393793821335,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 4.274901390075684,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670157068062828,
+ "grad_norm": 0.04090269282460213,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 4.330657958984375,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68411867364747,
+ "grad_norm": 0.04004506766796112,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 4.323106288909912,
+ "step": 1201
+ },
+ {
+ "epoch": 16.698080279232112,
+ "grad_norm": 0.041541460901498795,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 4.280947685241699,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712041884816752,
+ "grad_norm": 0.04033653065562248,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 4.341434955596924,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726003490401396,
+ "grad_norm": 0.04104280099272728,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 4.2931671142578125,
+ "step": 1204
+ },
+ {
+ "epoch": 16.739965095986037,
+ "grad_norm": 0.039398156106472015,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 4.316840648651123,
+ "step": 1205
+ },
+ {
+ "epoch": 16.75392670157068,
+ "grad_norm": 0.03980560600757599,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 4.325345039367676,
+ "step": 1206
+ },
+ {
+ "epoch": 16.767888307155324,
+ "grad_norm": 0.040656186640262604,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 4.32675838470459,
+ "step": 1207
+ },
+ {
+ "epoch": 16.781849912739965,
+ "grad_norm": 0.03775986284017563,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 4.314423561096191,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79581151832461,
+ "grad_norm": 0.04043284431099892,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 4.344202041625977,
+ "step": 1209
+ },
+ {
+ "epoch": 16.80977312390925,
+ "grad_norm": 0.04074503108859062,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 4.350123405456543,
+ "step": 1210
+ },
+ {
+ "epoch": 16.823734729493893,
+ "grad_norm": 0.04117420315742493,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 4.350917816162109,
+ "step": 1211
+ },
+ {
+ "epoch": 16.837696335078533,
+ "grad_norm": 0.03799610957503319,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 4.289987564086914,
+ "step": 1212
+ },
+ {
+ "epoch": 16.851657940663177,
+ "grad_norm": 0.04071265831589699,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 4.306744575500488,
+ "step": 1213
+ },
+ {
+ "epoch": 16.865619546247817,
+ "grad_norm": 0.04109715297818184,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 4.318753242492676,
+ "step": 1214
+ },
+ {
+ "epoch": 16.87958115183246,
+ "grad_norm": 0.03910145163536072,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 4.281325340270996,
+ "step": 1215
+ },
+ {
+ "epoch": 16.8935427574171,
+ "grad_norm": 0.03752733767032623,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 4.3231048583984375,
+ "step": 1216
+ },
+ {
+ "epoch": 16.907504363001745,
+ "grad_norm": 0.03752103075385094,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 4.351336479187012,
+ "step": 1217
+ },
+ {
+ "epoch": 16.921465968586386,
+ "grad_norm": 0.03675422817468643,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 4.282256126403809,
+ "step": 1218
+ },
+ {
+ "epoch": 16.93542757417103,
+ "grad_norm": 0.03850258141756058,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 4.322597503662109,
+ "step": 1219
+ },
+ {
+ "epoch": 16.949389179755673,
+ "grad_norm": 0.03692392632365227,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 4.320333957672119,
+ "step": 1220
+ },
+ {
+ "epoch": 16.963350785340314,
+ "grad_norm": 0.035366240888834,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 4.31341552734375,
+ "step": 1221
+ },
+ {
+ "epoch": 16.977312390924958,
+ "grad_norm": 0.037120968103408813,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 4.321141242980957,
+ "step": 1222
+ },
+ {
+ "epoch": 16.991273996509598,
+ "grad_norm": 0.03721754252910614,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 4.305757522583008,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.026698043569922447,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 2.716372489929199,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.5894472599029541,
+ "eval_runtime": 58.927,
+ "eval_samples_per_second": 41.441,
+ "eval_steps_per_second": 0.662,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013961605584644,
+ "grad_norm": 0.044532276690006256,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 4.243244171142578,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027923211169284,
+ "grad_norm": 0.050529688596725464,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 4.221382141113281,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041884816753928,
+ "grad_norm": 0.05690870061516762,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 4.240285873413086,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05584642233857,
+ "grad_norm": 0.054803814738988876,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 4.213047027587891,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069808027923212,
+ "grad_norm": 0.051825862377882004,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 4.226880073547363,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083769633507853,
+ "grad_norm": 0.049802035093307495,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 4.25303840637207,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097731239092496,
+ "grad_norm": 0.05290328338742256,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 4.199422836303711,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111692844677137,
+ "grad_norm": 0.0531814880669117,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 4.201648712158203,
+ "step": 1232
+ },
+ {
+ "epoch": 17.12565445026178,
+ "grad_norm": 0.048972588032484055,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 4.207578182220459,
+ "step": 1233
+ },
+ {
+ "epoch": 17.13961605584642,
+ "grad_norm": 0.051345713436603546,
+ "learning_rate": 0.000459203665939781,
+ "loss": 4.206205368041992,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153577661431065,
+ "grad_norm": 0.059012677520513535,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 4.228360176086426,
+ "step": 1235
+ },
+ {
+ "epoch": 17.167539267015705,
+ "grad_norm": 0.05984492972493172,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 4.234362602233887,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18150087260035,
+ "grad_norm": 0.060886185616254807,
+ "learning_rate": 0.000458522886595554,
+ "loss": 4.253525257110596,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195462478184993,
+ "grad_norm": 0.06476911902427673,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 4.249889373779297,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209424083769633,
+ "grad_norm": 0.059773124754428864,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 4.225249290466309,
+ "step": 1239
+ },
+ {
+ "epoch": 17.223385689354277,
+ "grad_norm": 0.06063077598810196,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 4.25465202331543,
+ "step": 1240
+ },
+ {
+ "epoch": 17.237347294938917,
+ "grad_norm": 0.059922896325588226,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 4.252101898193359,
+ "step": 1241
+ },
+ {
+ "epoch": 17.25130890052356,
+ "grad_norm": 0.05849695950746536,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 4.268218040466309,
+ "step": 1242
+ },
+ {
+ "epoch": 17.2652705061082,
+ "grad_norm": 0.05769657716155052,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 4.2395782470703125,
+ "step": 1243
+ },
+ {
+ "epoch": 17.279232111692846,
+ "grad_norm": 0.058189719915390015,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 4.257606506347656,
+ "step": 1244
+ },
+ {
+ "epoch": 17.293193717277486,
+ "grad_norm": 0.059679049998521805,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 4.278588771820068,
+ "step": 1245
+ },
+ {
+ "epoch": 17.30715532286213,
+ "grad_norm": 0.0579286590218544,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 4.25511360168457,
+ "step": 1246
+ },
+ {
+ "epoch": 17.32111692844677,
+ "grad_norm": 0.05269965901970863,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 4.241326332092285,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335078534031414,
+ "grad_norm": 0.04828358441591263,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 4.243182182312012,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349040139616054,
+ "grad_norm": 0.050172045826911926,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 4.295144081115723,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363001745200698,
+ "grad_norm": 0.047297943383455276,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 4.250682830810547,
+ "step": 1250
+ },
+ {
+ "epoch": 17.376963350785342,
+ "grad_norm": 0.04574039950966835,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 4.269444465637207,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390924956369982,
+ "grad_norm": 0.042310427874326706,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 4.255361557006836,
+ "step": 1252
+ },
+ {
+ "epoch": 17.404886561954626,
+ "grad_norm": 0.04268861189484596,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 4.252292156219482,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418848167539267,
+ "grad_norm": 0.0424286350607872,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 4.24439811706543,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43280977312391,
+ "grad_norm": 0.042049869894981384,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 4.236266136169434,
+ "step": 1255
+ },
+ {
+ "epoch": 17.44677137870855,
+ "grad_norm": 0.04185241088271141,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 4.268826484680176,
+ "step": 1256
+ },
+ {
+ "epoch": 17.460732984293195,
+ "grad_norm": 0.04423609375953674,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 4.240240573883057,
+ "step": 1257
+ },
+ {
+ "epoch": 17.474694589877835,
+ "grad_norm": 0.041972566395998,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 4.269453048706055,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48865619546248,
+ "grad_norm": 0.04493343085050583,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 4.270544052124023,
+ "step": 1259
+ },
+ {
+ "epoch": 17.50261780104712,
+ "grad_norm": 0.04502921551465988,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 4.271880149841309,
+ "step": 1260
+ },
+ {
+ "epoch": 17.516579406631763,
+ "grad_norm": 0.044518254697322845,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 4.249786376953125,
+ "step": 1261
+ },
+ {
+ "epoch": 17.530541012216403,
+ "grad_norm": 0.044482357800006866,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 4.300779819488525,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544502617801047,
+ "grad_norm": 0.0424671471118927,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 4.248597621917725,
+ "step": 1263
+ },
+ {
+ "epoch": 17.55846422338569,
+ "grad_norm": 0.04136480763554573,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 4.276763916015625,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57242582897033,
+ "grad_norm": 0.04094843938946724,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 4.263628005981445,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586387434554975,
+ "grad_norm": 0.04415388032793999,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 4.257946014404297,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600349040139616,
+ "grad_norm": 0.04177384823560715,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 4.289669036865234,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61431064572426,
+ "grad_norm": 0.040863484144210815,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 4.320903778076172,
+ "step": 1268
+ },
+ {
+ "epoch": 17.6282722513089,
+ "grad_norm": 0.04025879129767418,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 4.268717288970947,
+ "step": 1269
+ },
+ {
+ "epoch": 17.642233856893544,
+ "grad_norm": 0.04161325842142105,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 4.275268077850342,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656195462478184,
+ "grad_norm": 0.04260355979204178,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 4.263361930847168,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670157068062828,
+ "grad_norm": 0.043366871774196625,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 4.254581928253174,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68411867364747,
+ "grad_norm": 0.04299113526940346,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 4.268922328948975,
+ "step": 1273
+ },
+ {
+ "epoch": 17.698080279232112,
+ "grad_norm": 0.041865330189466476,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 4.290097236633301,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712041884816752,
+ "grad_norm": 0.04515853524208069,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 4.261376857757568,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726003490401396,
+ "grad_norm": 0.046030085533857346,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 4.290409564971924,
+ "step": 1276
+ },
+ {
+ "epoch": 17.739965095986037,
+ "grad_norm": 0.045034267008304596,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 4.28201961517334,
+ "step": 1277
+ },
+ {
+ "epoch": 17.75392670157068,
+ "grad_norm": 0.04293077066540718,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 4.282614707946777,
+ "step": 1278
+ },
+ {
+ "epoch": 17.767888307155324,
+ "grad_norm": 0.043866172432899475,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 4.2777557373046875,
+ "step": 1279
+ },
+ {
+ "epoch": 17.781849912739965,
+ "grad_norm": 0.04693951830267906,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 4.272597312927246,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79581151832461,
+ "grad_norm": 0.04072774201631546,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 4.284282684326172,
+ "step": 1281
+ },
+ {
+ "epoch": 17.80977312390925,
+ "grad_norm": 0.04241872951388359,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 4.254406452178955,
+ "step": 1282
+ },
+ {
+ "epoch": 17.823734729493893,
+ "grad_norm": 0.04614942893385887,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 4.297187805175781,
+ "step": 1283
+ },
+ {
+ "epoch": 17.837696335078533,
+ "grad_norm": 0.04326155036687851,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 4.292263984680176,
+ "step": 1284
+ },
+ {
+ "epoch": 17.851657940663177,
+ "grad_norm": 0.04053817689418793,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 4.2925920486450195,
+ "step": 1285
+ },
+ {
+ "epoch": 17.865619546247817,
+ "grad_norm": 0.040867388248443604,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 4.276965141296387,
+ "step": 1286
+ },
+ {
+ "epoch": 17.87958115183246,
+ "grad_norm": 0.041000183671712875,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 4.257638454437256,
+ "step": 1287
+ },
+ {
+ "epoch": 17.8935427574171,
+ "grad_norm": 0.04014866054058075,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 4.2734479904174805,
+ "step": 1288
+ },
+ {
+ "epoch": 17.907504363001745,
+ "grad_norm": 0.041237372905015945,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 4.263461112976074,
+ "step": 1289
+ },
+ {
+ "epoch": 17.921465968586386,
+ "grad_norm": 0.039109110832214355,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 4.295286655426025,
+ "step": 1290
+ },
+ {
+ "epoch": 17.93542757417103,
+ "grad_norm": 0.03911379352211952,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 4.254448413848877,
+ "step": 1291
+ },
+ {
+ "epoch": 17.949389179755673,
+ "grad_norm": 0.03997721150517464,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 4.284714698791504,
+ "step": 1292
+ },
+ {
+ "epoch": 17.963350785340314,
+ "grad_norm": 0.04011402651667595,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 4.278304100036621,
+ "step": 1293
+ },
+ {
+ "epoch": 17.977312390924958,
+ "grad_norm": 0.04040595516562462,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 4.288183212280273,
+ "step": 1294
+ },
+ {
+ "epoch": 17.991273996509598,
+ "grad_norm": 0.038857635110616684,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 4.274664402008057,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.028110472485423088,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 2.6625723838806152,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.5912551283836365,
+ "eval_runtime": 60.273,
+ "eval_samples_per_second": 40.516,
+ "eval_steps_per_second": 0.647,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013961605584644,
+ "grad_norm": 0.041634928435087204,
+ "learning_rate": 0.000444677103403854,
+ "loss": 4.181324005126953,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027923211169284,
+ "grad_norm": 0.051596369594335556,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 4.204722881317139,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041884816753928,
+ "grad_norm": 0.058580417186021805,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 4.207996845245361,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05584642233857,
+ "grad_norm": 0.05986522138118744,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 4.184011459350586,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069808027923212,
+ "grad_norm": 0.051045969128608704,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 4.148118019104004,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083769633507853,
+ "grad_norm": 0.045477766543626785,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 4.207469940185547,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097731239092496,
+ "grad_norm": 0.0488731674849987,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 4.201406478881836,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111692844677137,
+ "grad_norm": 0.05013517662882805,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 4.192239761352539,
+ "step": 1304
+ },
+ {
+ "epoch": 18.12565445026178,
+ "grad_norm": 0.05091651529073715,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 4.188252925872803,
+ "step": 1305
+ },
+ {
+ "epoch": 18.13961605584642,
+ "grad_norm": 0.051825884729623795,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 4.157929420471191,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153577661431065,
+ "grad_norm": 0.049625713378190994,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 4.208776473999023,
+ "step": 1307
+ },
+ {
+ "epoch": 18.167539267015705,
+ "grad_norm": 0.04816461354494095,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 4.180496692657471,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18150087260035,
+ "grad_norm": 0.04720531404018402,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 4.178536415100098,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195462478184993,
+ "grad_norm": 0.0475568026304245,
+ "learning_rate": 0.000441621628895411,
+ "loss": 4.203497886657715,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209424083769633,
+ "grad_norm": 0.04613947868347168,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 4.226914405822754,
+ "step": 1311
+ },
+ {
+ "epoch": 18.223385689354277,
+ "grad_norm": 0.04585929960012436,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 4.175119400024414,
+ "step": 1312
+ },
+ {
+ "epoch": 18.237347294938917,
+ "grad_norm": 0.045777060091495514,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 4.181760311126709,
+ "step": 1313
+ },
+ {
+ "epoch": 18.25130890052356,
+ "grad_norm": 0.04583602771162987,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 4.179868698120117,
+ "step": 1314
+ },
+ {
+ "epoch": 18.2652705061082,
+ "grad_norm": 0.04706503078341484,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 4.203325271606445,
+ "step": 1315
+ },
+ {
+ "epoch": 18.279232111692846,
+ "grad_norm": 0.045122336596250534,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 4.187152862548828,
+ "step": 1316
+ },
+ {
+ "epoch": 18.293193717277486,
+ "grad_norm": 0.04495825991034508,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 4.2080535888671875,
+ "step": 1317
+ },
+ {
+ "epoch": 18.30715532286213,
+ "grad_norm": 0.04767543077468872,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 4.207225322723389,
+ "step": 1318
+ },
+ {
+ "epoch": 18.32111692844677,
+ "grad_norm": 0.053396135568618774,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 4.227478981018066,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335078534031414,
+ "grad_norm": 0.0512089841067791,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 4.1904754638671875,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349040139616054,
+ "grad_norm": 0.04967977851629257,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 4.2079925537109375,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363001745200698,
+ "grad_norm": 0.04800970479846001,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 4.206037998199463,
+ "step": 1322
+ },
+ {
+ "epoch": 18.376963350785342,
+ "grad_norm": 0.04993556812405586,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 4.210054397583008,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390924956369982,
+ "grad_norm": 0.0480492040514946,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 4.229897499084473,
+ "step": 1324
+ },
+ {
+ "epoch": 18.404886561954626,
+ "grad_norm": 0.048699282109737396,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 4.17500114440918,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418848167539267,
+ "grad_norm": 0.04681490734219551,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 4.196526527404785,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43280977312391,
+ "grad_norm": 0.04674449935555458,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 4.246790409088135,
+ "step": 1327
+ },
+ {
+ "epoch": 18.44677137870855,
+ "grad_norm": 0.04650607705116272,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 4.225878715515137,
+ "step": 1328
+ },
+ {
+ "epoch": 18.460732984293195,
+ "grad_norm": 0.047200269997119904,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 4.20685338973999,
+ "step": 1329
+ },
+ {
+ "epoch": 18.474694589877835,
+ "grad_norm": 0.04322655498981476,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 4.212920665740967,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48865619546248,
+ "grad_norm": 0.04489128664135933,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 4.266653060913086,
+ "step": 1331
+ },
+ {
+ "epoch": 18.50261780104712,
+ "grad_norm": 0.048838935792446136,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 4.2364501953125,
+ "step": 1332
+ },
+ {
+ "epoch": 18.516579406631763,
+ "grad_norm": 0.04893286153674126,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 4.202394485473633,
+ "step": 1333
+ },
+ {
+ "epoch": 18.530541012216403,
+ "grad_norm": 0.046900101006031036,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 4.209832191467285,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544502617801047,
+ "grad_norm": 0.044056762009859085,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 4.229903221130371,
+ "step": 1335
+ },
+ {
+ "epoch": 18.55846422338569,
+ "grad_norm": 0.04681328684091568,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 4.22783899307251,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57242582897033,
+ "grad_norm": 0.046087440103292465,
+ "learning_rate": 0.000435215814386134,
+ "loss": 4.2078986167907715,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586387434554975,
+ "grad_norm": 0.04363090172410011,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 4.218686103820801,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600349040139616,
+ "grad_norm": 0.046255532652139664,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 4.2475128173828125,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61431064572426,
+ "grad_norm": 0.045545510947704315,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 4.233272552490234,
+ "step": 1340
+ },
+ {
+ "epoch": 18.6282722513089,
+ "grad_norm": 0.045259326696395874,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 4.252043724060059,
+ "step": 1341
+ },
+ {
+ "epoch": 18.642233856893544,
+ "grad_norm": 0.042144861072301865,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 4.247762203216553,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656195462478184,
+ "grad_norm": 0.04298558086156845,
+ "learning_rate": 0.000433781621332752,
+ "loss": 4.222329616546631,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670157068062828,
+ "grad_norm": 0.044070884585380554,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 4.222488880157471,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68411867364747,
+ "grad_norm": 0.044758375734090805,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 4.238160133361816,
+ "step": 1345
+ },
+ {
+ "epoch": 18.698080279232112,
+ "grad_norm": 0.04324857518076897,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 4.220003128051758,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712041884816752,
+ "grad_norm": 0.04349793121218681,
+ "learning_rate": 0.000432823382695327,
+ "loss": 4.251608848571777,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726003490401396,
+ "grad_norm": 0.04483242705464363,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 4.243319511413574,
+ "step": 1348
+ },
+ {
+ "epoch": 18.739965095986037,
+ "grad_norm": 0.04391994699835777,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 4.237479209899902,
+ "step": 1349
+ },
+ {
+ "epoch": 18.75392670157068,
+ "grad_norm": 0.042584922164678574,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 4.2309980392456055,
+ "step": 1350
+ },
+ {
+ "epoch": 18.767888307155324,
+ "grad_norm": 0.04363991320133209,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 4.216980934143066,
+ "step": 1351
+ },
+ {
+ "epoch": 18.781849912739965,
+ "grad_norm": 0.04662707448005676,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 4.257450103759766,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79581151832461,
+ "grad_norm": 0.0462074875831604,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 4.222329139709473,
+ "step": 1353
+ },
+ {
+ "epoch": 18.80977312390925,
+ "grad_norm": 0.043543554842472076,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 4.231960773468018,
+ "step": 1354
+ },
+ {
+ "epoch": 18.823734729493893,
+ "grad_norm": 0.04444567486643791,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 4.225714683532715,
+ "step": 1355
+ },
+ {
+ "epoch": 18.837696335078533,
+ "grad_norm": 0.04507977142930031,
+ "learning_rate": 0.000430661245733797,
+ "loss": 4.234722137451172,
+ "step": 1356
+ },
+ {
+ "epoch": 18.851657940663177,
+ "grad_norm": 0.043786536902189255,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 4.274998664855957,
+ "step": 1357
+ },
+ {
+ "epoch": 18.865619546247817,
+ "grad_norm": 0.04188325256109238,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 4.260570526123047,
+ "step": 1358
+ },
+ {
+ "epoch": 18.87958115183246,
+ "grad_norm": 0.0439927875995636,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 4.250492572784424,
+ "step": 1359
+ },
+ {
+ "epoch": 18.8935427574171,
+ "grad_norm": 0.04520147293806076,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 4.250911712646484,
+ "step": 1360
+ },
+ {
+ "epoch": 18.907504363001745,
+ "grad_norm": 0.04442249983549118,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 4.255832672119141,
+ "step": 1361
+ },
+ {
+ "epoch": 18.921465968586386,
+ "grad_norm": 0.044756144285202026,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 4.226641654968262,
+ "step": 1362
+ },
+ {
+ "epoch": 18.93542757417103,
+ "grad_norm": 0.04647127911448479,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 4.2110915184021,
+ "step": 1363
+ },
+ {
+ "epoch": 18.949389179755673,
+ "grad_norm": 0.045986004173755646,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 4.243263244628906,
+ "step": 1364
+ },
+ {
+ "epoch": 18.963350785340314,
+ "grad_norm": 0.04533308371901512,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 4.253750801086426,
+ "step": 1365
+ },
+ {
+ "epoch": 18.977312390924958,
+ "grad_norm": 0.04307471215724945,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 4.233024597167969,
+ "step": 1366
+ },
+ {
+ "epoch": 18.991273996509598,
+ "grad_norm": 0.0416078194975853,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 4.264965534210205,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.030926328152418137,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 2.651331901550293,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.592528223991394,
+ "eval_runtime": 59.67,
+ "eval_samples_per_second": 40.925,
+ "eval_steps_per_second": 0.654,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013961605584644,
+ "grad_norm": 0.04524356126785278,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 4.153796672821045,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027923211169284,
+ "grad_norm": 0.05105067789554596,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 4.154603004455566,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041884816753928,
+ "grad_norm": 0.052823506295681,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 4.1599273681640625,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05584642233857,
+ "grad_norm": 0.05688423290848732,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 4.124622344970703,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069808027923212,
+ "grad_norm": 0.053178343921899796,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 4.138644218444824,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083769633507853,
+ "grad_norm": 0.05473732203245163,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 4.138207912445068,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097731239092496,
+ "grad_norm": 0.055207233875989914,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 4.155735015869141,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111692844677137,
+ "grad_norm": 0.05484927073121071,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 4.140205383300781,
+ "step": 1376
+ },
+ {
+ "epoch": 19.12565445026178,
+ "grad_norm": 0.05589596927165985,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 4.160840034484863,
+ "step": 1377
+ },
+ {
+ "epoch": 19.13961605584642,
+ "grad_norm": 0.05222902446985245,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 4.154781341552734,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153577661431065,
+ "grad_norm": 0.051824700087308884,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 4.134243488311768,
+ "step": 1379
+ },
+ {
+ "epoch": 19.167539267015705,
+ "grad_norm": 0.0517590269446373,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 4.125945568084717,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18150087260035,
+ "grad_norm": 0.05399726703763008,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 4.146993160247803,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195462478184993,
+ "grad_norm": 0.054430052638053894,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 4.1310906410217285,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209424083769633,
+ "grad_norm": 0.0551968477666378,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 4.170640468597412,
+ "step": 1383
+ },
+ {
+ "epoch": 19.223385689354277,
+ "grad_norm": 0.06050962209701538,
+ "learning_rate": 0.000423881957237287,
+ "loss": 4.166876316070557,
+ "step": 1384
+ },
+ {
+ "epoch": 19.237347294938917,
+ "grad_norm": 0.055587392300367355,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 4.150339603424072,
+ "step": 1385
+ },
+ {
+ "epoch": 19.25130890052356,
+ "grad_norm": 0.05193328112363815,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 4.171754360198975,
+ "step": 1386
+ },
+ {
+ "epoch": 19.2652705061082,
+ "grad_norm": 0.05272402986884117,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 4.165149688720703,
+ "step": 1387
+ },
+ {
+ "epoch": 19.279232111692846,
+ "grad_norm": 0.056283943355083466,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 4.193273544311523,
+ "step": 1388
+ },
+ {
+ "epoch": 19.293193717277486,
+ "grad_norm": 0.05825977027416229,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 4.192652225494385,
+ "step": 1389
+ },
+ {
+ "epoch": 19.30715532286213,
+ "grad_norm": 0.05539379641413689,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 4.187044143676758,
+ "step": 1390
+ },
+ {
+ "epoch": 19.32111692844677,
+ "grad_norm": 0.05300498753786087,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 4.168220043182373,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335078534031414,
+ "grad_norm": 0.05425756797194481,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 4.185193061828613,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349040139616054,
+ "grad_norm": 0.05527182295918465,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 4.1618332862854,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363001745200698,
+ "grad_norm": 0.054422613233327866,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 4.195295810699463,
+ "step": 1394
+ },
+ {
+ "epoch": 19.376963350785342,
+ "grad_norm": 0.052717432379722595,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 4.143290996551514,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390924956369982,
+ "grad_norm": 0.05154350399971008,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 4.115989685058594,
+ "step": 1396
+ },
+ {
+ "epoch": 19.404886561954626,
+ "grad_norm": 0.0489451140165329,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 4.169506072998047,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418848167539267,
+ "grad_norm": 0.049093786627054214,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 4.164122104644775,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43280977312391,
+ "grad_norm": 0.04690805822610855,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 4.186679840087891,
+ "step": 1399
+ },
+ {
+ "epoch": 19.44677137870855,
+ "grad_norm": 0.046708665788173676,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 4.184325695037842,
+ "step": 1400
+ },
+ {
+ "epoch": 19.460732984293195,
+ "grad_norm": 0.04617927595973015,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 4.201678276062012,
+ "step": 1401
+ },
+ {
+ "epoch": 19.474694589877835,
+ "grad_norm": 0.05017530545592308,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 4.18928337097168,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48865619546248,
+ "grad_norm": 0.05197977274656296,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 4.188801288604736,
+ "step": 1403
+ },
+ {
+ "epoch": 19.50261780104712,
+ "grad_norm": 0.049691297113895416,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 4.209467887878418,
+ "step": 1404
+ },
+ {
+ "epoch": 19.516579406631763,
+ "grad_norm": 0.049709953367710114,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 4.19733190536499,
+ "step": 1405
+ },
+ {
+ "epoch": 19.530541012216403,
+ "grad_norm": 0.05274137482047081,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 4.221798896789551,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544502617801047,
+ "grad_norm": 0.05015392601490021,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 4.210565567016602,
+ "step": 1407
+ },
+ {
+ "epoch": 19.55846422338569,
+ "grad_norm": 0.05023058131337166,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 4.184835910797119,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57242582897033,
+ "grad_norm": 0.0479239858686924,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 4.199846267700195,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586387434554975,
+ "grad_norm": 0.0481228232383728,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 4.185157775878906,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600349040139616,
+ "grad_norm": 0.04909738525748253,
+ "learning_rate": 0.000417272427439646,
+ "loss": 4.207550525665283,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61431064572426,
+ "grad_norm": 0.0475567989051342,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 4.213225841522217,
+ "step": 1412
+ },
+ {
+ "epoch": 19.6282722513089,
+ "grad_norm": 0.048073310405015945,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 4.212669372558594,
+ "step": 1413
+ },
+ {
+ "epoch": 19.642233856893544,
+ "grad_norm": 0.04523796588182449,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 4.1950860023498535,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656195462478184,
+ "grad_norm": 0.04359745234251022,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 4.2136640548706055,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670157068062828,
+ "grad_norm": 0.047282423824071884,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 4.209680557250977,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68411867364747,
+ "grad_norm": 0.0474061481654644,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 4.188494682312012,
+ "step": 1417
+ },
+ {
+ "epoch": 19.698080279232112,
+ "grad_norm": 0.04402102902531624,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 4.207236289978027,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712041884816752,
+ "grad_norm": 0.04398655146360397,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 4.210899353027344,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726003490401396,
+ "grad_norm": 0.045767925679683685,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 4.171738624572754,
+ "step": 1420
+ },
+ {
+ "epoch": 19.739965095986037,
+ "grad_norm": 0.04498986527323723,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 4.222414970397949,
+ "step": 1421
+ },
+ {
+ "epoch": 19.75392670157068,
+ "grad_norm": 0.04282551258802414,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 4.202138900756836,
+ "step": 1422
+ },
+ {
+ "epoch": 19.767888307155324,
+ "grad_norm": 0.04392262175679207,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 4.184439659118652,
+ "step": 1423
+ },
+ {
+ "epoch": 19.781849912739965,
+ "grad_norm": 0.04343154653906822,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 4.204674243927002,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79581151832461,
+ "grad_norm": 0.04496419429779053,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 4.23316764831543,
+ "step": 1425
+ },
+ {
+ "epoch": 19.80977312390925,
+ "grad_norm": 0.04488370940089226,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 4.201659202575684,
+ "step": 1426
+ },
+ {
+ "epoch": 19.823734729493893,
+ "grad_norm": 0.045598775148391724,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 4.187412261962891,
+ "step": 1427
+ },
+ {
+ "epoch": 19.837696335078533,
+ "grad_norm": 0.043846048414707184,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 4.215444564819336,
+ "step": 1428
+ },
+ {
+ "epoch": 19.851657940663177,
+ "grad_norm": 0.043062254786491394,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 4.2074055671691895,
+ "step": 1429
+ },
+ {
+ "epoch": 19.865619546247817,
+ "grad_norm": 0.04544302448630333,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 4.184919357299805,
+ "step": 1430
+ },
+ {
+ "epoch": 19.87958115183246,
+ "grad_norm": 0.046144865453243256,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 4.2175750732421875,
+ "step": 1431
+ },
+ {
+ "epoch": 19.8935427574171,
+ "grad_norm": 0.04465445503592491,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 4.1851115226745605,
+ "step": 1432
+ },
+ {
+ "epoch": 19.907504363001745,
+ "grad_norm": 0.045913565903902054,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 4.2226152420043945,
+ "step": 1433
+ },
+ {
+ "epoch": 19.921465968586386,
+ "grad_norm": 0.04401300847530365,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 4.216772556304932,
+ "step": 1434
+ },
+ {
+ "epoch": 19.93542757417103,
+ "grad_norm": 0.042624711990356445,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 4.2133684158325195,
+ "step": 1435
+ },
+ {
+ "epoch": 19.949389179755673,
+ "grad_norm": 0.04533256217837334,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 4.218449592590332,
+ "step": 1436
+ },
+ {
+ "epoch": 19.963350785340314,
+ "grad_norm": 0.04764309898018837,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 4.191402912139893,
+ "step": 1437
+ },
+ {
+ "epoch": 19.977312390924958,
+ "grad_norm": 0.04689318686723709,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 4.2113037109375,
+ "step": 1438
+ },
+ {
+ "epoch": 19.991273996509598,
+ "grad_norm": 0.04479913413524628,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 4.194989204406738,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.03151386231184006,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 2.6285438537597656,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 0.5938133597373962,
+ "eval_runtime": 58.9656,
+ "eval_samples_per_second": 41.414,
+ "eval_steps_per_second": 0.661,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013961605584644,
+ "grad_norm": 0.04884875938296318,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 4.1257734298706055,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027923211169284,
+ "grad_norm": 0.05968257039785385,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 4.091641426086426,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041884816753928,
+ "grad_norm": 0.06437838077545166,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 4.122465133666992,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05584642233857,
+ "grad_norm": 0.057831212878227234,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 4.129654884338379,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069808027923212,
+ "grad_norm": 0.049407217651605606,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 4.116935729980469,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083769633507853,
+ "grad_norm": 0.051957421004772186,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 4.125039577484131,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097731239092496,
+ "grad_norm": 0.05425718426704407,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 4.113485336303711,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111692844677137,
+ "grad_norm": 0.055318959057331085,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 4.119422912597656,
+ "step": 1448
+ },
+ {
+ "epoch": 20.12565445026178,
+ "grad_norm": 0.0548059344291687,
+ "learning_rate": 0.000407857329622967,
+ "loss": 4.097405910491943,
+ "step": 1449
+ },
+ {
+ "epoch": 20.13961605584642,
+ "grad_norm": 0.051046356558799744,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 4.115051746368408,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153577661431065,
+ "grad_norm": 0.05079492926597595,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 4.126070499420166,
+ "step": 1451
+ },
+ {
+ "epoch": 20.167539267015705,
+ "grad_norm": 0.04777265712618828,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 4.118952751159668,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18150087260035,
+ "grad_norm": 0.04895703122019768,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 4.095477104187012,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195462478184993,
+ "grad_norm": 0.051196496933698654,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 4.116866111755371,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209424083769633,
+ "grad_norm": 0.05625293031334877,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 4.141980171203613,
+ "step": 1455
+ },
+ {
+ "epoch": 20.223385689354277,
+ "grad_norm": 0.05852660909295082,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 4.126821517944336,
+ "step": 1456
+ },
+ {
+ "epoch": 20.237347294938917,
+ "grad_norm": 0.05419151484966278,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 4.10935640335083,
+ "step": 1457
+ },
+ {
+ "epoch": 20.25130890052356,
+ "grad_norm": 0.05251913517713547,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 4.151289463043213,
+ "step": 1458
+ },
+ {
+ "epoch": 20.2652705061082,
+ "grad_norm": 0.05223092809319496,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 4.149935722351074,
+ "step": 1459
+ },
+ {
+ "epoch": 20.279232111692846,
+ "grad_norm": 0.05314765125513077,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 4.106821060180664,
+ "step": 1460
+ },
+ {
+ "epoch": 20.293193717277486,
+ "grad_norm": 0.05472502112388611,
+ "learning_rate": 0.000404858275823277,
+ "loss": 4.120498180389404,
+ "step": 1461
+ },
+ {
+ "epoch": 20.30715532286213,
+ "grad_norm": 0.05233139917254448,
+ "learning_rate": 0.000404607816444578,
+ "loss": 4.1237335205078125,
+ "step": 1462
+ },
+ {
+ "epoch": 20.32111692844677,
+ "grad_norm": 0.05210232734680176,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 4.1271209716796875,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335078534031414,
+ "grad_norm": 0.049779925495386124,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 4.121427536010742,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349040139616054,
+ "grad_norm": 0.05232279747724533,
+ "learning_rate": 0.000403855947934478,
+ "loss": 4.136429786682129,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363001745200698,
+ "grad_norm": 0.05235981196165085,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 4.1808600425720215,
+ "step": 1466
+ },
+ {
+ "epoch": 20.376963350785342,
+ "grad_norm": 0.04902610555291176,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 4.138566017150879,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390924956369982,
+ "grad_norm": 0.04923798143863678,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 4.135126113891602,
+ "step": 1468
+ },
+ {
+ "epoch": 20.404886561954626,
+ "grad_norm": 0.05059202015399933,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 4.1548285484313965,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418848167539267,
+ "grad_norm": 0.05058307573199272,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 4.149041652679443,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43280977312391,
+ "grad_norm": 0.050172705203294754,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 4.148054122924805,
+ "step": 1471
+ },
+ {
+ "epoch": 20.44677137870855,
+ "grad_norm": 0.05158299580216408,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 4.1425018310546875,
+ "step": 1472
+ },
+ {
+ "epoch": 20.460732984293195,
+ "grad_norm": 0.052176836878061295,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 4.129712104797363,
+ "step": 1473
+ },
+ {
+ "epoch": 20.474694589877835,
+ "grad_norm": 0.052095964550971985,
+ "learning_rate": 0.000401595976318565,
+ "loss": 4.134920597076416,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48865619546248,
+ "grad_norm": 0.04944197088479996,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 4.165334701538086,
+ "step": 1475
+ },
+ {
+ "epoch": 20.50261780104712,
+ "grad_norm": 0.049351636320352554,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 4.144362926483154,
+ "step": 1476
+ },
+ {
+ "epoch": 20.516579406631763,
+ "grad_norm": 0.053657080978155136,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 4.164234161376953,
+ "step": 1477
+ },
+ {
+ "epoch": 20.530541012216403,
+ "grad_norm": 0.052578140050172806,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 4.131514549255371,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544502617801047,
+ "grad_norm": 0.049683794379234314,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 4.129961967468262,
+ "step": 1479
+ },
+ {
+ "epoch": 20.55846422338569,
+ "grad_norm": 0.05115316063165665,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 4.164663314819336,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57242582897033,
+ "grad_norm": 0.051620446145534515,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 4.161067485809326,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586387434554975,
+ "grad_norm": 0.04873501881957054,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 4.161318302154541,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600349040139616,
+ "grad_norm": 0.04968130216002464,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 4.151819229125977,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61431064572426,
+ "grad_norm": 0.05476681888103485,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 4.14247989654541,
+ "step": 1484
+ },
+ {
+ "epoch": 20.6282722513089,
+ "grad_norm": 0.05098642408847809,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 4.178083419799805,
+ "step": 1485
+ },
+ {
+ "epoch": 20.642233856893544,
+ "grad_norm": 0.047264356166124344,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 4.15170955657959,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656195462478184,
+ "grad_norm": 0.050382886081933975,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 4.166364669799805,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670157068062828,
+ "grad_norm": 0.04956239461898804,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 4.164124488830566,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68411867364747,
+ "grad_norm": 0.049608584493398666,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 4.130239486694336,
+ "step": 1489
+ },
+ {
+ "epoch": 20.698080279232112,
+ "grad_norm": 0.0489494763314724,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 4.163503646850586,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712041884816752,
+ "grad_norm": 0.04847714304924011,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 4.1744585037231445,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726003490401396,
+ "grad_norm": 0.050256017595529556,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 4.1502227783203125,
+ "step": 1492
+ },
+ {
+ "epoch": 20.739965095986037,
+ "grad_norm": 0.048107463866472244,
+ "learning_rate": 0.000396803974909638,
+ "loss": 4.1790971755981445,
+ "step": 1493
+ },
+ {
+ "epoch": 20.75392670157068,
+ "grad_norm": 0.04828944429755211,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 4.128883361816406,
+ "step": 1494
+ },
+ {
+ "epoch": 20.767888307155324,
+ "grad_norm": 0.04720750078558922,
+ "learning_rate": 0.000396297942393265,
+ "loss": 4.1661577224731445,
+ "step": 1495
+ },
+ {
+ "epoch": 20.781849912739965,
+ "grad_norm": 0.047148339450359344,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 4.172522068023682,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79581151832461,
+ "grad_norm": 0.04612366110086441,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 4.149024963378906,
+ "step": 1497
+ },
+ {
+ "epoch": 20.80977312390925,
+ "grad_norm": 0.048896901309490204,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 4.179235458374023,
+ "step": 1498
+ },
+ {
+ "epoch": 20.823734729493893,
+ "grad_norm": 0.04779870808124542,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 4.1649932861328125,
+ "step": 1499
+ },
+ {
+ "epoch": 20.837696335078533,
+ "grad_norm": 0.04949707165360451,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 4.17645263671875,
+ "step": 1500
+ },
+ {
+ "epoch": 20.851657940663177,
+ "grad_norm": 0.04645023122429848,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 4.216496467590332,
+ "step": 1501
+ },
+ {
+ "epoch": 20.865619546247817,
+ "grad_norm": 0.046520136296749115,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 4.152667999267578,
+ "step": 1502
+ },
+ {
+ "epoch": 20.87958115183246,
+ "grad_norm": 0.04882653057575226,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 4.187082767486572,
+ "step": 1503
+ },
+ {
+ "epoch": 20.8935427574171,
+ "grad_norm": 0.04669157415628433,
+ "learning_rate": 0.000394017098798633,
+ "loss": 4.172348499298096,
+ "step": 1504
+ },
+ {
+ "epoch": 20.907504363001745,
+ "grad_norm": 0.04595336690545082,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 4.180559158325195,
+ "step": 1505
+ },
+ {
+ "epoch": 20.921465968586386,
+ "grad_norm": 0.048478737473487854,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 4.142880439758301,
+ "step": 1506
+ },
+ {
+ "epoch": 20.93542757417103,
+ "grad_norm": 0.04609905555844307,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 4.1459641456604,
+ "step": 1507
+ },
+ {
+ "epoch": 20.949389179755673,
+ "grad_norm": 0.04524408280849457,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 4.185099124908447,
+ "step": 1508
+ },
+ {
+ "epoch": 20.963350785340314,
+ "grad_norm": 0.0475911870598793,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 4.2085161209106445,
+ "step": 1509
+ },
+ {
+ "epoch": 20.977312390924958,
+ "grad_norm": 0.047796521335840225,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 4.165589332580566,
+ "step": 1510
+ },
+ {
+ "epoch": 20.991273996509598,
+ "grad_norm": 0.047358278185129166,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 4.1634368896484375,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.03328625112771988,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 2.6023788452148438,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.5952173471450806,
+ "eval_runtime": 58.6816,
+ "eval_samples_per_second": 41.614,
+ "eval_steps_per_second": 0.665,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013961605584644,
+ "grad_norm": 0.04970651492476463,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 4.077130317687988,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027923211169284,
+ "grad_norm": 0.060882844030857086,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 4.074376106262207,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041884816753928,
+ "grad_norm": 0.06609825044870377,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 4.107635021209717,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05584642233857,
+ "grad_norm": 0.056066352874040604,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 4.068084716796875,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069808027923212,
+ "grad_norm": 0.053013063967227936,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 4.069967269897461,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083769633507853,
+ "grad_norm": 0.059536613523960114,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 4.0639142990112305,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097731239092496,
+ "grad_norm": 0.06084490567445755,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 4.093890190124512,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111692844677137,
+ "grad_norm": 0.05299340933561325,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 4.100567817687988,
+ "step": 1520
+ },
+ {
+ "epoch": 21.12565445026178,
+ "grad_norm": 0.05136161297559738,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 4.08171272277832,
+ "step": 1521
+ },
+ {
+ "epoch": 21.13961605584642,
+ "grad_norm": 0.05201329290866852,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 4.0869975090026855,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153577661431065,
+ "grad_norm": 0.053958263248205185,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 4.0928850173950195,
+ "step": 1523
+ },
+ {
+ "epoch": 21.167539267015705,
+ "grad_norm": 0.053465038537979126,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 4.097828388214111,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18150087260035,
+ "grad_norm": 0.050980277359485626,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 4.101456165313721,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195462478184993,
+ "grad_norm": 0.05415358766913414,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 4.09096097946167,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209424083769633,
+ "grad_norm": 0.05567048117518425,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 4.061287879943848,
+ "step": 1527
+ },
+ {
+ "epoch": 21.223385689354277,
+ "grad_norm": 0.052690375596284866,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 4.080180644989014,
+ "step": 1528
+ },
+ {
+ "epoch": 21.237347294938917,
+ "grad_norm": 0.05448971316218376,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 4.079794883728027,
+ "step": 1529
+ },
+ {
+ "epoch": 21.25130890052356,
+ "grad_norm": 0.055472876876592636,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 4.072025299072266,
+ "step": 1530
+ },
+ {
+ "epoch": 21.2652705061082,
+ "grad_norm": 0.05418326333165169,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 4.103858470916748,
+ "step": 1531
+ },
+ {
+ "epoch": 21.279232111692846,
+ "grad_norm": 0.0524221733212471,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 4.0808916091918945,
+ "step": 1532
+ },
+ {
+ "epoch": 21.293193717277486,
+ "grad_norm": 0.05224563553929329,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 4.090270042419434,
+ "step": 1533
+ },
+ {
+ "epoch": 21.30715532286213,
+ "grad_norm": 0.052917953580617905,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 4.081887245178223,
+ "step": 1534
+ },
+ {
+ "epoch": 21.32111692844677,
+ "grad_norm": 0.05619358271360397,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 4.089571475982666,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335078534031414,
+ "grad_norm": 0.05455326288938522,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 4.096992015838623,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349040139616054,
+ "grad_norm": 0.049311455339193344,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 4.114621639251709,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363001745200698,
+ "grad_norm": 0.05175276845693588,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 4.107684135437012,
+ "step": 1538
+ },
+ {
+ "epoch": 21.376963350785342,
+ "grad_norm": 0.05236738547682762,
+ "learning_rate": 0.00038509205478744,
+ "loss": 4.075709342956543,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390924956369982,
+ "grad_norm": 0.04899081960320473,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 4.108396530151367,
+ "step": 1540
+ },
+ {
+ "epoch": 21.404886561954626,
+ "grad_norm": 0.05129663646221161,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 4.1135711669921875,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418848167539267,
+ "grad_norm": 0.05478592589497566,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 4.11839485168457,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43280977312391,
+ "grad_norm": 0.055830392986536026,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 4.11065149307251,
+ "step": 1543
+ },
+ {
+ "epoch": 21.44677137870855,
+ "grad_norm": 0.05370553582906723,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 4.113103866577148,
+ "step": 1544
+ },
+ {
+ "epoch": 21.460732984293195,
+ "grad_norm": 0.04968908056616783,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 4.119443893432617,
+ "step": 1545
+ },
+ {
+ "epoch": 21.474694589877835,
+ "grad_norm": 0.056149110198020935,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 4.126479625701904,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48865619546248,
+ "grad_norm": 0.053045690059661865,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 4.138535499572754,
+ "step": 1547
+ },
+ {
+ "epoch": 21.50261780104712,
+ "grad_norm": 0.04864223673939705,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 4.099347114562988,
+ "step": 1548
+ },
+ {
+ "epoch": 21.516579406631763,
+ "grad_norm": 0.05234768986701965,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 4.112222194671631,
+ "step": 1549
+ },
+ {
+ "epoch": 21.530541012216403,
+ "grad_norm": 0.050486158579587936,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 4.107913970947266,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544502617801047,
+ "grad_norm": 0.048641227185726166,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 4.114014148712158,
+ "step": 1551
+ },
+ {
+ "epoch": 21.55846422338569,
+ "grad_norm": 0.047759316861629486,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 4.128541946411133,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57242582897033,
+ "grad_norm": 0.046098388731479645,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 4.12142276763916,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586387434554975,
+ "grad_norm": 0.04934769123792648,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 4.13035249710083,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600349040139616,
+ "grad_norm": 0.048911672085523605,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 4.111344337463379,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61431064572426,
+ "grad_norm": 0.049368683248758316,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 4.1345672607421875,
+ "step": 1556
+ },
+ {
+ "epoch": 21.6282722513089,
+ "grad_norm": 0.04890736937522888,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 4.1286821365356445,
+ "step": 1557
+ },
+ {
+ "epoch": 21.642233856893544,
+ "grad_norm": 0.04812583327293396,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 4.09993314743042,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656195462478184,
+ "grad_norm": 0.05025152117013931,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 4.150769233703613,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670157068062828,
+ "grad_norm": 0.05007397383451462,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 4.074525833129883,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68411867364747,
+ "grad_norm": 0.050168465822935104,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 4.0933732986450195,
+ "step": 1561
+ },
+ {
+ "epoch": 21.698080279232112,
+ "grad_norm": 0.04990607872605324,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 4.126741886138916,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712041884816752,
+ "grad_norm": 0.05023440346121788,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 4.120314598083496,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726003490401396,
+ "grad_norm": 0.048100546002388,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 4.1242356300354,
+ "step": 1564
+ },
+ {
+ "epoch": 21.739965095986037,
+ "grad_norm": 0.047999415546655655,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 4.127745628356934,
+ "step": 1565
+ },
+ {
+ "epoch": 21.75392670157068,
+ "grad_norm": 0.05180094763636589,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 4.112565994262695,
+ "step": 1566
+ },
+ {
+ "epoch": 21.767888307155324,
+ "grad_norm": 0.053367964923381805,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 4.109419822692871,
+ "step": 1567
+ },
+ {
+ "epoch": 21.781849912739965,
+ "grad_norm": 0.05205400660634041,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 4.111886024475098,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79581151832461,
+ "grad_norm": 0.0464242659509182,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 4.104648590087891,
+ "step": 1569
+ },
+ {
+ "epoch": 21.80977312390925,
+ "grad_norm": 0.047767721116542816,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 4.156681060791016,
+ "step": 1570
+ },
+ {
+ "epoch": 21.823734729493893,
+ "grad_norm": 0.05028790235519409,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 4.132200241088867,
+ "step": 1571
+ },
+ {
+ "epoch": 21.837696335078533,
+ "grad_norm": 0.04714380204677582,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 4.132688999176025,
+ "step": 1572
+ },
+ {
+ "epoch": 21.851657940663177,
+ "grad_norm": 0.04726320132613182,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 4.173277378082275,
+ "step": 1573
+ },
+ {
+ "epoch": 21.865619546247817,
+ "grad_norm": 0.049558594822883606,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 4.15473747253418,
+ "step": 1574
+ },
+ {
+ "epoch": 21.87958115183246,
+ "grad_norm": 0.04742267727851868,
+ "learning_rate": 0.000375827577564042,
+ "loss": 4.10634708404541,
+ "step": 1575
+ },
+ {
+ "epoch": 21.8935427574171,
+ "grad_norm": 0.048223357647657394,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 4.133949279785156,
+ "step": 1576
+ },
+ {
+ "epoch": 21.907504363001745,
+ "grad_norm": 0.046766068786382675,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 4.153807640075684,
+ "step": 1577
+ },
+ {
+ "epoch": 21.921465968586386,
+ "grad_norm": 0.04586656019091606,
+ "learning_rate": 0.000375051971546236,
+ "loss": 4.169903755187988,
+ "step": 1578
+ },
+ {
+ "epoch": 21.93542757417103,
+ "grad_norm": 0.047912538051605225,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 4.148102283477783,
+ "step": 1579
+ },
+ {
+ "epoch": 21.949389179755673,
+ "grad_norm": 0.048949215561151505,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 4.149707317352295,
+ "step": 1580
+ },
+ {
+ "epoch": 21.963350785340314,
+ "grad_norm": 0.04765952378511429,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 4.128008842468262,
+ "step": 1581
+ },
+ {
+ "epoch": 21.977312390924958,
+ "grad_norm": 0.04726482927799225,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 4.1268720626831055,
+ "step": 1582
+ },
+ {
+ "epoch": 21.991273996509598,
+ "grad_norm": 0.0468299426138401,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 4.1332926750183105,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.03415689244866371,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 2.5875940322875977,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.5966277122497559,
+ "eval_runtime": 59.7704,
+ "eval_samples_per_second": 40.856,
+ "eval_steps_per_second": 0.652,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013961605584644,
+ "grad_norm": 0.04954535514116287,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 4.054527759552002,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027923211169284,
+ "grad_norm": 0.0498427115380764,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 4.048842430114746,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041884816753928,
+ "grad_norm": 0.051013074815273285,
+ "learning_rate": 0.000372722041257003,
+ "loss": 4.036187171936035,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05584642233857,
+ "grad_norm": 0.04973271116614342,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 4.050318717956543,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069808027923212,
+ "grad_norm": 0.051144856959581375,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 4.036319732666016,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083769633507853,
+ "grad_norm": 0.051874030381441116,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 4.017857551574707,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097731239092496,
+ "grad_norm": 0.05181736871600151,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 4.0519819259643555,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111692844677137,
+ "grad_norm": 0.05467049032449722,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 4.064291954040527,
+ "step": 1592
+ },
+ {
+ "epoch": 22.12565445026178,
+ "grad_norm": 0.05251661688089371,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 4.056951522827148,
+ "step": 1593
+ },
+ {
+ "epoch": 22.13961605584642,
+ "grad_norm": 0.05278680846095085,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 4.037125587463379,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153577661431065,
+ "grad_norm": 0.05276667699217796,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 4.053675174713135,
+ "step": 1595
+ },
+ {
+ "epoch": 22.167539267015705,
+ "grad_norm": 0.05007271468639374,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 4.044122695922852,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18150087260035,
+ "grad_norm": 0.051073770970106125,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 4.0265960693359375,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195462478184993,
+ "grad_norm": 0.05045647919178009,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 4.060938835144043,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209424083769633,
+ "grad_norm": 0.05075158178806305,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 4.035902976989746,
+ "step": 1599
+ },
+ {
+ "epoch": 22.223385689354277,
+ "grad_norm": 0.04867244139313698,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 4.055220603942871,
+ "step": 1600
+ },
+ {
+ "epoch": 22.237347294938917,
+ "grad_norm": 0.049976419657468796,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 4.026124954223633,
+ "step": 1601
+ },
+ {
+ "epoch": 22.25130890052356,
+ "grad_norm": 0.04968957230448723,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 4.081721782684326,
+ "step": 1602
+ },
+ {
+ "epoch": 22.2652705061082,
+ "grad_norm": 0.0503058098256588,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 4.0384721755981445,
+ "step": 1603
+ },
+ {
+ "epoch": 22.279232111692846,
+ "grad_norm": 0.050643209367990494,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 4.049397945404053,
+ "step": 1604
+ },
+ {
+ "epoch": 22.293193717277486,
+ "grad_norm": 0.05297046899795532,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 4.046389579772949,
+ "step": 1605
+ },
+ {
+ "epoch": 22.30715532286213,
+ "grad_norm": 0.053165238350629807,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 4.090221405029297,
+ "step": 1606
+ },
+ {
+ "epoch": 22.32111692844677,
+ "grad_norm": 0.05104832351207733,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 4.023227691650391,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335078534031414,
+ "grad_norm": 0.048736415803432465,
+ "learning_rate": 0.000367268105959126,
+ "loss": 4.07478666305542,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349040139616054,
+ "grad_norm": 0.05190873518586159,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 4.068147659301758,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363001745200698,
+ "grad_norm": 0.05409928783774376,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 4.060364723205566,
+ "step": 1610
+ },
+ {
+ "epoch": 22.376963350785342,
+ "grad_norm": 0.05112519487738609,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 4.072710990905762,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390924956369982,
+ "grad_norm": 0.04818312078714371,
+ "learning_rate": 0.000366226612810024,
+ "loss": 4.042757034301758,
+ "step": 1612
+ },
+ {
+ "epoch": 22.404886561954626,
+ "grad_norm": 0.050733599811792374,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 4.064810752868652,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418848167539267,
+ "grad_norm": 0.054070211946964264,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 4.054926872253418,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43280977312391,
+ "grad_norm": 0.05459332838654518,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 4.0652923583984375,
+ "step": 1615
+ },
+ {
+ "epoch": 22.44677137870855,
+ "grad_norm": 0.05064154416322708,
+ "learning_rate": 0.000365184304613104,
+ "loss": 4.076216220855713,
+ "step": 1616
+ },
+ {
+ "epoch": 22.460732984293195,
+ "grad_norm": 0.05071895569562912,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 4.070466041564941,
+ "step": 1617
+ },
+ {
+ "epoch": 22.474694589877835,
+ "grad_norm": 0.05049987882375717,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 4.062691688537598,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48865619546248,
+ "grad_norm": 0.05174451321363449,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 4.07501220703125,
+ "step": 1619
+ },
+ {
+ "epoch": 22.50261780104712,
+ "grad_norm": 0.05360838770866394,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 4.054711818695068,
+ "step": 1620
+ },
+ {
+ "epoch": 22.516579406631763,
+ "grad_norm": 0.051809027791023254,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 4.075811386108398,
+ "step": 1621
+ },
+ {
+ "epoch": 22.530541012216403,
+ "grad_norm": 0.05586795508861542,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 4.0638203620910645,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544502617801047,
+ "grad_norm": 0.05571979284286499,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 4.089414596557617,
+ "step": 1623
+ },
+ {
+ "epoch": 22.55846422338569,
+ "grad_norm": 0.05098723992705345,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 4.091159820556641,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57242582897033,
+ "grad_norm": 0.05359277129173279,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 4.060613632202148,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586387434554975,
+ "grad_norm": 0.05281100794672966,
+ "learning_rate": 0.000362575056487632,
+ "loss": 4.08416748046875,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600349040139616,
+ "grad_norm": 0.053913842886686325,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 4.088703155517578,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61431064572426,
+ "grad_norm": 0.05476798862218857,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 4.064074516296387,
+ "step": 1628
+ },
+ {
+ "epoch": 22.6282722513089,
+ "grad_norm": 0.051888495683670044,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 4.116634368896484,
+ "step": 1629
+ },
+ {
+ "epoch": 22.642233856893544,
+ "grad_norm": 0.051054876297712326,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 4.0800018310546875,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656195462478184,
+ "grad_norm": 0.054063159972429276,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 4.112773418426514,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670157068062828,
+ "grad_norm": 0.05173034593462944,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 4.062599182128906,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68411867364747,
+ "grad_norm": 0.054618336260318756,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 4.110188961029053,
+ "step": 1633
+ },
+ {
+ "epoch": 22.698080279232112,
+ "grad_norm": 0.055602896958589554,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 4.140998840332031,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712041884816752,
+ "grad_norm": 0.05332222580909729,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 4.119804382324219,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726003490401396,
+ "grad_norm": 0.05162122845649719,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 4.0731282234191895,
+ "step": 1636
+ },
+ {
+ "epoch": 22.739965095986037,
+ "grad_norm": 0.0514204278588295,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 4.068007946014404,
+ "step": 1637
+ },
+ {
+ "epoch": 22.75392670157068,
+ "grad_norm": 0.05027637258172035,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 4.097658157348633,
+ "step": 1638
+ },
+ {
+ "epoch": 22.767888307155324,
+ "grad_norm": 0.051812347024679184,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 4.083312511444092,
+ "step": 1639
+ },
+ {
+ "epoch": 22.781849912739965,
+ "grad_norm": 0.05145815759897232,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 4.129471778869629,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79581151832461,
+ "grad_norm": 0.05093684792518616,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 4.091841220855713,
+ "step": 1641
+ },
+ {
+ "epoch": 22.80977312390925,
+ "grad_norm": 0.05100737139582634,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 4.076615810394287,
+ "step": 1642
+ },
+ {
+ "epoch": 22.823734729493893,
+ "grad_norm": 0.05079149082303047,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 4.067737579345703,
+ "step": 1643
+ },
+ {
+ "epoch": 22.837696335078533,
+ "grad_norm": 0.05044509842991829,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 4.118824481964111,
+ "step": 1644
+ },
+ {
+ "epoch": 22.851657940663177,
+ "grad_norm": 0.05010584369301796,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 4.113405227661133,
+ "step": 1645
+ },
+ {
+ "epoch": 22.865619546247817,
+ "grad_norm": 0.052232302725315094,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 4.101553440093994,
+ "step": 1646
+ },
+ {
+ "epoch": 22.87958115183246,
+ "grad_norm": 0.05080661550164223,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 4.116809844970703,
+ "step": 1647
+ },
+ {
+ "epoch": 22.8935427574171,
+ "grad_norm": 0.05051033943891525,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 4.128484725952148,
+ "step": 1648
+ },
+ {
+ "epoch": 22.907504363001745,
+ "grad_norm": 0.051927294582128525,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 4.092285633087158,
+ "step": 1649
+ },
+ {
+ "epoch": 22.921465968586386,
+ "grad_norm": 0.049063827842473984,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 4.088473796844482,
+ "step": 1650
+ },
+ {
+ "epoch": 22.93542757417103,
+ "grad_norm": 0.05048854276537895,
+ "learning_rate": 0.000356031397755195,
+ "loss": 4.081521987915039,
+ "step": 1651
+ },
+ {
+ "epoch": 22.949389179755673,
+ "grad_norm": 0.049307651817798615,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 4.122158050537109,
+ "step": 1652
+ },
+ {
+ "epoch": 22.963350785340314,
+ "grad_norm": 0.048311345279216766,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 4.100488662719727,
+ "step": 1653
+ },
+ {
+ "epoch": 22.977312390924958,
+ "grad_norm": 0.05016734451055527,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 4.088587760925293,
+ "step": 1654
+ },
+ {
+ "epoch": 22.991273996509598,
+ "grad_norm": 0.04944450408220291,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 4.114686012268066,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.03592754527926445,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 2.545376777648926,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.5981184840202332,
+ "eval_runtime": 59.9558,
+ "eval_samples_per_second": 40.73,
+ "eval_steps_per_second": 0.65,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013961605584644,
+ "grad_norm": 0.05112785845994949,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 4.00495719909668,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027923211169284,
+ "grad_norm": 0.05807499587535858,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 4.015654563903809,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041884816753928,
+ "grad_norm": 0.05383148789405823,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.979349136352539,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05584642233857,
+ "grad_norm": 0.05221369490027428,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 4.009493827819824,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069808027923212,
+ "grad_norm": 0.05144324526190758,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 4.045680999755859,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083769633507853,
+ "grad_norm": 0.051422782242298126,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.994011878967285,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097731239092496,
+ "grad_norm": 0.05142863467335701,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 4.03209114074707,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111692844677137,
+ "grad_norm": 0.050954073667526245,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 4.00391960144043,
+ "step": 1664
+ },
+ {
+ "epoch": 23.12565445026178,
+ "grad_norm": 0.05365024879574776,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.9839367866516113,
+ "step": 1665
+ },
+ {
+ "epoch": 23.13961605584642,
+ "grad_norm": 0.051181286573410034,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 4.017707824707031,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153577661431065,
+ "grad_norm": 0.054279860109090805,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.989722490310669,
+ "step": 1667
+ },
+ {
+ "epoch": 23.167539267015705,
+ "grad_norm": 0.052196286618709564,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 4.014932155609131,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18150087260035,
+ "grad_norm": 0.05098043382167816,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 4.026510238647461,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195462478184993,
+ "grad_norm": 0.05241888761520386,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.979708433151245,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209424083769633,
+ "grad_norm": 0.051801104098558426,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 4.029015064239502,
+ "step": 1671
+ },
+ {
+ "epoch": 23.223385689354277,
+ "grad_norm": 0.052666619420051575,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.9932432174682617,
+ "step": 1672
+ },
+ {
+ "epoch": 23.237347294938917,
+ "grad_norm": 0.054418303072452545,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 4.05676794052124,
+ "step": 1673
+ },
+ {
+ "epoch": 23.25130890052356,
+ "grad_norm": 0.055074021220207214,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.999929666519165,
+ "step": 1674
+ },
+ {
+ "epoch": 23.2652705061082,
+ "grad_norm": 0.05626136437058449,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 4.013473987579346,
+ "step": 1675
+ },
+ {
+ "epoch": 23.279232111692846,
+ "grad_norm": 0.054028257727622986,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 4.015802383422852,
+ "step": 1676
+ },
+ {
+ "epoch": 23.293193717277486,
+ "grad_norm": 0.054044198244810104,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 4.020464897155762,
+ "step": 1677
+ },
+ {
+ "epoch": 23.30715532286213,
+ "grad_norm": 0.05465259402990341,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 4.010867118835449,
+ "step": 1678
+ },
+ {
+ "epoch": 23.32111692844677,
+ "grad_norm": 0.05545585975050926,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 4.054412841796875,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335078534031414,
+ "grad_norm": 0.055265430361032486,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 4.020756244659424,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349040139616054,
+ "grad_norm": 0.053363267332315445,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 4.033337116241455,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363001745200698,
+ "grad_norm": 0.05347830802202225,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 4.024794578552246,
+ "step": 1682
+ },
+ {
+ "epoch": 23.376963350785342,
+ "grad_norm": 0.05496951565146446,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 4.013302803039551,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390924956369982,
+ "grad_norm": 0.05179804563522339,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 4.030914306640625,
+ "step": 1684
+ },
+ {
+ "epoch": 23.404886561954626,
+ "grad_norm": 0.054958634078502655,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 4.012727737426758,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418848167539267,
+ "grad_norm": 0.05533262714743614,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 4.0543365478515625,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43280977312391,
+ "grad_norm": 0.052329424768686295,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 4.031486511230469,
+ "step": 1687
+ },
+ {
+ "epoch": 23.44677137870855,
+ "grad_norm": 0.05273434892296791,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 4.040292739868164,
+ "step": 1688
+ },
+ {
+ "epoch": 23.460732984293195,
+ "grad_norm": 0.05445481091737747,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 4.025630950927734,
+ "step": 1689
+ },
+ {
+ "epoch": 23.474694589877835,
+ "grad_norm": 0.052102912217378616,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 4.049562454223633,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48865619546248,
+ "grad_norm": 0.051272716373205185,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 4.0431365966796875,
+ "step": 1691
+ },
+ {
+ "epoch": 23.50261780104712,
+ "grad_norm": 0.051311176270246506,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 4.042119026184082,
+ "step": 1692
+ },
+ {
+ "epoch": 23.516579406631763,
+ "grad_norm": 0.05042742192745209,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 4.044881820678711,
+ "step": 1693
+ },
+ {
+ "epoch": 23.530541012216403,
+ "grad_norm": 0.05103316903114319,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 4.033669471740723,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544502617801047,
+ "grad_norm": 0.05441150814294815,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 4.076536178588867,
+ "step": 1695
+ },
+ {
+ "epoch": 23.55846422338569,
+ "grad_norm": 0.05529651418328285,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 4.073678016662598,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57242582897033,
+ "grad_norm": 0.0526033453643322,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 4.067898750305176,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586387434554975,
+ "grad_norm": 0.052817948162555695,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 4.017403602600098,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600349040139616,
+ "grad_norm": 0.052461300045251846,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 4.026311874389648,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61431064572426,
+ "grad_norm": 0.05097689852118492,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 4.076074600219727,
+ "step": 1700
+ },
+ {
+ "epoch": 23.6282722513089,
+ "grad_norm": 0.05387786403298378,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 4.022763729095459,
+ "step": 1701
+ },
+ {
+ "epoch": 23.642233856893544,
+ "grad_norm": 0.05132855102419853,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 4.077712535858154,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656195462478184,
+ "grad_norm": 0.05380886048078537,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 4.038937091827393,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670157068062828,
+ "grad_norm": 0.053299430757761,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 4.063364028930664,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68411867364747,
+ "grad_norm": 0.051913514733314514,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 4.05495548248291,
+ "step": 1705
+ },
+ {
+ "epoch": 23.698080279232112,
+ "grad_norm": 0.05660073831677437,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 4.057949066162109,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712041884816752,
+ "grad_norm": 0.054757390171289444,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 4.05377721786499,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726003490401396,
+ "grad_norm": 0.0517287515103817,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 4.035928726196289,
+ "step": 1708
+ },
+ {
+ "epoch": 23.739965095986037,
+ "grad_norm": 0.05187627300620079,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 4.043362617492676,
+ "step": 1709
+ },
+ {
+ "epoch": 23.75392670157068,
+ "grad_norm": 0.053502220660448074,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 4.007593154907227,
+ "step": 1710
+ },
+ {
+ "epoch": 23.767888307155324,
+ "grad_norm": 0.05755629390478134,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 4.076575756072998,
+ "step": 1711
+ },
+ {
+ "epoch": 23.781849912739965,
+ "grad_norm": 0.05619359388947487,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 4.072240352630615,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79581151832461,
+ "grad_norm": 0.05604318529367447,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 4.0486650466918945,
+ "step": 1713
+ },
+ {
+ "epoch": 23.80977312390925,
+ "grad_norm": 0.05630200728774071,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 4.086282730102539,
+ "step": 1714
+ },
+ {
+ "epoch": 23.823734729493893,
+ "grad_norm": 0.053262822329998016,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 4.089612007141113,
+ "step": 1715
+ },
+ {
+ "epoch": 23.837696335078533,
+ "grad_norm": 0.05699621140956879,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 4.053062438964844,
+ "step": 1716
+ },
+ {
+ "epoch": 23.851657940663177,
+ "grad_norm": 0.05963658541440964,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 4.072884559631348,
+ "step": 1717
+ },
+ {
+ "epoch": 23.865619546247817,
+ "grad_norm": 0.0547492615878582,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 4.066218852996826,
+ "step": 1718
+ },
+ {
+ "epoch": 23.87958115183246,
+ "grad_norm": 0.054908450692892075,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 4.035344123840332,
+ "step": 1719
+ },
+ {
+ "epoch": 23.8935427574171,
+ "grad_norm": 0.055191367864608765,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 4.08471155166626,
+ "step": 1720
+ },
+ {
+ "epoch": 23.907504363001745,
+ "grad_norm": 0.05334226414561272,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 4.076174736022949,
+ "step": 1721
+ },
+ {
+ "epoch": 23.921465968586386,
+ "grad_norm": 0.05579005926847458,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 4.05291223526001,
+ "step": 1722
+ },
+ {
+ "epoch": 23.93542757417103,
+ "grad_norm": 0.05464969575405121,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 4.07335090637207,
+ "step": 1723
+ },
+ {
+ "epoch": 23.949389179755673,
+ "grad_norm": 0.05240754038095474,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 4.0995073318481445,
+ "step": 1724
+ },
+ {
+ "epoch": 23.963350785340314,
+ "grad_norm": 0.05416056886315346,
+ "learning_rate": 0.000336518346307424,
+ "loss": 4.072793960571289,
+ "step": 1725
+ },
+ {
+ "epoch": 23.977312390924958,
+ "grad_norm": 0.05496834218502045,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 4.059562683105469,
+ "step": 1726
+ },
+ {
+ "epoch": 23.991273996509598,
+ "grad_norm": 0.05537887290120125,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 4.050772666931152,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.036677468568086624,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 2.5407257080078125,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.5995977520942688,
+ "eval_runtime": 60.2704,
+ "eval_samples_per_second": 40.517,
+ "eval_steps_per_second": 0.647,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013961605584644,
+ "grad_norm": 0.05583682656288147,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.9873995780944824,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027923211169284,
+ "grad_norm": 0.0605769157409668,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.999206066131592,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041884816753928,
+ "grad_norm": 0.057975225150585175,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.9923095703125,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05584642233857,
+ "grad_norm": 0.056104276329278946,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.946317672729492,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069808027923212,
+ "grad_norm": 0.05481491982936859,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.9973483085632324,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083769633507853,
+ "grad_norm": 0.05460227280855179,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.984617233276367,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097731239092496,
+ "grad_norm": 0.058481909334659576,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 4.0138773918151855,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111692844677137,
+ "grad_norm": 0.05672750994563103,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.9683098793029785,
+ "step": 1736
+ },
+ {
+ "epoch": 24.12565445026178,
+ "grad_norm": 0.05315157398581505,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.9909560680389404,
+ "step": 1737
+ },
+ {
+ "epoch": 24.13961605584642,
+ "grad_norm": 0.05565650761127472,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.954531192779541,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153577661431065,
+ "grad_norm": 0.05550679191946983,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.97672700881958,
+ "step": 1739
+ },
+ {
+ "epoch": 24.167539267015705,
+ "grad_norm": 0.05608576908707619,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.987941265106201,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18150087260035,
+ "grad_norm": 0.05266846343874931,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.97526216506958,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195462478184993,
+ "grad_norm": 0.05439139902591705,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 4.012054920196533,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209424083769633,
+ "grad_norm": 0.05473192036151886,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 4.022945404052734,
+ "step": 1743
+ },
+ {
+ "epoch": 24.223385689354277,
+ "grad_norm": 0.054737839847803116,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.981072425842285,
+ "step": 1744
+ },
+ {
+ "epoch": 24.237347294938917,
+ "grad_norm": 0.05696691572666168,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.9708170890808105,
+ "step": 1745
+ },
+ {
+ "epoch": 24.25130890052356,
+ "grad_norm": 0.05677401274442673,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.985995292663574,
+ "step": 1746
+ },
+ {
+ "epoch": 24.2652705061082,
+ "grad_norm": 0.05895088613033295,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.991887092590332,
+ "step": 1747
+ },
+ {
+ "epoch": 24.279232111692846,
+ "grad_norm": 0.05758961662650108,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.9695687294006348,
+ "step": 1748
+ },
+ {
+ "epoch": 24.293193717277486,
+ "grad_norm": 0.05174405500292778,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.959138870239258,
+ "step": 1749
+ },
+ {
+ "epoch": 24.30715532286213,
+ "grad_norm": 0.055284611880779266,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 4.012698173522949,
+ "step": 1750
+ },
+ {
+ "epoch": 24.32111692844677,
+ "grad_norm": 0.06042446941137314,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.9892425537109375,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335078534031414,
+ "grad_norm": 0.05489135906100273,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.9962222576141357,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349040139616054,
+ "grad_norm": 0.0542326495051384,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.9787094593048096,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363001745200698,
+ "grad_norm": 0.05612313374876976,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.9881339073181152,
+ "step": 1754
+ },
+ {
+ "epoch": 24.376963350785342,
+ "grad_norm": 0.05661709979176521,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 4.009440898895264,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390924956369982,
+ "grad_norm": 0.055537596344947815,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 4.00221061706543,
+ "step": 1756
+ },
+ {
+ "epoch": 24.404886561954626,
+ "grad_norm": 0.054210033267736435,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.9807987213134766,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418848167539267,
+ "grad_norm": 0.05847020074725151,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.960838794708252,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43280977312391,
+ "grad_norm": 0.05386386811733246,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 4.017860412597656,
+ "step": 1759
+ },
+ {
+ "epoch": 24.44677137870855,
+ "grad_norm": 0.056131865829229355,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.980543613433838,
+ "step": 1760
+ },
+ {
+ "epoch": 24.460732984293195,
+ "grad_norm": 0.05648239329457283,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.9948508739471436,
+ "step": 1761
+ },
+ {
+ "epoch": 24.474694589877835,
+ "grad_norm": 0.054138120263814926,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 4.002998352050781,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48865619546248,
+ "grad_norm": 0.05725586414337158,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 4.01864767074585,
+ "step": 1763
+ },
+ {
+ "epoch": 24.50261780104712,
+ "grad_norm": 0.05680665746331215,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 4.023143768310547,
+ "step": 1764
+ },
+ {
+ "epoch": 24.516579406631763,
+ "grad_norm": 0.052627697587013245,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 4.006389617919922,
+ "step": 1765
+ },
+ {
+ "epoch": 24.530541012216403,
+ "grad_norm": 0.053234782069921494,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 4.007419586181641,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544502617801047,
+ "grad_norm": 0.05756475403904915,
+ "learning_rate": 0.000325372061241796,
+ "loss": 4.0226945877075195,
+ "step": 1767
+ },
+ {
+ "epoch": 24.55846422338569,
+ "grad_norm": 0.055328067392110825,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 4.022767066955566,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57242582897033,
+ "grad_norm": 0.05378621071577072,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 4.005022048950195,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586387434554975,
+ "grad_norm": 0.056049644947052,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 4.025008201599121,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600349040139616,
+ "grad_norm": 0.057860117405653,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 4.003593444824219,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61431064572426,
+ "grad_norm": 0.056685276329517365,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.997868299484253,
+ "step": 1772
+ },
+ {
+ "epoch": 24.6282722513089,
+ "grad_norm": 0.053294967859983444,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 4.0211992263793945,
+ "step": 1773
+ },
+ {
+ "epoch": 24.642233856893544,
+ "grad_norm": 0.053967639803886414,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 4.022367477416992,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656195462478184,
+ "grad_norm": 0.052357546985149384,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 4.021792411804199,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670157068062828,
+ "grad_norm": 0.0547296479344368,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 4.008403778076172,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68411867364747,
+ "grad_norm": 0.056104280054569244,
+ "learning_rate": 0.000322712903929477,
+ "loss": 4.033827781677246,
+ "step": 1777
+ },
+ {
+ "epoch": 24.698080279232112,
+ "grad_norm": 0.055415406823158264,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 4.015371322631836,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712041884816752,
+ "grad_norm": 0.054506298154592514,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 4.0000152587890625,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726003490401396,
+ "grad_norm": 0.05468682199716568,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.9905481338500977,
+ "step": 1780
+ },
+ {
+ "epoch": 24.739965095986037,
+ "grad_norm": 0.05323269963264465,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 4.040729999542236,
+ "step": 1781
+ },
+ {
+ "epoch": 24.75392670157068,
+ "grad_norm": 0.052569612860679626,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 4.028299331665039,
+ "step": 1782
+ },
+ {
+ "epoch": 24.767888307155324,
+ "grad_norm": 0.052494753152132034,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 4.028860569000244,
+ "step": 1783
+ },
+ {
+ "epoch": 24.781849912739965,
+ "grad_norm": 0.05346379801630974,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 4.000349044799805,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79581151832461,
+ "grad_norm": 0.051543332636356354,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 4.016782760620117,
+ "step": 1785
+ },
+ {
+ "epoch": 24.80977312390925,
+ "grad_norm": 0.05330873280763626,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 4.024371147155762,
+ "step": 1786
+ },
+ {
+ "epoch": 24.823734729493893,
+ "grad_norm": 0.05306831747293472,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 4.0101213455200195,
+ "step": 1787
+ },
+ {
+ "epoch": 24.837696335078533,
+ "grad_norm": 0.05148417130112648,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 4.020768642425537,
+ "step": 1788
+ },
+ {
+ "epoch": 24.851657940663177,
+ "grad_norm": 0.050848908722400665,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 4.030246734619141,
+ "step": 1789
+ },
+ {
+ "epoch": 24.865619546247817,
+ "grad_norm": 0.05279014632105827,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 4.04910135269165,
+ "step": 1790
+ },
+ {
+ "epoch": 24.87958115183246,
+ "grad_norm": 0.0514594204723835,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 4.025290489196777,
+ "step": 1791
+ },
+ {
+ "epoch": 24.8935427574171,
+ "grad_norm": 0.05200032889842987,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 4.019560813903809,
+ "step": 1792
+ },
+ {
+ "epoch": 24.907504363001745,
+ "grad_norm": 0.052474118769168854,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 4.035177230834961,
+ "step": 1793
+ },
+ {
+ "epoch": 24.921465968586386,
+ "grad_norm": 0.05411219224333763,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.9920859336853027,
+ "step": 1794
+ },
+ {
+ "epoch": 24.93542757417103,
+ "grad_norm": 0.0543992817401886,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 4.017391204833984,
+ "step": 1795
+ },
+ {
+ "epoch": 24.949389179755673,
+ "grad_norm": 0.05187678337097168,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 4.042266368865967,
+ "step": 1796
+ },
+ {
+ "epoch": 24.963350785340314,
+ "grad_norm": 0.053762417286634445,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 4.029013633728027,
+ "step": 1797
+ },
+ {
+ "epoch": 24.977312390924958,
+ "grad_norm": 0.051723796874284744,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 4.038886070251465,
+ "step": 1798
+ },
+ {
+ "epoch": 24.991273996509598,
+ "grad_norm": 0.05242295190691948,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 4.056063652038574,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.03998562693595886,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 2.534770965576172,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 0.6009227633476257,
+ "eval_runtime": 58.9421,
+ "eval_samples_per_second": 41.43,
+ "eval_steps_per_second": 0.662,
+ "step": 1800
+ },
+ {
+ "epoch": 25.013961605584644,
+ "grad_norm": 0.05450186878442764,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 3.936816692352295,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027923211169284,
+ "grad_norm": 0.06389589607715607,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 3.9368391036987305,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041884816753928,
+ "grad_norm": 0.06028348580002785,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 3.953296661376953,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05584642233857,
+ "grad_norm": 0.05359262600541115,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 3.924558639526367,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069808027923212,
+ "grad_norm": 0.059837762266397476,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 3.9608402252197266,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083769633507853,
+ "grad_norm": 0.05523517727851868,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 3.943776845932007,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097731239092496,
+ "grad_norm": 0.05617433041334152,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 3.9635634422302246,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111692844677137,
+ "grad_norm": 0.06066269427537918,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 3.9647364616394043,
+ "step": 1808
+ },
+ {
+ "epoch": 25.12565445026178,
+ "grad_norm": 0.05491397902369499,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 3.95222806930542,
+ "step": 1809
+ },
+ {
+ "epoch": 25.13961605584642,
+ "grad_norm": 0.05731203034520149,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 3.9440197944641113,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153577661431065,
+ "grad_norm": 0.05428403243422508,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 3.9428441524505615,
+ "step": 1811
+ },
+ {
+ "epoch": 25.167539267015705,
+ "grad_norm": 0.05487453565001488,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 3.9489784240722656,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18150087260035,
+ "grad_norm": 0.05725884810090065,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 3.9643478393554688,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195462478184993,
+ "grad_norm": 0.05752462521195412,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 3.963301181793213,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209424083769633,
+ "grad_norm": 0.057387594133615494,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 3.9427294731140137,
+ "step": 1815
+ },
+ {
+ "epoch": 25.223385689354277,
+ "grad_norm": 0.057492151856422424,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 3.9229259490966797,
+ "step": 1816
+ },
+ {
+ "epoch": 25.237347294938917,
+ "grad_norm": 0.05202317237854004,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 3.938037395477295,
+ "step": 1817
+ },
+ {
+ "epoch": 25.25130890052356,
+ "grad_norm": 0.05574658885598183,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 3.9532949924468994,
+ "step": 1818
+ },
+ {
+ "epoch": 25.2652705061082,
+ "grad_norm": 0.05714752897620201,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 3.962207794189453,
+ "step": 1819
+ },
+ {
+ "epoch": 25.279232111692846,
+ "grad_norm": 0.05540093779563904,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 3.96329402923584,
+ "step": 1820
+ },
+ {
+ "epoch": 25.293193717277486,
+ "grad_norm": 0.056597232818603516,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 3.948793411254883,
+ "step": 1821
+ },
+ {
+ "epoch": 25.30715532286213,
+ "grad_norm": 0.055902402848005295,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 3.9111557006835938,
+ "step": 1822
+ },
+ {
+ "epoch": 25.32111692844677,
+ "grad_norm": 0.05509798601269722,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 3.929591655731201,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335078534031414,
+ "grad_norm": 0.05763949453830719,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 3.9758448600769043,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349040139616054,
+ "grad_norm": 0.053838349878787994,
+ "learning_rate": 0.000309930621834911,
+ "loss": 3.92457914352417,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363001745200698,
+ "grad_norm": 0.05467558279633522,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 3.9501285552978516,
+ "step": 1826
+ },
+ {
+ "epoch": 25.376963350785342,
+ "grad_norm": 0.05715392529964447,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 3.9734368324279785,
+ "step": 1827
+ },
+ {
+ "epoch": 25.390924956369982,
+ "grad_norm": 0.0566536970436573,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 3.9933152198791504,
+ "step": 1828
+ },
+ {
+ "epoch": 25.404886561954626,
+ "grad_norm": 0.053221236914396286,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 3.967787265777588,
+ "step": 1829
+ },
+ {
+ "epoch": 25.418848167539267,
+ "grad_norm": 0.05779031664133072,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 3.9415433406829834,
+ "step": 1830
+ },
+ {
+ "epoch": 25.43280977312391,
+ "grad_norm": 0.062013521790504456,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 3.9801456928253174,
+ "step": 1831
+ },
+ {
+ "epoch": 25.44677137870855,
+ "grad_norm": 0.05819004401564598,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 3.966928005218506,
+ "step": 1832
+ },
+ {
+ "epoch": 25.460732984293195,
+ "grad_norm": 0.05754564702510834,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 3.9612770080566406,
+ "step": 1833
+ },
+ {
+ "epoch": 25.474694589877835,
+ "grad_norm": 0.05707791447639465,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 3.935521125793457,
+ "step": 1834
+ },
+ {
+ "epoch": 25.48865619546248,
+ "grad_norm": 0.056380245834589005,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 4.006120681762695,
+ "step": 1835
+ },
+ {
+ "epoch": 25.50261780104712,
+ "grad_norm": 0.0551077239215374,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 3.9544754028320312,
+ "step": 1836
+ },
+ {
+ "epoch": 25.516579406631763,
+ "grad_norm": 0.05720360204577446,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 3.9564664363861084,
+ "step": 1837
+ },
+ {
+ "epoch": 25.530541012216403,
+ "grad_norm": 0.05522238463163376,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 4.0157880783081055,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544502617801047,
+ "grad_norm": 0.05304659530520439,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 3.957207202911377,
+ "step": 1839
+ },
+ {
+ "epoch": 25.55846422338569,
+ "grad_norm": 0.05569292604923248,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 3.9909608364105225,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57242582897033,
+ "grad_norm": 0.05347870662808418,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 3.9537010192871094,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586387434554975,
+ "grad_norm": 0.05572560802102089,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 3.9818642139434814,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600349040139616,
+ "grad_norm": 0.05624900385737419,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 3.995117425918579,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61431064572426,
+ "grad_norm": 0.05600332096219063,
+ "learning_rate": 0.000304866093757771,
+ "loss": 3.9480154514312744,
+ "step": 1844
+ },
+ {
+ "epoch": 25.6282722513089,
+ "grad_norm": 0.05488704890012741,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 3.993014335632324,
+ "step": 1845
+ },
+ {
+ "epoch": 25.642233856893544,
+ "grad_norm": 0.05647030100226402,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 3.994615077972412,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656195462478184,
+ "grad_norm": 0.05460594967007637,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 3.9714808464050293,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670157068062828,
+ "grad_norm": 0.05629213526844978,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 3.962033271789551,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68411867364747,
+ "grad_norm": 0.05666861683130264,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 3.995985269546509,
+ "step": 1849
+ },
+ {
+ "epoch": 25.698080279232112,
+ "grad_norm": 0.05580414459109306,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 4.001766681671143,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712041884816752,
+ "grad_norm": 0.05638023093342781,
+ "learning_rate": 0.000303,
+ "loss": 4.005331993103027,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726003490401396,
+ "grad_norm": 0.05684564262628555,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 4.017581939697266,
+ "step": 1852
+ },
+ {
+ "epoch": 25.739965095986037,
+ "grad_norm": 0.057602159678936005,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 3.9668312072753906,
+ "step": 1853
+ },
+ {
+ "epoch": 25.75392670157068,
+ "grad_norm": 0.055540211498737335,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 3.9885573387145996,
+ "step": 1854
+ },
+ {
+ "epoch": 25.767888307155324,
+ "grad_norm": 0.05655810981988907,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 3.990997314453125,
+ "step": 1855
+ },
+ {
+ "epoch": 25.781849912739965,
+ "grad_norm": 0.05443689972162247,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 3.9554123878479004,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79581151832461,
+ "grad_norm": 0.05570529028773308,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 3.983283519744873,
+ "step": 1857
+ },
+ {
+ "epoch": 25.80977312390925,
+ "grad_norm": 0.05399419739842415,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 4.004507064819336,
+ "step": 1858
+ },
+ {
+ "epoch": 25.823734729493893,
+ "grad_norm": 0.05587748438119888,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 4.00900936126709,
+ "step": 1859
+ },
+ {
+ "epoch": 25.837696335078533,
+ "grad_norm": 0.05465575307607651,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 3.9826579093933105,
+ "step": 1860
+ },
+ {
+ "epoch": 25.851657940663177,
+ "grad_norm": 0.05286499112844467,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 3.973453998565674,
+ "step": 1861
+ },
+ {
+ "epoch": 25.865619546247817,
+ "grad_norm": 0.05495590344071388,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 4.003095626831055,
+ "step": 1862
+ },
+ {
+ "epoch": 25.87958115183246,
+ "grad_norm": 0.053388018161058426,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 4.00042200088501,
+ "step": 1863
+ },
+ {
+ "epoch": 25.8935427574171,
+ "grad_norm": 0.05323237553238869,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 3.992929458618164,
+ "step": 1864
+ },
+ {
+ "epoch": 25.907504363001745,
+ "grad_norm": 0.053946711122989655,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 3.996519088745117,
+ "step": 1865
+ },
+ {
+ "epoch": 25.921465968586386,
+ "grad_norm": 0.05455862358212471,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 4.007123947143555,
+ "step": 1866
+ },
+ {
+ "epoch": 25.93542757417103,
+ "grad_norm": 0.05290355160832405,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 4.001724720001221,
+ "step": 1867
+ },
+ {
+ "epoch": 25.949389179755673,
+ "grad_norm": 0.052808016538619995,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 3.9964821338653564,
+ "step": 1868
+ },
+ {
+ "epoch": 25.963350785340314,
+ "grad_norm": 0.05416771396994591,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 3.98549222946167,
+ "step": 1869
+ },
+ {
+ "epoch": 25.977312390924958,
+ "grad_norm": 0.05264521390199661,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 3.9729766845703125,
+ "step": 1870
+ },
+ {
+ "epoch": 25.991273996509598,
+ "grad_norm": 0.05483962222933769,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 4.018877029418945,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.03927824646234512,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 2.456951856613159,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 0.6026532053947449,
+ "eval_runtime": 60.4079,
+ "eval_samples_per_second": 40.425,
+ "eval_steps_per_second": 0.646,
+ "step": 1872
+ },
+ {
+ "epoch": 26.013961605584644,
+ "grad_norm": 0.05756894499063492,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 3.93898606300354,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027923211169284,
+ "grad_norm": 0.06288876384496689,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 3.9142136573791504,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041884816753928,
+ "grad_norm": 0.0600322000682354,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 3.918700695037842,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05584642233857,
+ "grad_norm": 0.05729348957538605,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 3.9040162563323975,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069808027923212,
+ "grad_norm": 0.05658194422721863,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 3.8809690475463867,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083769633507853,
+ "grad_norm": 0.06274262815713882,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 3.928222179412842,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097731239092496,
+ "grad_norm": 0.06313030421733856,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 3.9395527839660645,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111692844677137,
+ "grad_norm": 0.0554068423807621,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 3.9314563274383545,
+ "step": 1880
+ },
+ {
+ "epoch": 26.12565445026178,
+ "grad_norm": 0.05697336047887802,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 3.9071288108825684,
+ "step": 1881
+ },
+ {
+ "epoch": 26.13961605584642,
+ "grad_norm": 0.05603424087166786,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 3.932054042816162,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153577661431065,
+ "grad_norm": 0.055603109300136566,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 3.9169063568115234,
+ "step": 1883
+ },
+ {
+ "epoch": 26.167539267015705,
+ "grad_norm": 0.055022306740283966,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 3.8985838890075684,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18150087260035,
+ "grad_norm": 0.05383468046784401,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 3.896359443664551,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195462478184993,
+ "grad_norm": 0.055076900869607925,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 3.905672073364258,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209424083769633,
+ "grad_norm": 0.05629882961511612,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 3.9222521781921387,
+ "step": 1887
+ },
+ {
+ "epoch": 26.223385689354277,
+ "grad_norm": 0.056736111640930176,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 3.8918867111206055,
+ "step": 1888
+ },
+ {
+ "epoch": 26.237347294938917,
+ "grad_norm": 0.05625338479876518,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 3.9249935150146484,
+ "step": 1889
+ },
+ {
+ "epoch": 26.25130890052356,
+ "grad_norm": 0.06145375967025757,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 3.904536724090576,
+ "step": 1890
+ },
+ {
+ "epoch": 26.2652705061082,
+ "grad_norm": 0.05601578950881958,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 3.8845863342285156,
+ "step": 1891
+ },
+ {
+ "epoch": 26.279232111692846,
+ "grad_norm": 0.056899815797805786,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 3.917966842651367,
+ "step": 1892
+ },
+ {
+ "epoch": 26.293193717277486,
+ "grad_norm": 0.060314927250146866,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 3.9230940341949463,
+ "step": 1893
+ },
+ {
+ "epoch": 26.30715532286213,
+ "grad_norm": 0.05449546128511429,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 3.9082441329956055,
+ "step": 1894
+ },
+ {
+ "epoch": 26.32111692844677,
+ "grad_norm": 0.05601051449775696,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 3.9174978733062744,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335078534031414,
+ "grad_norm": 0.05770742893218994,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 3.9183058738708496,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349040139616054,
+ "grad_norm": 0.05718272924423218,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 3.9438772201538086,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363001745200698,
+ "grad_norm": 0.052873238921165466,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 3.9554827213287354,
+ "step": 1898
+ },
+ {
+ "epoch": 26.376963350785342,
+ "grad_norm": 0.056853555142879486,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 3.886812210083008,
+ "step": 1899
+ },
+ {
+ "epoch": 26.390924956369982,
+ "grad_norm": 0.05662025138735771,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 3.9384284019470215,
+ "step": 1900
+ },
+ {
+ "epoch": 26.404886561954626,
+ "grad_norm": 0.05693329498171806,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 3.9123764038085938,
+ "step": 1901
+ },
+ {
+ "epoch": 26.418848167539267,
+ "grad_norm": 0.05425408110022545,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 3.9441678524017334,
+ "step": 1902
+ },
+ {
+ "epoch": 26.43280977312391,
+ "grad_norm": 0.05533420667052269,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 3.9086289405822754,
+ "step": 1903
+ },
+ {
+ "epoch": 26.44677137870855,
+ "grad_norm": 0.05527901649475098,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 3.9469594955444336,
+ "step": 1904
+ },
+ {
+ "epoch": 26.460732984293195,
+ "grad_norm": 0.052982039749622345,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 3.9182326793670654,
+ "step": 1905
+ },
+ {
+ "epoch": 26.474694589877835,
+ "grad_norm": 0.05520346015691757,
+ "learning_rate": 0.000288343693342466,
+ "loss": 3.954651117324829,
+ "step": 1906
+ },
+ {
+ "epoch": 26.48865619546248,
+ "grad_norm": 0.05564865842461586,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 3.951840400695801,
+ "step": 1907
+ },
+ {
+ "epoch": 26.50261780104712,
+ "grad_norm": 0.05588578060269356,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 3.933762550354004,
+ "step": 1908
+ },
+ {
+ "epoch": 26.516579406631763,
+ "grad_norm": 0.05727043375372887,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 3.9299421310424805,
+ "step": 1909
+ },
+ {
+ "epoch": 26.530541012216403,
+ "grad_norm": 0.05533654987812042,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 3.941525936126709,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544502617801047,
+ "grad_norm": 0.05524713918566704,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 3.935666084289551,
+ "step": 1911
+ },
+ {
+ "epoch": 26.55846422338569,
+ "grad_norm": 0.055435653775930405,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 3.951631546020508,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57242582897033,
+ "grad_norm": 0.05678093433380127,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 3.960881233215332,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586387434554975,
+ "grad_norm": 0.054896436631679535,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 3.9262163639068604,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600349040139616,
+ "grad_norm": 0.055213335901498795,
+ "learning_rate": 0.000285947841605349,
+ "loss": 3.9284324645996094,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61431064572426,
+ "grad_norm": 0.05644496902823448,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 3.9727559089660645,
+ "step": 1916
+ },
+ {
+ "epoch": 26.6282722513089,
+ "grad_norm": 0.05483568087220192,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 3.9410910606384277,
+ "step": 1917
+ },
+ {
+ "epoch": 26.642233856893544,
+ "grad_norm": 0.054636143147945404,
+ "learning_rate": 0.000285149463934261,
+ "loss": 3.945554733276367,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656195462478184,
+ "grad_norm": 0.056062210351228714,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 3.950249671936035,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670157068062828,
+ "grad_norm": 0.05734739080071449,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 3.9527649879455566,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68411867364747,
+ "grad_norm": 0.05562867224216461,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 3.9563350677490234,
+ "step": 1921
+ },
+ {
+ "epoch": 26.698080279232112,
+ "grad_norm": 0.05565498396754265,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 3.9478225708007812,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712041884816752,
+ "grad_norm": 0.057103481143713,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 3.9302189350128174,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726003490401396,
+ "grad_norm": 0.058241210877895355,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 3.9773902893066406,
+ "step": 1924
+ },
+ {
+ "epoch": 26.739965095986037,
+ "grad_norm": 0.055981557816267014,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 3.923896074295044,
+ "step": 1925
+ },
+ {
+ "epoch": 26.75392670157068,
+ "grad_norm": 0.05711016431450844,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 3.972334861755371,
+ "step": 1926
+ },
+ {
+ "epoch": 26.767888307155324,
+ "grad_norm": 0.05707632377743721,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 3.958393096923828,
+ "step": 1927
+ },
+ {
+ "epoch": 26.781849912739965,
+ "grad_norm": 0.055073633790016174,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 3.9449703693389893,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79581151832461,
+ "grad_norm": 0.05765518546104431,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 3.9442358016967773,
+ "step": 1929
+ },
+ {
+ "epoch": 26.80977312390925,
+ "grad_norm": 0.05692309886217117,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 3.9665679931640625,
+ "step": 1930
+ },
+ {
+ "epoch": 26.823734729493893,
+ "grad_norm": 0.05776236951351166,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 3.9153542518615723,
+ "step": 1931
+ },
+ {
+ "epoch": 26.837696335078533,
+ "grad_norm": 0.05465872213244438,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 3.919874429702759,
+ "step": 1932
+ },
+ {
+ "epoch": 26.851657940663177,
+ "grad_norm": 0.057827629148960114,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 3.9578394889831543,
+ "step": 1933
+ },
+ {
+ "epoch": 26.865619546247817,
+ "grad_norm": 0.055902253836393356,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 3.995711326599121,
+ "step": 1934
+ },
+ {
+ "epoch": 26.87958115183246,
+ "grad_norm": 0.05868638679385185,
+ "learning_rate": 0.000280627938758204,
+ "loss": 3.96519136428833,
+ "step": 1935
+ },
+ {
+ "epoch": 26.8935427574171,
+ "grad_norm": 0.059722673147916794,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 3.9566941261291504,
+ "step": 1936
+ },
+ {
+ "epoch": 26.907504363001745,
+ "grad_norm": 0.055209770798683167,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 3.9929540157318115,
+ "step": 1937
+ },
+ {
+ "epoch": 26.921465968586386,
+ "grad_norm": 0.05694293603301048,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 3.9529929161071777,
+ "step": 1938
+ },
+ {
+ "epoch": 26.93542757417103,
+ "grad_norm": 0.058492012321949005,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 3.9465792179107666,
+ "step": 1939
+ },
+ {
+ "epoch": 26.949389179755673,
+ "grad_norm": 0.056036073714494705,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 3.9590697288513184,
+ "step": 1940
+ },
+ {
+ "epoch": 26.963350785340314,
+ "grad_norm": 0.0547766387462616,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 3.969907522201538,
+ "step": 1941
+ },
+ {
+ "epoch": 26.977312390924958,
+ "grad_norm": 0.05592350661754608,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 3.9509897232055664,
+ "step": 1942
+ },
+ {
+ "epoch": 26.991273996509598,
+ "grad_norm": 0.05589321628212929,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 3.9648609161376953,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.03934016078710556,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 2.4784152507781982,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 0.6035652160644531,
+ "eval_runtime": 60.0656,
+ "eval_samples_per_second": 40.656,
+ "eval_steps_per_second": 0.649,
+ "step": 1944
+ },
+ {
+ "epoch": 27.013961605584644,
+ "grad_norm": 0.05773136392235756,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 3.8625826835632324,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027923211169284,
+ "grad_norm": 0.05844702199101448,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 3.8796658515930176,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041884816753928,
+ "grad_norm": 0.05718810483813286,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 3.8808112144470215,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05584642233857,
+ "grad_norm": 0.05264506861567497,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 3.8776121139526367,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069808027923212,
+ "grad_norm": 0.055468373000621796,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 3.866316318511963,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083769633507853,
+ "grad_norm": 0.053565360605716705,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 3.8870849609375,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097731239092496,
+ "grad_norm": 0.05468900501728058,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 3.8581550121307373,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111692844677137,
+ "grad_norm": 0.05454741418361664,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 3.8752541542053223,
+ "step": 1952
+ },
+ {
+ "epoch": 27.12565445026178,
+ "grad_norm": 0.05647914111614227,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 3.8818912506103516,
+ "step": 1953
+ },
+ {
+ "epoch": 27.13961605584642,
+ "grad_norm": 0.05325528234243393,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 3.8562560081481934,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153577661431065,
+ "grad_norm": 0.05581580847501755,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 3.879669189453125,
+ "step": 1955
+ },
+ {
+ "epoch": 27.167539267015705,
+ "grad_norm": 0.058219097554683685,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 3.8530635833740234,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18150087260035,
+ "grad_norm": 0.05613216012716293,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 3.8905630111694336,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195462478184993,
+ "grad_norm": 0.05486270785331726,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 3.8945157527923584,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209424083769633,
+ "grad_norm": 0.05338508263230324,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 3.909775733947754,
+ "step": 1959
+ },
+ {
+ "epoch": 27.223385689354277,
+ "grad_norm": 0.05602312833070755,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 3.8643651008605957,
+ "step": 1960
+ },
+ {
+ "epoch": 27.237347294938917,
+ "grad_norm": 0.05373037979006767,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 3.8915491104125977,
+ "step": 1961
+ },
+ {
+ "epoch": 27.25130890052356,
+ "grad_norm": 0.05805065110325813,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 3.8836984634399414,
+ "step": 1962
+ },
+ {
+ "epoch": 27.2652705061082,
+ "grad_norm": 0.05469847097992897,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 3.8812215328216553,
+ "step": 1963
+ },
+ {
+ "epoch": 27.279232111692846,
+ "grad_norm": 0.05575987696647644,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 3.8887271881103516,
+ "step": 1964
+ },
+ {
+ "epoch": 27.293193717277486,
+ "grad_norm": 0.05577889457345009,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 3.8917694091796875,
+ "step": 1965
+ },
+ {
+ "epoch": 27.30715532286213,
+ "grad_norm": 0.05604076758027077,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 3.881073474884033,
+ "step": 1966
+ },
+ {
+ "epoch": 27.32111692844677,
+ "grad_norm": 0.055674705654382706,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 3.9303488731384277,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335078534031414,
+ "grad_norm": 0.05300940200686455,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 3.871678113937378,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349040139616054,
+ "grad_norm": 0.05647176504135132,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 3.8848423957824707,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363001745200698,
+ "grad_norm": 0.05647560954093933,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 3.931574583053589,
+ "step": 1970
+ },
+ {
+ "epoch": 27.376963350785342,
+ "grad_norm": 0.055338893085718155,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 3.908456563949585,
+ "step": 1971
+ },
+ {
+ "epoch": 27.390924956369982,
+ "grad_norm": 0.05508219450712204,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 3.8933262825012207,
+ "step": 1972
+ },
+ {
+ "epoch": 27.404886561954626,
+ "grad_norm": 0.05787097290158272,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 3.8908863067626953,
+ "step": 1973
+ },
+ {
+ "epoch": 27.418848167539267,
+ "grad_norm": 0.05587272346019745,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 3.9261322021484375,
+ "step": 1974
+ },
+ {
+ "epoch": 27.43280977312391,
+ "grad_norm": 0.05573277547955513,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 3.9129767417907715,
+ "step": 1975
+ },
+ {
+ "epoch": 27.44677137870855,
+ "grad_norm": 0.05628194287419319,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 3.906001091003418,
+ "step": 1976
+ },
+ {
+ "epoch": 27.460732984293195,
+ "grad_norm": 0.055410996079444885,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 3.9163825511932373,
+ "step": 1977
+ },
+ {
+ "epoch": 27.474694589877835,
+ "grad_norm": 0.05579163134098053,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 3.9286627769470215,
+ "step": 1978
+ },
+ {
+ "epoch": 27.48865619546248,
+ "grad_norm": 0.05793057754635811,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 3.8955917358398438,
+ "step": 1979
+ },
+ {
+ "epoch": 27.50261780104712,
+ "grad_norm": 0.05592824146151543,
+ "learning_rate": 0.00026868712586269,
+ "loss": 3.922128677368164,
+ "step": 1980
+ },
+ {
+ "epoch": 27.516579406631763,
+ "grad_norm": 0.05768568441271782,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 3.897780418395996,
+ "step": 1981
+ },
+ {
+ "epoch": 27.530541012216403,
+ "grad_norm": 0.055304404348134995,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 3.886094570159912,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544502617801047,
+ "grad_norm": 0.054745838046073914,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 3.9272961616516113,
+ "step": 1983
+ },
+ {
+ "epoch": 27.55846422338569,
+ "grad_norm": 0.05616277828812599,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 3.8794174194335938,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57242582897033,
+ "grad_norm": 0.05605581775307655,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 3.9165992736816406,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586387434554975,
+ "grad_norm": 0.057485803961753845,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 3.902085542678833,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600349040139616,
+ "grad_norm": 0.0575849711894989,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 3.92405366897583,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61431064572426,
+ "grad_norm": 0.05563105270266533,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 3.8700900077819824,
+ "step": 1988
+ },
+ {
+ "epoch": 27.6282722513089,
+ "grad_norm": 0.05615481734275818,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 3.9526686668395996,
+ "step": 1989
+ },
+ {
+ "epoch": 27.642233856893544,
+ "grad_norm": 0.056746870279312134,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 3.9000678062438965,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656195462478184,
+ "grad_norm": 0.057345934212207794,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 3.910407543182373,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670157068062828,
+ "grad_norm": 0.05434483289718628,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 3.9132578372955322,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68411867364747,
+ "grad_norm": 0.0582718588411808,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 3.9188318252563477,
+ "step": 1993
+ },
+ {
+ "epoch": 27.698080279232112,
+ "grad_norm": 0.05566178634762764,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 3.8843226432800293,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712041884816752,
+ "grad_norm": 0.05550086498260498,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 3.904834032058716,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726003490401396,
+ "grad_norm": 0.05517689883708954,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 3.9082438945770264,
+ "step": 1996
+ },
+ {
+ "epoch": 27.739965095986037,
+ "grad_norm": 0.05598316714167595,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 3.909560203552246,
+ "step": 1997
+ },
+ {
+ "epoch": 27.75392670157068,
+ "grad_norm": 0.054991044104099274,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 3.903012275695801,
+ "step": 1998
+ },
+ {
+ "epoch": 27.767888307155324,
+ "grad_norm": 0.05612454190850258,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 3.9183692932128906,
+ "step": 1999
+ },
+ {
+ "epoch": 27.781849912739965,
+ "grad_norm": 0.055587366223335266,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 3.918454647064209,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79581151832461,
+ "grad_norm": 0.05595824122428894,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 3.895993232727051,
+ "step": 2001
+ },
+ {
+ "epoch": 27.80977312390925,
+ "grad_norm": 0.054730597883462906,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 3.9305458068847656,
+ "step": 2002
+ },
+ {
+ "epoch": 27.823734729493893,
+ "grad_norm": 0.05543222650885582,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 3.9375224113464355,
+ "step": 2003
+ },
+ {
+ "epoch": 27.837696335078533,
+ "grad_norm": 0.055638376623392105,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 3.9242658615112305,
+ "step": 2004
+ },
+ {
+ "epoch": 27.851657940663177,
+ "grad_norm": 0.05635111406445503,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 3.889681339263916,
+ "step": 2005
+ },
+ {
+ "epoch": 27.865619546247817,
+ "grad_norm": 0.056509099900722504,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 3.9191479682922363,
+ "step": 2006
+ },
+ {
+ "epoch": 27.87958115183246,
+ "grad_norm": 0.055455707013607025,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 3.936049222946167,
+ "step": 2007
+ },
+ {
+ "epoch": 27.8935427574171,
+ "grad_norm": 0.05508789047598839,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 3.901210069656372,
+ "step": 2008
+ },
+ {
+ "epoch": 27.907504363001745,
+ "grad_norm": 0.05542607232928276,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 3.9348654747009277,
+ "step": 2009
+ },
+ {
+ "epoch": 27.921465968586386,
+ "grad_norm": 0.05693232640624046,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 3.940847158432007,
+ "step": 2010
+ },
+ {
+ "epoch": 27.93542757417103,
+ "grad_norm": 0.05707051232457161,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 3.9295883178710938,
+ "step": 2011
+ },
+ {
+ "epoch": 27.949389179755673,
+ "grad_norm": 0.05778150260448456,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 3.9317305088043213,
+ "step": 2012
+ },
+ {
+ "epoch": 27.963350785340314,
+ "grad_norm": 0.05821850523352623,
+ "learning_rate": 0.00025996500713765,
+ "loss": 3.9197943210601807,
+ "step": 2013
+ },
+ {
+ "epoch": 27.977312390924958,
+ "grad_norm": 0.05592802166938782,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 3.9471435546875,
+ "step": 2014
+ },
+ {
+ "epoch": 27.991273996509598,
+ "grad_norm": 0.059447627514600754,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 3.9220168590545654,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.04082372784614563,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 2.433501958847046,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 0.6050393581390381,
+ "eval_runtime": 57.4222,
+ "eval_samples_per_second": 42.527,
+ "eval_steps_per_second": 0.679,
+ "step": 2016
+ },
+ {
+ "epoch": 28.013961605584644,
+ "grad_norm": 0.059585411101579666,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 3.8270139694213867,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027923211169284,
+ "grad_norm": 0.0652594342827797,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 3.842228412628174,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041884816753928,
+ "grad_norm": 0.05607042834162712,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 3.8498799800872803,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05584642233857,
+ "grad_norm": 0.05797959491610527,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 3.8537564277648926,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069808027923212,
+ "grad_norm": 0.06011393293738365,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 3.865638256072998,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083769633507853,
+ "grad_norm": 0.05946647375822067,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 3.8406717777252197,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097731239092496,
+ "grad_norm": 0.05657580494880676,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 3.8516461849212646,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111692844677137,
+ "grad_norm": 0.06000044196844101,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 3.859069347381592,
+ "step": 2024
+ },
+ {
+ "epoch": 28.12565445026178,
+ "grad_norm": 0.05713335797190666,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 3.8558874130249023,
+ "step": 2025
+ },
+ {
+ "epoch": 28.13961605584642,
+ "grad_norm": 0.05555511265993118,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 3.882650852203369,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153577661431065,
+ "grad_norm": 0.05772312358021736,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 3.847323179244995,
+ "step": 2027
+ },
+ {
+ "epoch": 28.167539267015705,
+ "grad_norm": 0.056661207228899,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 3.8922581672668457,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18150087260035,
+ "grad_norm": 0.05714283883571625,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 3.849281072616577,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195462478184993,
+ "grad_norm": 0.05582018196582794,
+ "learning_rate": 0.000255486047794226,
+ "loss": 3.8659892082214355,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209424083769633,
+ "grad_norm": 0.05674750730395317,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 3.8591721057891846,
+ "step": 2031
+ },
+ {
+ "epoch": 28.223385689354277,
+ "grad_norm": 0.060421206057071686,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 3.8599629402160645,
+ "step": 2032
+ },
+ {
+ "epoch": 28.237347294938917,
+ "grad_norm": 0.05727452412247658,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 3.8559670448303223,
+ "step": 2033
+ },
+ {
+ "epoch": 28.25130890052356,
+ "grad_norm": 0.05708890035748482,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 3.8643712997436523,
+ "step": 2034
+ },
+ {
+ "epoch": 28.2652705061082,
+ "grad_norm": 0.05768970027565956,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 3.842012405395508,
+ "step": 2035
+ },
+ {
+ "epoch": 28.279232111692846,
+ "grad_norm": 0.05973518267273903,
+ "learning_rate": 0.000253907826333243,
+ "loss": 3.841464042663574,
+ "step": 2036
+ },
+ {
+ "epoch": 28.293193717277486,
+ "grad_norm": 0.057536788284778595,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 3.8655552864074707,
+ "step": 2037
+ },
+ {
+ "epoch": 28.30715532286213,
+ "grad_norm": 0.05864274501800537,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 3.852839469909668,
+ "step": 2038
+ },
+ {
+ "epoch": 28.32111692844677,
+ "grad_norm": 0.056395065039396286,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 3.859243392944336,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335078534031414,
+ "grad_norm": 0.05933082476258278,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 3.843926191329956,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349040139616054,
+ "grad_norm": 0.05862351134419441,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 3.8284354209899902,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363001745200698,
+ "grad_norm": 0.05564470961689949,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 3.8802285194396973,
+ "step": 2042
+ },
+ {
+ "epoch": 28.376963350785342,
+ "grad_norm": 0.06139146164059639,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 3.867100954055786,
+ "step": 2043
+ },
+ {
+ "epoch": 28.390924956369982,
+ "grad_norm": 0.058729205280542374,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 3.8322975635528564,
+ "step": 2044
+ },
+ {
+ "epoch": 28.404886561954626,
+ "grad_norm": 0.05771554261445999,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 3.855440616607666,
+ "step": 2045
+ },
+ {
+ "epoch": 28.418848167539267,
+ "grad_norm": 0.06000557541847229,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 3.8510828018188477,
+ "step": 2046
+ },
+ {
+ "epoch": 28.43280977312391,
+ "grad_norm": 0.05935550108551979,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 3.8572049140930176,
+ "step": 2047
+ },
+ {
+ "epoch": 28.44677137870855,
+ "grad_norm": 0.056745223701000214,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 3.8892972469329834,
+ "step": 2048
+ },
+ {
+ "epoch": 28.460732984293195,
+ "grad_norm": 0.0588800348341465,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 3.865957260131836,
+ "step": 2049
+ },
+ {
+ "epoch": 28.474694589877835,
+ "grad_norm": 0.05756748467683792,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 3.8530519008636475,
+ "step": 2050
+ },
+ {
+ "epoch": 28.48865619546248,
+ "grad_norm": 0.057398825883865356,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 3.9022018909454346,
+ "step": 2051
+ },
+ {
+ "epoch": 28.50261780104712,
+ "grad_norm": 0.05809267610311508,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 3.853647232055664,
+ "step": 2052
+ },
+ {
+ "epoch": 28.516579406631763,
+ "grad_norm": 0.05625335872173309,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 3.890363931655884,
+ "step": 2053
+ },
+ {
+ "epoch": 28.530541012216403,
+ "grad_norm": 0.05852413922548294,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 3.8836288452148438,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544502617801047,
+ "grad_norm": 0.05545232817530632,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 3.902006149291992,
+ "step": 2055
+ },
+ {
+ "epoch": 28.55846422338569,
+ "grad_norm": 0.05776599794626236,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 3.8910727500915527,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57242582897033,
+ "grad_norm": 0.05847984179854393,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 3.890106201171875,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586387434554975,
+ "grad_norm": 0.056724466383457184,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 3.836899995803833,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600349040139616,
+ "grad_norm": 0.05690842866897583,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 3.834292411804199,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61431064572426,
+ "grad_norm": 0.058501504361629486,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 3.891890048980713,
+ "step": 2060
+ },
+ {
+ "epoch": 28.6282722513089,
+ "grad_norm": 0.059110380709171295,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 3.883753538131714,
+ "step": 2061
+ },
+ {
+ "epoch": 28.642233856893544,
+ "grad_norm": 0.0574255995452404,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 3.8829569816589355,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656195462478184,
+ "grad_norm": 0.05842957645654678,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 3.845698118209839,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670157068062828,
+ "grad_norm": 0.05536387115716934,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 3.9008209705352783,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68411867364747,
+ "grad_norm": 0.05769932642579079,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 3.890972137451172,
+ "step": 2065
+ },
+ {
+ "epoch": 28.698080279232112,
+ "grad_norm": 0.058036014437675476,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 3.873745918273926,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712041884816752,
+ "grad_norm": 0.05829216167330742,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 3.879685401916504,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726003490401396,
+ "grad_norm": 0.06169239804148674,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 3.9156579971313477,
+ "step": 2068
+ },
+ {
+ "epoch": 28.739965095986037,
+ "grad_norm": 0.05861479043960571,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 3.876373052597046,
+ "step": 2069
+ },
+ {
+ "epoch": 28.75392670157068,
+ "grad_norm": 0.05878114700317383,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 3.918684959411621,
+ "step": 2070
+ },
+ {
+ "epoch": 28.767888307155324,
+ "grad_norm": 0.059904664754867554,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 3.8893160820007324,
+ "step": 2071
+ },
+ {
+ "epoch": 28.781849912739965,
+ "grad_norm": 0.057385317981243134,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 3.880828619003296,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79581151832461,
+ "grad_norm": 0.05815180763602257,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 3.8638038635253906,
+ "step": 2073
+ },
+ {
+ "epoch": 28.80977312390925,
+ "grad_norm": 0.05871085450053215,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 3.8427436351776123,
+ "step": 2074
+ },
+ {
+ "epoch": 28.823734729493893,
+ "grad_norm": 0.056275319308042526,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 3.858442544937134,
+ "step": 2075
+ },
+ {
+ "epoch": 28.837696335078533,
+ "grad_norm": 0.05850474536418915,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 3.871570110321045,
+ "step": 2076
+ },
+ {
+ "epoch": 28.851657940663177,
+ "grad_norm": 0.0580497644841671,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 3.89687442779541,
+ "step": 2077
+ },
+ {
+ "epoch": 28.865619546247817,
+ "grad_norm": 0.05944157764315605,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 3.8877720832824707,
+ "step": 2078
+ },
+ {
+ "epoch": 28.87958115183246,
+ "grad_norm": 0.05749465152621269,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 3.8787431716918945,
+ "step": 2079
+ },
+ {
+ "epoch": 28.8935427574171,
+ "grad_norm": 0.05645301938056946,
+ "learning_rate": 0.000242380656502223,
+ "loss": 3.9244091510772705,
+ "step": 2080
+ },
+ {
+ "epoch": 28.907504363001745,
+ "grad_norm": 0.05700884759426117,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 3.9192817211151123,
+ "step": 2081
+ },
+ {
+ "epoch": 28.921465968586386,
+ "grad_norm": 0.05823809653520584,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 3.905806064605713,
+ "step": 2082
+ },
+ {
+ "epoch": 28.93542757417103,
+ "grad_norm": 0.05736592411994934,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 3.8907763957977295,
+ "step": 2083
+ },
+ {
+ "epoch": 28.949389179755673,
+ "grad_norm": 0.05646584555506706,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 3.892714500427246,
+ "step": 2084
+ },
+ {
+ "epoch": 28.963350785340314,
+ "grad_norm": 0.06081141158938408,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 3.864286184310913,
+ "step": 2085
+ },
+ {
+ "epoch": 28.977312390924958,
+ "grad_norm": 0.05633515864610672,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 3.8791451454162598,
+ "step": 2086
+ },
+ {
+ "epoch": 28.991273996509598,
+ "grad_norm": 0.05825566127896309,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 3.8895483016967773,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.042954832315444946,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 2.408613681793213,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 0.6065697073936462,
+ "eval_runtime": 59.1276,
+ "eval_samples_per_second": 41.301,
+ "eval_steps_per_second": 0.66,
+ "step": 2088
+ },
+ {
+ "epoch": 29.013961605584644,
+ "grad_norm": 0.06000809744000435,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 3.8419973850250244,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027923211169284,
+ "grad_norm": 0.06233612447977066,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 3.83025860786438,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041884816753928,
+ "grad_norm": 0.0551699623465538,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 3.8283262252807617,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05584642233857,
+ "grad_norm": 0.06225701421499252,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 3.822310447692871,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069808027923212,
+ "grad_norm": 0.05747336149215698,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 3.8208513259887695,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083769633507853,
+ "grad_norm": 0.0582275427877903,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 3.8013901710510254,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097731239092496,
+ "grad_norm": 0.059701140969991684,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 3.829238176345825,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111692844677137,
+ "grad_norm": 0.057550523430109024,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 3.7963826656341553,
+ "step": 2096
+ },
+ {
+ "epoch": 29.12565445026178,
+ "grad_norm": 0.05855116620659828,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 3.824028491973877,
+ "step": 2097
+ },
+ {
+ "epoch": 29.13961605584642,
+ "grad_norm": 0.056533779948949814,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 3.8151779174804688,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153577661431065,
+ "grad_norm": 0.0573289655148983,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 3.7990314960479736,
+ "step": 2099
+ },
+ {
+ "epoch": 29.167539267015705,
+ "grad_norm": 0.060061171650886536,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 3.8221707344055176,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18150087260035,
+ "grad_norm": 0.05851743370294571,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 3.793701648712158,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195462478184993,
+ "grad_norm": 0.059191569685935974,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 3.8194010257720947,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209424083769633,
+ "grad_norm": 0.05996483936905861,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 3.8316292762756348,
+ "step": 2103
+ },
+ {
+ "epoch": 29.223385689354277,
+ "grad_norm": 0.0555657260119915,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 3.821362018585205,
+ "step": 2104
+ },
+ {
+ "epoch": 29.237347294938917,
+ "grad_norm": 0.058476634323596954,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 3.8098394870758057,
+ "step": 2105
+ },
+ {
+ "epoch": 29.25130890052356,
+ "grad_norm": 0.05669668689370155,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 3.843484878540039,
+ "step": 2106
+ },
+ {
+ "epoch": 29.2652705061082,
+ "grad_norm": 0.05648602545261383,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 3.811458110809326,
+ "step": 2107
+ },
+ {
+ "epoch": 29.279232111692846,
+ "grad_norm": 0.05574130266904831,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 3.837676525115967,
+ "step": 2108
+ },
+ {
+ "epoch": 29.293193717277486,
+ "grad_norm": 0.05538785457611084,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 3.810798168182373,
+ "step": 2109
+ },
+ {
+ "epoch": 29.30715532286213,
+ "grad_norm": 0.05477031692862511,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 3.8156371116638184,
+ "step": 2110
+ },
+ {
+ "epoch": 29.32111692844677,
+ "grad_norm": 0.05664678290486336,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 3.8472559452056885,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335078534031414,
+ "grad_norm": 0.05649891495704651,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 3.824376106262207,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349040139616054,
+ "grad_norm": 0.05665554851293564,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 3.7963147163391113,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363001745200698,
+ "grad_norm": 0.054111722856760025,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 3.819735527038574,
+ "step": 2114
+ },
+ {
+ "epoch": 29.376963350785342,
+ "grad_norm": 0.05805648863315582,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 3.790189743041992,
+ "step": 2115
+ },
+ {
+ "epoch": 29.390924956369982,
+ "grad_norm": 0.05634011700749397,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 3.852161169052124,
+ "step": 2116
+ },
+ {
+ "epoch": 29.404886561954626,
+ "grad_norm": 0.0555867962539196,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 3.846884250640869,
+ "step": 2117
+ },
+ {
+ "epoch": 29.418848167539267,
+ "grad_norm": 0.057229746133089066,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 3.8432483673095703,
+ "step": 2118
+ },
+ {
+ "epoch": 29.43280977312391,
+ "grad_norm": 0.05664287880063057,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 3.842709541320801,
+ "step": 2119
+ },
+ {
+ "epoch": 29.44677137870855,
+ "grad_norm": 0.0567023791372776,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 3.844529628753662,
+ "step": 2120
+ },
+ {
+ "epoch": 29.460732984293195,
+ "grad_norm": 0.057870302349328995,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 3.853799343109131,
+ "step": 2121
+ },
+ {
+ "epoch": 29.474694589877835,
+ "grad_norm": 0.05681667849421501,
+ "learning_rate": 0.000231465389734324,
+ "loss": 3.849405288696289,
+ "step": 2122
+ },
+ {
+ "epoch": 29.48865619546248,
+ "grad_norm": 0.05738653615117073,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 3.8083410263061523,
+ "step": 2123
+ },
+ {
+ "epoch": 29.50261780104712,
+ "grad_norm": 0.058900732547044754,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 3.840458869934082,
+ "step": 2124
+ },
+ {
+ "epoch": 29.516579406631763,
+ "grad_norm": 0.05823393166065216,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 3.821415662765503,
+ "step": 2125
+ },
+ {
+ "epoch": 29.530541012216403,
+ "grad_norm": 0.057413019239902496,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 3.8713812828063965,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544502617801047,
+ "grad_norm": 0.05935502052307129,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 3.8488903045654297,
+ "step": 2127
+ },
+ {
+ "epoch": 29.55846422338569,
+ "grad_norm": 0.055285241454839706,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 3.8196067810058594,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57242582897033,
+ "grad_norm": 0.05894080176949501,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 3.8444557189941406,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586387434554975,
+ "grad_norm": 0.05878668278455734,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 3.8731908798217773,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600349040139616,
+ "grad_norm": 0.05855663865804672,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 3.831028938293457,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61431064572426,
+ "grad_norm": 0.057419586926698685,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 3.849435806274414,
+ "step": 2132
+ },
+ {
+ "epoch": 29.6282722513089,
+ "grad_norm": 0.058863408863544464,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 3.8495519161224365,
+ "step": 2133
+ },
+ {
+ "epoch": 29.642233856893544,
+ "grad_norm": 0.06081194058060646,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 3.844421863555908,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656195462478184,
+ "grad_norm": 0.05597590655088425,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 3.8413243293762207,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670157068062828,
+ "grad_norm": 0.058956339955329895,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 3.861703872680664,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68411867364747,
+ "grad_norm": 0.05883771926164627,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 3.8756394386291504,
+ "step": 2137
+ },
+ {
+ "epoch": 29.698080279232112,
+ "grad_norm": 0.06036122143268585,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 3.855403423309326,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712041884816752,
+ "grad_norm": 0.061243657022714615,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 3.861691474914551,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726003490401396,
+ "grad_norm": 0.05917691811919212,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 3.870516777038574,
+ "step": 2140
+ },
+ {
+ "epoch": 29.739965095986037,
+ "grad_norm": 0.06157418712973595,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 3.8292267322540283,
+ "step": 2141
+ },
+ {
+ "epoch": 29.75392670157068,
+ "grad_norm": 0.0605802945792675,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 3.836577892303467,
+ "step": 2142
+ },
+ {
+ "epoch": 29.767888307155324,
+ "grad_norm": 0.05750525742769241,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 3.8472204208374023,
+ "step": 2143
+ },
+ {
+ "epoch": 29.781849912739965,
+ "grad_norm": 0.057982541620731354,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 3.8682217597961426,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79581151832461,
+ "grad_norm": 0.057891640812158585,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 3.8397727012634277,
+ "step": 2145
+ },
+ {
+ "epoch": 29.80977312390925,
+ "grad_norm": 0.058175452053546906,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 3.8578720092773438,
+ "step": 2146
+ },
+ {
+ "epoch": 29.823734729493893,
+ "grad_norm": 0.05686056986451149,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 3.8467812538146973,
+ "step": 2147
+ },
+ {
+ "epoch": 29.837696335078533,
+ "grad_norm": 0.05755440890789032,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 3.844052314758301,
+ "step": 2148
+ },
+ {
+ "epoch": 29.851657940663177,
+ "grad_norm": 0.057706039398908615,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 3.818631172180176,
+ "step": 2149
+ },
+ {
+ "epoch": 29.865619546247817,
+ "grad_norm": 0.058824241161346436,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 3.8732900619506836,
+ "step": 2150
+ },
+ {
+ "epoch": 29.87958115183246,
+ "grad_norm": 0.058583300560712814,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 3.8373632431030273,
+ "step": 2151
+ },
+ {
+ "epoch": 29.8935427574171,
+ "grad_norm": 0.057403020560741425,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 3.8761391639709473,
+ "step": 2152
+ },
+ {
+ "epoch": 29.907504363001745,
+ "grad_norm": 0.0586736761033535,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 3.8518989086151123,
+ "step": 2153
+ },
+ {
+ "epoch": 29.921465968586386,
+ "grad_norm": 0.05727249011397362,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 3.839515209197998,
+ "step": 2154
+ },
+ {
+ "epoch": 29.93542757417103,
+ "grad_norm": 0.0594087690114975,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 3.834868907928467,
+ "step": 2155
+ },
+ {
+ "epoch": 29.949389179755673,
+ "grad_norm": 0.057742297649383545,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 3.861978530883789,
+ "step": 2156
+ },
+ {
+ "epoch": 29.963350785340314,
+ "grad_norm": 0.05707386136054993,
+ "learning_rate": 0.00022244633283095,
+ "loss": 3.886831760406494,
+ "step": 2157
+ },
+ {
+ "epoch": 29.977312390924958,
+ "grad_norm": 0.05631854757666588,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 3.83940052986145,
+ "step": 2158
+ },
+ {
+ "epoch": 29.991273996509598,
+ "grad_norm": 0.057860348373651505,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 3.8477416038513184,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.04080710560083389,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 2.4169907569885254,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 0.6078308820724487,
+ "eval_runtime": 59.4599,
+ "eval_samples_per_second": 41.07,
+ "eval_steps_per_second": 0.656,
+ "step": 2160
+ },
+ {
+ "epoch": 30.013961605584644,
+ "grad_norm": 0.05596645176410675,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 3.798748016357422,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027923211169284,
+ "grad_norm": 0.054763782769441605,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 3.7787699699401855,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041884816753928,
+ "grad_norm": 0.053927499800920486,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 3.779787063598633,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05584642233857,
+ "grad_norm": 0.05513143539428711,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 3.782046318054199,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069808027923212,
+ "grad_norm": 0.055234167724847794,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 3.8064064979553223,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083769633507853,
+ "grad_norm": 0.05320988595485687,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 3.8084359169006348,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097731239092496,
+ "grad_norm": 0.05671925097703934,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 3.7805888652801514,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111692844677137,
+ "grad_norm": 0.05502697452902794,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 3.791616916656494,
+ "step": 2168
+ },
+ {
+ "epoch": 30.12565445026178,
+ "grad_norm": 0.055426280945539474,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 3.7947494983673096,
+ "step": 2169
+ },
+ {
+ "epoch": 30.13961605584642,
+ "grad_norm": 0.05564495921134949,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 3.7958974838256836,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153577661431065,
+ "grad_norm": 0.055089402943849564,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 3.783299446105957,
+ "step": 2171
+ },
+ {
+ "epoch": 30.167539267015705,
+ "grad_norm": 0.055240437388420105,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 3.769829750061035,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18150087260035,
+ "grad_norm": 0.055886462330818176,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 3.805971622467041,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195462478184993,
+ "grad_norm": 0.05560130998492241,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 3.801429271697998,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209424083769633,
+ "grad_norm": 0.054780587553977966,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 3.794417381286621,
+ "step": 2175
+ },
+ {
+ "epoch": 30.223385689354277,
+ "grad_norm": 0.05751654878258705,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 3.7895469665527344,
+ "step": 2176
+ },
+ {
+ "epoch": 30.237347294938917,
+ "grad_norm": 0.054392315447330475,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 3.786479949951172,
+ "step": 2177
+ },
+ {
+ "epoch": 30.25130890052356,
+ "grad_norm": 0.057609833776950836,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 3.809680700302124,
+ "step": 2178
+ },
+ {
+ "epoch": 30.2652705061082,
+ "grad_norm": 0.054882537573575974,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 3.7928991317749023,
+ "step": 2179
+ },
+ {
+ "epoch": 30.279232111692846,
+ "grad_norm": 0.056957781314849854,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 3.790977716445923,
+ "step": 2180
+ },
+ {
+ "epoch": 30.293193717277486,
+ "grad_norm": 0.0580085888504982,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 3.851022958755493,
+ "step": 2181
+ },
+ {
+ "epoch": 30.30715532286213,
+ "grad_norm": 0.055905163288116455,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 3.784822463989258,
+ "step": 2182
+ },
+ {
+ "epoch": 30.32111692844677,
+ "grad_norm": 0.05998636782169342,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 3.811324119567871,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335078534031414,
+ "grad_norm": 0.057061366736888885,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 3.8131840229034424,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349040139616054,
+ "grad_norm": 0.05680330842733383,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 3.806260824203491,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363001745200698,
+ "grad_norm": 0.05539876967668533,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 3.793149471282959,
+ "step": 2186
+ },
+ {
+ "epoch": 30.376963350785342,
+ "grad_norm": 0.05771343782544136,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 3.7772932052612305,
+ "step": 2187
+ },
+ {
+ "epoch": 30.390924956369982,
+ "grad_norm": 0.055599041283130646,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 3.7704858779907227,
+ "step": 2188
+ },
+ {
+ "epoch": 30.404886561954626,
+ "grad_norm": 0.05746382102370262,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 3.8296337127685547,
+ "step": 2189
+ },
+ {
+ "epoch": 30.418848167539267,
+ "grad_norm": 0.05619506537914276,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 3.8134541511535645,
+ "step": 2190
+ },
+ {
+ "epoch": 30.43280977312391,
+ "grad_norm": 0.05699177086353302,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 3.8070058822631836,
+ "step": 2191
+ },
+ {
+ "epoch": 30.44677137870855,
+ "grad_norm": 0.056783925741910934,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 3.7931368350982666,
+ "step": 2192
+ },
+ {
+ "epoch": 30.460732984293195,
+ "grad_norm": 0.055537473410367966,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 3.809480905532837,
+ "step": 2193
+ },
+ {
+ "epoch": 30.474694589877835,
+ "grad_norm": 0.057826802134513855,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 3.8422718048095703,
+ "step": 2194
+ },
+ {
+ "epoch": 30.48865619546248,
+ "grad_norm": 0.0562913753092289,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 3.803966760635376,
+ "step": 2195
+ },
+ {
+ "epoch": 30.50261780104712,
+ "grad_norm": 0.05720103159546852,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 3.767852306365967,
+ "step": 2196
+ },
+ {
+ "epoch": 30.516579406631763,
+ "grad_norm": 0.05813048407435417,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 3.8423361778259277,
+ "step": 2197
+ },
+ {
+ "epoch": 30.530541012216403,
+ "grad_norm": 0.057173099368810654,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 3.8059489727020264,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544502617801047,
+ "grad_norm": 0.059024229645729065,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 3.7705254554748535,
+ "step": 2199
+ },
+ {
+ "epoch": 30.55846422338569,
+ "grad_norm": 0.056807395070791245,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 3.8056063652038574,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57242582897033,
+ "grad_norm": 0.05868417024612427,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 3.829719066619873,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586387434554975,
+ "grad_norm": 0.06023262441158295,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 3.8277781009674072,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600349040139616,
+ "grad_norm": 0.057253073900938034,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 3.8088555335998535,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61431064572426,
+ "grad_norm": 0.05946561321616173,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 3.7907581329345703,
+ "step": 2204
+ },
+ {
+ "epoch": 30.6282722513089,
+ "grad_norm": 0.05909013748168945,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 3.812042236328125,
+ "step": 2205
+ },
+ {
+ "epoch": 30.642233856893544,
+ "grad_norm": 0.05843796953558922,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 3.814551830291748,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656195462478184,
+ "grad_norm": 0.05953318253159523,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 3.8058853149414062,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670157068062828,
+ "grad_norm": 0.0616983063519001,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 3.7986061573028564,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68411867364747,
+ "grad_norm": 0.0588211789727211,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 3.8080081939697266,
+ "step": 2209
+ },
+ {
+ "epoch": 30.698080279232112,
+ "grad_norm": 0.05937618389725685,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 3.7851452827453613,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712041884816752,
+ "grad_norm": 0.05869529768824577,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 3.7893905639648438,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726003490401396,
+ "grad_norm": 0.060255810618400574,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 3.7998719215393066,
+ "step": 2212
+ },
+ {
+ "epoch": 30.739965095986037,
+ "grad_norm": 0.05810222774744034,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 3.800131320953369,
+ "step": 2213
+ },
+ {
+ "epoch": 30.75392670157068,
+ "grad_norm": 0.057548508048057556,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 3.7940425872802734,
+ "step": 2214
+ },
+ {
+ "epoch": 30.767888307155324,
+ "grad_norm": 0.05792916938662529,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 3.8089685440063477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.781849912739965,
+ "grad_norm": 0.06035939231514931,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 3.8355417251586914,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79581151832461,
+ "grad_norm": 0.058805160224437714,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 3.8146378993988037,
+ "step": 2217
+ },
+ {
+ "epoch": 30.80977312390925,
+ "grad_norm": 0.05828496813774109,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 3.8210601806640625,
+ "step": 2218
+ },
+ {
+ "epoch": 30.823734729493893,
+ "grad_norm": 0.059257738292217255,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 3.807644844055176,
+ "step": 2219
+ },
+ {
+ "epoch": 30.837696335078533,
+ "grad_norm": 0.05937729775905609,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 3.8312151432037354,
+ "step": 2220
+ },
+ {
+ "epoch": 30.851657940663177,
+ "grad_norm": 0.058580249547958374,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 3.7979660034179688,
+ "step": 2221
+ },
+ {
+ "epoch": 30.865619546247817,
+ "grad_norm": 0.0585496611893177,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 3.827874183654785,
+ "step": 2222
+ },
+ {
+ "epoch": 30.87958115183246,
+ "grad_norm": 0.05949759483337402,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 3.808143138885498,
+ "step": 2223
+ },
+ {
+ "epoch": 30.8935427574171,
+ "grad_norm": 0.058828581124544144,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 3.8351407051086426,
+ "step": 2224
+ },
+ {
+ "epoch": 30.907504363001745,
+ "grad_norm": 0.0600498728454113,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 3.8025779724121094,
+ "step": 2225
+ },
+ {
+ "epoch": 30.921465968586386,
+ "grad_norm": 0.06004577502608299,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 3.8157856464385986,
+ "step": 2226
+ },
+ {
+ "epoch": 30.93542757417103,
+ "grad_norm": 0.057676251977682114,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 3.8136918544769287,
+ "step": 2227
+ },
+ {
+ "epoch": 30.949389179755673,
+ "grad_norm": 0.0627770945429802,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 3.8470370769500732,
+ "step": 2228
+ },
+ {
+ "epoch": 30.963350785340314,
+ "grad_norm": 0.058836981654167175,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 3.8422586917877197,
+ "step": 2229
+ },
+ {
+ "epoch": 30.977312390924958,
+ "grad_norm": 0.060720525681972504,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 3.7992711067199707,
+ "step": 2230
+ },
+ {
+ "epoch": 30.991273996509598,
+ "grad_norm": 0.05844535678625107,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 3.823277473449707,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.0424773246049881,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 2.3872625827789307,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 0.6090936064720154,
+ "eval_runtime": 58.7367,
+ "eval_samples_per_second": 41.575,
+ "eval_steps_per_second": 0.664,
+ "step": 2232
+ },
+ {
+ "epoch": 31.013961605584644,
+ "grad_norm": 0.057239554822444916,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 3.764132022857666,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027923211169284,
+ "grad_norm": 0.05666050314903259,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 3.767040729522705,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041884816753928,
+ "grad_norm": 0.054038409143686295,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 3.7878732681274414,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05584642233857,
+ "grad_norm": 0.056939832866191864,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 3.7827179431915283,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069808027923212,
+ "grad_norm": 0.05488067492842674,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 3.7544898986816406,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083769633507853,
+ "grad_norm": 0.05604229122400284,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 3.7706053256988525,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097731239092496,
+ "grad_norm": 0.0550883486866951,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 3.755401611328125,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111692844677137,
+ "grad_norm": 0.055959347635507584,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 3.741966724395752,
+ "step": 2240
+ },
+ {
+ "epoch": 31.12565445026178,
+ "grad_norm": 0.05733809620141983,
+ "learning_rate": 0.000201141724176723,
+ "loss": 3.7537970542907715,
+ "step": 2241
+ },
+ {
+ "epoch": 31.13961605584642,
+ "grad_norm": 0.0551145002245903,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 3.763530731201172,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153577661431065,
+ "grad_norm": 0.055499665439128876,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 3.747159481048584,
+ "step": 2243
+ },
+ {
+ "epoch": 31.167539267015705,
+ "grad_norm": 0.0553385354578495,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 3.769044876098633,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18150087260035,
+ "grad_norm": 0.05678795278072357,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 3.7528388500213623,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195462478184993,
+ "grad_norm": 0.055931974202394485,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 3.742429733276367,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209424083769633,
+ "grad_norm": 0.05553249269723892,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 3.7760162353515625,
+ "step": 2247
+ },
+ {
+ "epoch": 31.223385689354277,
+ "grad_norm": 0.05742773041129112,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 3.7682301998138428,
+ "step": 2248
+ },
+ {
+ "epoch": 31.237347294938917,
+ "grad_norm": 0.056315284222364426,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 3.755894660949707,
+ "step": 2249
+ },
+ {
+ "epoch": 31.25130890052356,
+ "grad_norm": 0.05695716291666031,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 3.733503580093384,
+ "step": 2250
+ },
+ {
+ "epoch": 31.2652705061082,
+ "grad_norm": 0.05714774876832962,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 3.7504918575286865,
+ "step": 2251
+ },
+ {
+ "epoch": 31.279232111692846,
+ "grad_norm": 0.058987151831388474,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 3.7579331398010254,
+ "step": 2252
+ },
+ {
+ "epoch": 31.293193717277486,
+ "grad_norm": 0.05853817239403725,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 3.781409740447998,
+ "step": 2253
+ },
+ {
+ "epoch": 31.30715532286213,
+ "grad_norm": 0.057014722377061844,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 3.763247013092041,
+ "step": 2254
+ },
+ {
+ "epoch": 31.32111692844677,
+ "grad_norm": 0.05816958099603653,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 3.7801408767700195,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335078534031414,
+ "grad_norm": 0.05724027752876282,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 3.7466397285461426,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349040139616054,
+ "grad_norm": 0.055836353451013565,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 3.7770347595214844,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363001745200698,
+ "grad_norm": 0.05831955373287201,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 3.7769415378570557,
+ "step": 2258
+ },
+ {
+ "epoch": 31.376963350785342,
+ "grad_norm": 0.056602735072374344,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 3.781635284423828,
+ "step": 2259
+ },
+ {
+ "epoch": 31.390924956369982,
+ "grad_norm": 0.05784839391708374,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 3.7605199813842773,
+ "step": 2260
+ },
+ {
+ "epoch": 31.404886561954626,
+ "grad_norm": 0.057923249900341034,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 3.7671380043029785,
+ "step": 2261
+ },
+ {
+ "epoch": 31.418848167539267,
+ "grad_norm": 0.05663329362869263,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 3.7952239513397217,
+ "step": 2262
+ },
+ {
+ "epoch": 31.43280977312391,
+ "grad_norm": 0.057000353932380676,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 3.7924318313598633,
+ "step": 2263
+ },
+ {
+ "epoch": 31.44677137870855,
+ "grad_norm": 0.059976726770401,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 3.7782416343688965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.460732984293195,
+ "grad_norm": 0.05681120976805687,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 3.7918577194213867,
+ "step": 2265
+ },
+ {
+ "epoch": 31.474694589877835,
+ "grad_norm": 0.05824654549360275,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 3.7725377082824707,
+ "step": 2266
+ },
+ {
+ "epoch": 31.48865619546248,
+ "grad_norm": 0.058513302356004715,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 3.769562244415283,
+ "step": 2267
+ },
+ {
+ "epoch": 31.50261780104712,
+ "grad_norm": 0.05720348283648491,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 3.773402214050293,
+ "step": 2268
+ },
+ {
+ "epoch": 31.516579406631763,
+ "grad_norm": 0.06029265746474266,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 3.7562055587768555,
+ "step": 2269
+ },
+ {
+ "epoch": 31.530541012216403,
+ "grad_norm": 0.05892042815685272,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 3.779987335205078,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544502617801047,
+ "grad_norm": 0.06146431341767311,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 3.749899387359619,
+ "step": 2271
+ },
+ {
+ "epoch": 31.55846422338569,
+ "grad_norm": 0.056403227150440216,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 3.763558864593506,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57242582897033,
+ "grad_norm": 0.058624617755413055,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 3.7811503410339355,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586387434554975,
+ "grad_norm": 0.05978401005268097,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 3.7820377349853516,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600349040139616,
+ "grad_norm": 0.05971162021160126,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 3.7685751914978027,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61431064572426,
+ "grad_norm": 0.05576494708657265,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 3.7588582038879395,
+ "step": 2276
+ },
+ {
+ "epoch": 31.6282722513089,
+ "grad_norm": 0.05953748896718025,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 3.7525811195373535,
+ "step": 2277
+ },
+ {
+ "epoch": 31.642233856893544,
+ "grad_norm": 0.05771994963288307,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 3.7891383171081543,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656195462478184,
+ "grad_norm": 0.05822686105966568,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 3.7933766841888428,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670157068062828,
+ "grad_norm": 0.06044064089655876,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 3.783848762512207,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68411867364747,
+ "grad_norm": 0.05953317880630493,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 3.775938034057617,
+ "step": 2281
+ },
+ {
+ "epoch": 31.698080279232112,
+ "grad_norm": 0.05928120017051697,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 3.7867183685302734,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712041884816752,
+ "grad_norm": 0.05855962261557579,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 3.7872838973999023,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726003490401396,
+ "grad_norm": 0.058542776852846146,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 3.8022043704986572,
+ "step": 2284
+ },
+ {
+ "epoch": 31.739965095986037,
+ "grad_norm": 0.059078946709632874,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 3.79337739944458,
+ "step": 2285
+ },
+ {
+ "epoch": 31.75392670157068,
+ "grad_norm": 0.05854807421565056,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 3.769434928894043,
+ "step": 2286
+ },
+ {
+ "epoch": 31.767888307155324,
+ "grad_norm": 0.058214519172906876,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 3.778744697570801,
+ "step": 2287
+ },
+ {
+ "epoch": 31.781849912739965,
+ "grad_norm": 0.05981272831559181,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 3.8015875816345215,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79581151832461,
+ "grad_norm": 0.05833584442734718,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 3.7718634605407715,
+ "step": 2289
+ },
+ {
+ "epoch": 31.80977312390925,
+ "grad_norm": 0.058713216334581375,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 3.759026050567627,
+ "step": 2290
+ },
+ {
+ "epoch": 31.823734729493893,
+ "grad_norm": 0.06085464730858803,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 3.762263059616089,
+ "step": 2291
+ },
+ {
+ "epoch": 31.837696335078533,
+ "grad_norm": 0.05838872492313385,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 3.7770943641662598,
+ "step": 2292
+ },
+ {
+ "epoch": 31.851657940663177,
+ "grad_norm": 0.05941537022590637,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 3.8041505813598633,
+ "step": 2293
+ },
+ {
+ "epoch": 31.865619546247817,
+ "grad_norm": 0.058818284422159195,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 3.7724623680114746,
+ "step": 2294
+ },
+ {
+ "epoch": 31.87958115183246,
+ "grad_norm": 0.05925480276346207,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 3.7881040573120117,
+ "step": 2295
+ },
+ {
+ "epoch": 31.8935427574171,
+ "grad_norm": 0.058119483292102814,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 3.7992963790893555,
+ "step": 2296
+ },
+ {
+ "epoch": 31.907504363001745,
+ "grad_norm": 0.06139586120843887,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 3.800816297531128,
+ "step": 2297
+ },
+ {
+ "epoch": 31.921465968586386,
+ "grad_norm": 0.05812571942806244,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 3.7683157920837402,
+ "step": 2298
+ },
+ {
+ "epoch": 31.93542757417103,
+ "grad_norm": 0.05929575115442276,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 3.808281660079956,
+ "step": 2299
+ },
+ {
+ "epoch": 31.949389179755673,
+ "grad_norm": 0.057833533734083176,
+ "learning_rate": 0.000186516746349841,
+ "loss": 3.821108818054199,
+ "step": 2300
+ },
+ {
+ "epoch": 31.963350785340314,
+ "grad_norm": 0.05857503414154053,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 3.8082172870635986,
+ "step": 2301
+ },
+ {
+ "epoch": 31.977312390924958,
+ "grad_norm": 0.05893079936504364,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 3.802304267883301,
+ "step": 2302
+ },
+ {
+ "epoch": 31.991273996509598,
+ "grad_norm": 0.057888105511665344,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 3.7589073181152344,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.04306843504309654,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 2.359179735183716,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 0.6101394891738892,
+ "eval_runtime": 59.2767,
+ "eval_samples_per_second": 41.197,
+ "eval_steps_per_second": 0.658,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01396160558464,
+ "grad_norm": 0.059457872062921524,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 3.746081829071045,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02792321116929,
+ "grad_norm": 0.05826259404420853,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 3.7450389862060547,
+ "step": 2306
+ },
+ {
+ "epoch": 32.04188481675393,
+ "grad_norm": 0.05683758482336998,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 3.718315601348877,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05584642233857,
+ "grad_norm": 0.05709684267640114,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 3.7391161918640137,
+ "step": 2308
+ },
+ {
+ "epoch": 32.06980802792321,
+ "grad_norm": 0.0563763789832592,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 3.714268684387207,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083769633507856,
+ "grad_norm": 0.058482736349105835,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 3.7212839126586914,
+ "step": 2310
+ },
+ {
+ "epoch": 32.0977312390925,
+ "grad_norm": 0.05536908283829689,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 3.7573728561401367,
+ "step": 2311
+ },
+ {
+ "epoch": 32.11169284467714,
+ "grad_norm": 0.05950985103845596,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 3.715609312057495,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12565445026178,
+ "grad_norm": 0.05600661039352417,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 3.7144830226898193,
+ "step": 2313
+ },
+ {
+ "epoch": 32.139616055846425,
+ "grad_norm": 0.05814438313245773,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 3.7489242553710938,
+ "step": 2314
+ },
+ {
+ "epoch": 32.153577661431065,
+ "grad_norm": 0.0568280853331089,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 3.729698657989502,
+ "step": 2315
+ },
+ {
+ "epoch": 32.167539267015705,
+ "grad_norm": 0.05647209286689758,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 3.7484517097473145,
+ "step": 2316
+ },
+ {
+ "epoch": 32.181500872600346,
+ "grad_norm": 0.05811113119125366,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 3.7534449100494385,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19546247818499,
+ "grad_norm": 0.055132538080215454,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 3.695220470428467,
+ "step": 2318
+ },
+ {
+ "epoch": 32.20942408376963,
+ "grad_norm": 0.05722372978925705,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 3.7204041481018066,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223385689354274,
+ "grad_norm": 0.05533589795231819,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 3.7357606887817383,
+ "step": 2320
+ },
+ {
+ "epoch": 32.23734729493892,
+ "grad_norm": 0.056150201708078384,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 3.742290735244751,
+ "step": 2321
+ },
+ {
+ "epoch": 32.25130890052356,
+ "grad_norm": 0.05737832188606262,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 3.7484631538391113,
+ "step": 2322
+ },
+ {
+ "epoch": 32.2652705061082,
+ "grad_norm": 0.056017421185970306,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 3.744821071624756,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27923211169284,
+ "grad_norm": 0.05658293142914772,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 3.717287063598633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29319371727749,
+ "grad_norm": 0.05632270127534866,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 3.7172422409057617,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30715532286213,
+ "grad_norm": 0.05777734890580177,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 3.7551779747009277,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32111692844677,
+ "grad_norm": 0.05593368038535118,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 3.7148218154907227,
+ "step": 2327
+ },
+ {
+ "epoch": 32.33507853403141,
+ "grad_norm": 0.05844123661518097,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 3.7418324947357178,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34904013961606,
+ "grad_norm": 0.05616185814142227,
+ "learning_rate": 0.000179445406945268,
+ "loss": 3.7213454246520996,
+ "step": 2329
+ },
+ {
+ "epoch": 32.3630017452007,
+ "grad_norm": 0.058453164994716644,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 3.730037212371826,
+ "step": 2330
+ },
+ {
+ "epoch": 32.37696335078534,
+ "grad_norm": 0.05819960683584213,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 3.7571628093719482,
+ "step": 2331
+ },
+ {
+ "epoch": 32.390924956369986,
+ "grad_norm": 0.05889322608709335,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 3.7306394577026367,
+ "step": 2332
+ },
+ {
+ "epoch": 32.404886561954626,
+ "grad_norm": 0.05727435275912285,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 3.7304887771606445,
+ "step": 2333
+ },
+ {
+ "epoch": 32.41884816753927,
+ "grad_norm": 0.058260343968868256,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 3.714399814605713,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43280977312391,
+ "grad_norm": 0.06092117354273796,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 3.743901491165161,
+ "step": 2335
+ },
+ {
+ "epoch": 32.446771378708554,
+ "grad_norm": 0.05715249851346016,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 3.747819423675537,
+ "step": 2336
+ },
+ {
+ "epoch": 32.460732984293195,
+ "grad_norm": 0.060237716883420944,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 3.770688533782959,
+ "step": 2337
+ },
+ {
+ "epoch": 32.474694589877835,
+ "grad_norm": 0.059161923825740814,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 3.724484443664551,
+ "step": 2338
+ },
+ {
+ "epoch": 32.488656195462475,
+ "grad_norm": 0.05693944916129112,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 3.7228128910064697,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50261780104712,
+ "grad_norm": 0.05745118111371994,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 3.7756295204162598,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51657940663176,
+ "grad_norm": 0.05782153829932213,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 3.757032871246338,
+ "step": 2341
+ },
+ {
+ "epoch": 32.5305410122164,
+ "grad_norm": 0.058188699185848236,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 3.7462215423583984,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544502617801044,
+ "grad_norm": 0.05856512114405632,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 3.748447895050049,
+ "step": 2343
+ },
+ {
+ "epoch": 32.55846422338569,
+ "grad_norm": 0.05581355467438698,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 3.726031541824341,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57242582897033,
+ "grad_norm": 0.05866248160600662,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 3.7235984802246094,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58638743455497,
+ "grad_norm": 0.05732253938913345,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 3.7372875213623047,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60034904013962,
+ "grad_norm": 0.057102739810943604,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 3.768643617630005,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61431064572426,
+ "grad_norm": 0.05914342775940895,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 3.748710870742798,
+ "step": 2348
+ },
+ {
+ "epoch": 32.6282722513089,
+ "grad_norm": 0.05618954077363014,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 3.7640533447265625,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64223385689354,
+ "grad_norm": 0.05810703709721565,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 3.7230985164642334,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65619546247819,
+ "grad_norm": 0.05668202042579651,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 3.736083984375,
+ "step": 2351
+ },
+ {
+ "epoch": 32.67015706806283,
+ "grad_norm": 0.05779054015874863,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 3.7652668952941895,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68411867364747,
+ "grad_norm": 0.05945349857211113,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 3.75258207321167,
+ "step": 2353
+ },
+ {
+ "epoch": 32.69808027923211,
+ "grad_norm": 0.05697475001215935,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 3.7300591468811035,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712041884816756,
+ "grad_norm": 0.05820303037762642,
+ "learning_rate": 0.000173176617304673,
+ "loss": 3.7426252365112305,
+ "step": 2355
+ },
+ {
+ "epoch": 32.726003490401396,
+ "grad_norm": 0.05830353870987892,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 3.7716498374938965,
+ "step": 2356
+ },
+ {
+ "epoch": 32.73996509598604,
+ "grad_norm": 0.05690579116344452,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 3.7565040588378906,
+ "step": 2357
+ },
+ {
+ "epoch": 32.753926701570684,
+ "grad_norm": 0.059615254402160645,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 3.7484030723571777,
+ "step": 2358
+ },
+ {
+ "epoch": 32.767888307155324,
+ "grad_norm": 0.058043934404850006,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 3.76991868019104,
+ "step": 2359
+ },
+ {
+ "epoch": 32.781849912739965,
+ "grad_norm": 0.059192877262830734,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 3.7510385513305664,
+ "step": 2360
+ },
+ {
+ "epoch": 32.795811518324605,
+ "grad_norm": 0.05763391777873039,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 3.7479896545410156,
+ "step": 2361
+ },
+ {
+ "epoch": 32.80977312390925,
+ "grad_norm": 0.05946139991283417,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 3.7780566215515137,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82373472949389,
+ "grad_norm": 0.05837136134505272,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 3.7610299587249756,
+ "step": 2363
+ },
+ {
+ "epoch": 32.83769633507853,
+ "grad_norm": 0.05832294747233391,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 3.7216475009918213,
+ "step": 2364
+ },
+ {
+ "epoch": 32.85165794066317,
+ "grad_norm": 0.05887595936655998,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 3.767125129699707,
+ "step": 2365
+ },
+ {
+ "epoch": 32.86561954624782,
+ "grad_norm": 0.058594878762960434,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 3.751053810119629,
+ "step": 2366
+ },
+ {
+ "epoch": 32.87958115183246,
+ "grad_norm": 0.05895577743649483,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 3.7889444828033447,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8935427574171,
+ "grad_norm": 0.05831810459494591,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 3.7719945907592773,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90750436300174,
+ "grad_norm": 0.06090841069817543,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 3.7372164726257324,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92146596858639,
+ "grad_norm": 0.05728885903954506,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 3.782083511352539,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93542757417103,
+ "grad_norm": 0.061425331979990005,
+ "learning_rate": 0.00016935382741164,
+ "loss": 3.7716455459594727,
+ "step": 2371
+ },
+ {
+ "epoch": 32.94938917975567,
+ "grad_norm": 0.05963108688592911,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 3.7798497676849365,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96335078534032,
+ "grad_norm": 0.05818144232034683,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 3.746483325958252,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97731239092496,
+ "grad_norm": 0.05970250442624092,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 3.7578821182250977,
+ "step": 2374
+ },
+ {
+ "epoch": 32.9912739965096,
+ "grad_norm": 0.05898123234510422,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 3.762807846069336,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.042496208101511,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 2.353346347808838,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 0.611886203289032,
+ "eval_runtime": 58.794,
+ "eval_samples_per_second": 41.535,
+ "eval_steps_per_second": 0.663,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01396160558464,
+ "grad_norm": 0.05598964914679527,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 3.7107529640197754,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02792321116929,
+ "grad_norm": 0.057739973068237305,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 3.6825904846191406,
+ "step": 2378
+ },
+ {
+ "epoch": 33.04188481675393,
+ "grad_norm": 0.05335357040166855,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 3.684080123901367,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05584642233857,
+ "grad_norm": 0.05573783814907074,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 3.69419527053833,
+ "step": 2380
+ },
+ {
+ "epoch": 33.06980802792321,
+ "grad_norm": 0.055793728679418564,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 3.730604648590088,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083769633507856,
+ "grad_norm": 0.054736193269491196,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 3.6922354698181152,
+ "step": 2382
+ },
+ {
+ "epoch": 33.0977312390925,
+ "grad_norm": 0.05806516483426094,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 3.7343029975891113,
+ "step": 2383
+ },
+ {
+ "epoch": 33.11169284467714,
+ "grad_norm": 0.0541117899119854,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 3.7044200897216797,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12565445026178,
+ "grad_norm": 0.057081446051597595,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 3.6954641342163086,
+ "step": 2385
+ },
+ {
+ "epoch": 33.139616055846425,
+ "grad_norm": 0.05561508610844612,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 3.709800958633423,
+ "step": 2386
+ },
+ {
+ "epoch": 33.153577661431065,
+ "grad_norm": 0.05675433203577995,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 3.7057042121887207,
+ "step": 2387
+ },
+ {
+ "epoch": 33.167539267015705,
+ "grad_norm": 0.05598882585763931,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 3.7071051597595215,
+ "step": 2388
+ },
+ {
+ "epoch": 33.181500872600346,
+ "grad_norm": 0.05500704050064087,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 3.7088711261749268,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19546247818499,
+ "grad_norm": 0.057210177183151245,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 3.685708522796631,
+ "step": 2390
+ },
+ {
+ "epoch": 33.20942408376963,
+ "grad_norm": 0.05615527555346489,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 3.695154905319214,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223385689354274,
+ "grad_norm": 0.056796107441186905,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 3.7289557456970215,
+ "step": 2392
+ },
+ {
+ "epoch": 33.23734729493892,
+ "grad_norm": 0.056371185928583145,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 3.7177929878234863,
+ "step": 2393
+ },
+ {
+ "epoch": 33.25130890052356,
+ "grad_norm": 0.05646727234125137,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 3.7265491485595703,
+ "step": 2394
+ },
+ {
+ "epoch": 33.2652705061082,
+ "grad_norm": 0.05717485025525093,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 3.7198410034179688,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27923211169284,
+ "grad_norm": 0.056603338569402695,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 3.692781686782837,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29319371727749,
+ "grad_norm": 0.05612906441092491,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 3.7237164974212646,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30715532286213,
+ "grad_norm": 0.05704745277762413,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 3.698744058609009,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32111692844677,
+ "grad_norm": 0.055712319910526276,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 3.7240562438964844,
+ "step": 2399
+ },
+ {
+ "epoch": 33.33507853403141,
+ "grad_norm": 0.05763087794184685,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 3.709852695465088,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34904013961606,
+ "grad_norm": 0.05521798133850098,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 3.7102928161621094,
+ "step": 2401
+ },
+ {
+ "epoch": 33.3630017452007,
+ "grad_norm": 0.05802507698535919,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 3.697176933288574,
+ "step": 2402
+ },
+ {
+ "epoch": 33.37696335078534,
+ "grad_norm": 0.05547751113772392,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 3.7416112422943115,
+ "step": 2403
+ },
+ {
+ "epoch": 33.390924956369986,
+ "grad_norm": 0.05753115564584732,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 3.7048134803771973,
+ "step": 2404
+ },
+ {
+ "epoch": 33.404886561954626,
+ "grad_norm": 0.05642146244645119,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 3.690187931060791,
+ "step": 2405
+ },
+ {
+ "epoch": 33.41884816753927,
+ "grad_norm": 0.05780908465385437,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 3.7087855339050293,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43280977312391,
+ "grad_norm": 0.05772170051932335,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 3.7156143188476562,
+ "step": 2407
+ },
+ {
+ "epoch": 33.446771378708554,
+ "grad_norm": 0.05813417211174965,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 3.7155094146728516,
+ "step": 2408
+ },
+ {
+ "epoch": 33.460732984293195,
+ "grad_norm": 0.056307602673769,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 3.7018208503723145,
+ "step": 2409
+ },
+ {
+ "epoch": 33.474694589877835,
+ "grad_norm": 0.05811525136232376,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 3.7002341747283936,
+ "step": 2410
+ },
+ {
+ "epoch": 33.488656195462475,
+ "grad_norm": 0.056484196335077286,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 3.715003490447998,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50261780104712,
+ "grad_norm": 0.057813555002212524,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 3.744851589202881,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51657940663176,
+ "grad_norm": 0.05758954957127571,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 3.712819814682007,
+ "step": 2413
+ },
+ {
+ "epoch": 33.5305410122164,
+ "grad_norm": 0.05723990499973297,
+ "learning_rate": 0.000159218843594243,
+ "loss": 3.7235021591186523,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544502617801044,
+ "grad_norm": 0.058723341673612595,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 3.681633949279785,
+ "step": 2415
+ },
+ {
+ "epoch": 33.55846422338569,
+ "grad_norm": 0.05725013092160225,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 3.7118563652038574,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57242582897033,
+ "grad_norm": 0.05801474303007126,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 3.7071094512939453,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58638743455497,
+ "grad_norm": 0.05690481513738632,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 3.7366995811462402,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60034904013962,
+ "grad_norm": 0.05696716532111168,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 3.7125186920166016,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61431064572426,
+ "grad_norm": 0.05725441873073578,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 3.729522705078125,
+ "step": 2420
+ },
+ {
+ "epoch": 33.6282722513089,
+ "grad_norm": 0.05838564410805702,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 3.713085412979126,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64223385689354,
+ "grad_norm": 0.05807472765445709,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 3.722982883453369,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65619546247819,
+ "grad_norm": 0.05865970253944397,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 3.7254886627197266,
+ "step": 2423
+ },
+ {
+ "epoch": 33.67015706806283,
+ "grad_norm": 0.05879775062203407,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 3.7419567108154297,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68411867364747,
+ "grad_norm": 0.058425575494766235,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 3.7334160804748535,
+ "step": 2425
+ },
+ {
+ "epoch": 33.69808027923211,
+ "grad_norm": 0.05795733258128166,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 3.6824467182159424,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712041884816756,
+ "grad_norm": 0.05840831995010376,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 3.7226808071136475,
+ "step": 2427
+ },
+ {
+ "epoch": 33.726003490401396,
+ "grad_norm": 0.05915387347340584,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 3.7192320823669434,
+ "step": 2428
+ },
+ {
+ "epoch": 33.73996509598604,
+ "grad_norm": 0.05898639187216759,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 3.731309175491333,
+ "step": 2429
+ },
+ {
+ "epoch": 33.753926701570684,
+ "grad_norm": 0.058383479714393616,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 3.6850638389587402,
+ "step": 2430
+ },
+ {
+ "epoch": 33.767888307155324,
+ "grad_norm": 0.06078571453690529,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 3.7126986980438232,
+ "step": 2431
+ },
+ {
+ "epoch": 33.781849912739965,
+ "grad_norm": 0.05873129144310951,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 3.726377010345459,
+ "step": 2432
+ },
+ {
+ "epoch": 33.795811518324605,
+ "grad_norm": 0.060784872621297836,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 3.7208423614501953,
+ "step": 2433
+ },
+ {
+ "epoch": 33.80977312390925,
+ "grad_norm": 0.05992938578128815,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 3.7019124031066895,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82373472949389,
+ "grad_norm": 0.060495104640722275,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 3.7536425590515137,
+ "step": 2435
+ },
+ {
+ "epoch": 33.83769633507853,
+ "grad_norm": 0.059453774243593216,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 3.7357583045959473,
+ "step": 2436
+ },
+ {
+ "epoch": 33.85165794066317,
+ "grad_norm": 0.06152774766087532,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 3.7266006469726562,
+ "step": 2437
+ },
+ {
+ "epoch": 33.86561954624782,
+ "grad_norm": 0.05948188528418541,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 3.712332248687744,
+ "step": 2438
+ },
+ {
+ "epoch": 33.87958115183246,
+ "grad_norm": 0.05947373807430267,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 3.696099042892456,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8935427574171,
+ "grad_norm": 0.060062918812036514,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 3.7297117710113525,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90750436300174,
+ "grad_norm": 0.057695310562849045,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 3.7090835571289062,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92146596858639,
+ "grad_norm": 0.05904509872198105,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 3.763152599334717,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93542757417103,
+ "grad_norm": 0.05886756256222725,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 3.7213687896728516,
+ "step": 2443
+ },
+ {
+ "epoch": 33.94938917975567,
+ "grad_norm": 0.06159936264157295,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 3.7479395866394043,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96335078534032,
+ "grad_norm": 0.057995621114969254,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 3.7327473163604736,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97731239092496,
+ "grad_norm": 0.058990150690078735,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 3.699246883392334,
+ "step": 2446
+ },
+ {
+ "epoch": 33.9912739965096,
+ "grad_norm": 0.05924935266375542,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 3.730475425720215,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.04313633218407631,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 2.336765766143799,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 0.6134145855903625,
+ "eval_runtime": 59.4711,
+ "eval_samples_per_second": 41.062,
+ "eval_steps_per_second": 0.656,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01396160558464,
+ "grad_norm": 0.05632878839969635,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 3.667578935623169,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02792321116929,
+ "grad_norm": 0.05653175711631775,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 3.6973605155944824,
+ "step": 2450
+ },
+ {
+ "epoch": 34.04188481675393,
+ "grad_norm": 0.05482367426156998,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 3.675959587097168,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05584642233857,
+ "grad_norm": 0.05636861175298691,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 3.673849582672119,
+ "step": 2452
+ },
+ {
+ "epoch": 34.06980802792321,
+ "grad_norm": 0.05500229448080063,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 3.6991372108459473,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083769633507856,
+ "grad_norm": 0.056783828884363174,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 3.6838951110839844,
+ "step": 2454
+ },
+ {
+ "epoch": 34.0977312390925,
+ "grad_norm": 0.05543818324804306,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 3.6698226928710938,
+ "step": 2455
+ },
+ {
+ "epoch": 34.11169284467714,
+ "grad_norm": 0.05481055751442909,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 3.678802967071533,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12565445026178,
+ "grad_norm": 0.055152181535959244,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 3.706040620803833,
+ "step": 2457
+ },
+ {
+ "epoch": 34.139616055846425,
+ "grad_norm": 0.05652368441224098,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 3.669009208679199,
+ "step": 2458
+ },
+ {
+ "epoch": 34.153577661431065,
+ "grad_norm": 0.05562841519713402,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 3.6889967918395996,
+ "step": 2459
+ },
+ {
+ "epoch": 34.167539267015705,
+ "grad_norm": 0.05579658970236778,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 3.68369197845459,
+ "step": 2460
+ },
+ {
+ "epoch": 34.181500872600346,
+ "grad_norm": 0.055683478713035583,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 3.670243740081787,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19546247818499,
+ "grad_norm": 0.05624355748295784,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 3.633751392364502,
+ "step": 2462
+ },
+ {
+ "epoch": 34.20942408376963,
+ "grad_norm": 0.05518807843327522,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 3.689868927001953,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223385689354274,
+ "grad_norm": 0.05650133267045021,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 3.694890022277832,
+ "step": 2464
+ },
+ {
+ "epoch": 34.23734729493892,
+ "grad_norm": 0.0555979385972023,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 3.6806082725524902,
+ "step": 2465
+ },
+ {
+ "epoch": 34.25130890052356,
+ "grad_norm": 0.0558185949921608,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 3.6637582778930664,
+ "step": 2466
+ },
+ {
+ "epoch": 34.2652705061082,
+ "grad_norm": 0.056465502828359604,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 3.693030834197998,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27923211169284,
+ "grad_norm": 0.0570668987929821,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 3.6703829765319824,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29319371727749,
+ "grad_norm": 0.05592618137598038,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 3.6874935626983643,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30715532286213,
+ "grad_norm": 0.055560462176799774,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 3.656379222869873,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32111692844677,
+ "grad_norm": 0.056786514818668365,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 3.685967445373535,
+ "step": 2471
+ },
+ {
+ "epoch": 34.33507853403141,
+ "grad_norm": 0.055693406611680984,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 3.693864345550537,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34904013961606,
+ "grad_norm": 0.05816012993454933,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 3.7092270851135254,
+ "step": 2473
+ },
+ {
+ "epoch": 34.3630017452007,
+ "grad_norm": 0.05769353732466698,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 3.6529183387756348,
+ "step": 2474
+ },
+ {
+ "epoch": 34.37696335078534,
+ "grad_norm": 0.058298323303461075,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 3.7092885971069336,
+ "step": 2475
+ },
+ {
+ "epoch": 34.390924956369986,
+ "grad_norm": 0.055963121354579926,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 3.6982855796813965,
+ "step": 2476
+ },
+ {
+ "epoch": 34.404886561954626,
+ "grad_norm": 0.057205602526664734,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 3.690000534057617,
+ "step": 2477
+ },
+ {
+ "epoch": 34.41884816753927,
+ "grad_norm": 0.05652588978409767,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 3.6800036430358887,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43280977312391,
+ "grad_norm": 0.059187956154346466,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 3.677968978881836,
+ "step": 2479
+ },
+ {
+ "epoch": 34.446771378708554,
+ "grad_norm": 0.05786960572004318,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 3.7025253772735596,
+ "step": 2480
+ },
+ {
+ "epoch": 34.460732984293195,
+ "grad_norm": 0.06170958653092384,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 3.6797211170196533,
+ "step": 2481
+ },
+ {
+ "epoch": 34.474694589877835,
+ "grad_norm": 0.0579230934381485,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 3.6861648559570312,
+ "step": 2482
+ },
+ {
+ "epoch": 34.488656195462475,
+ "grad_norm": 0.059319235384464264,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 3.6565022468566895,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50261780104712,
+ "grad_norm": 0.05783943459391594,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 3.722566604614258,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51657940663176,
+ "grad_norm": 0.062192801386117935,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 3.6969285011291504,
+ "step": 2485
+ },
+ {
+ "epoch": 34.5305410122164,
+ "grad_norm": 0.05853543430566788,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 3.6851119995117188,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544502617801044,
+ "grad_norm": 0.05946554243564606,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 3.6720314025878906,
+ "step": 2487
+ },
+ {
+ "epoch": 34.55846422338569,
+ "grad_norm": 0.05766003578901291,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 3.665830135345459,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57242582897033,
+ "grad_norm": 0.06046074256300926,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 3.6955373287200928,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58638743455497,
+ "grad_norm": 0.057485196739435196,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 3.7058653831481934,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60034904013962,
+ "grad_norm": 0.0580572746694088,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 3.6663808822631836,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61431064572426,
+ "grad_norm": 0.05744212865829468,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 3.7024950981140137,
+ "step": 2492
+ },
+ {
+ "epoch": 34.6282722513089,
+ "grad_norm": 0.058456238359212875,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 3.6953721046447754,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64223385689354,
+ "grad_norm": 0.05971676856279373,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 3.6731884479522705,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65619546247819,
+ "grad_norm": 0.058337483555078506,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 3.7169854640960693,
+ "step": 2495
+ },
+ {
+ "epoch": 34.67015706806283,
+ "grad_norm": 0.059567082673311234,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 3.6867635250091553,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68411867364747,
+ "grad_norm": 0.05710845813155174,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 3.6897926330566406,
+ "step": 2497
+ },
+ {
+ "epoch": 34.69808027923211,
+ "grad_norm": 0.05975678935647011,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 3.699439764022827,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712041884816756,
+ "grad_norm": 0.05828933045268059,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 3.677755355834961,
+ "step": 2499
+ },
+ {
+ "epoch": 34.726003490401396,
+ "grad_norm": 0.05896460637450218,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 3.687589645385742,
+ "step": 2500
+ },
+ {
+ "epoch": 34.73996509598604,
+ "grad_norm": 0.05770917236804962,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 3.7188401222229004,
+ "step": 2501
+ },
+ {
+ "epoch": 34.753926701570684,
+ "grad_norm": 0.05971298739314079,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 3.6803629398345947,
+ "step": 2502
+ },
+ {
+ "epoch": 34.767888307155324,
+ "grad_norm": 0.05826874077320099,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 3.7006959915161133,
+ "step": 2503
+ },
+ {
+ "epoch": 34.781849912739965,
+ "grad_norm": 0.05863497033715248,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 3.680962562561035,
+ "step": 2504
+ },
+ {
+ "epoch": 34.795811518324605,
+ "grad_norm": 0.058764465153217316,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 3.7000298500061035,
+ "step": 2505
+ },
+ {
+ "epoch": 34.80977312390925,
+ "grad_norm": 0.05830385908484459,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 3.6838178634643555,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82373472949389,
+ "grad_norm": 0.05855685845017433,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 3.709963321685791,
+ "step": 2507
+ },
+ {
+ "epoch": 34.83769633507853,
+ "grad_norm": 0.05872069671750069,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 3.702852725982666,
+ "step": 2508
+ },
+ {
+ "epoch": 34.85165794066317,
+ "grad_norm": 0.05955277755856514,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 3.699869155883789,
+ "step": 2509
+ },
+ {
+ "epoch": 34.86561954624782,
+ "grad_norm": 0.061080168932676315,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 3.7028279304504395,
+ "step": 2510
+ },
+ {
+ "epoch": 34.87958115183246,
+ "grad_norm": 0.05777806416153908,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 3.6681065559387207,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8935427574171,
+ "grad_norm": 0.0598529689013958,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 3.7238996028900146,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90750436300174,
+ "grad_norm": 0.05878746509552002,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 3.696312427520752,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92146596858639,
+ "grad_norm": 0.06028176471590996,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 3.6749677658081055,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93542757417103,
+ "grad_norm": 0.05837864801287651,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 3.675783634185791,
+ "step": 2515
+ },
+ {
+ "epoch": 34.94938917975567,
+ "grad_norm": 0.060354143381118774,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 3.6795129776000977,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96335078534032,
+ "grad_norm": 0.05837022140622139,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 3.711968183517456,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97731239092496,
+ "grad_norm": 0.06035105511546135,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 3.6994686126708984,
+ "step": 2518
+ },
+ {
+ "epoch": 34.9912739965096,
+ "grad_norm": 0.05695690959692001,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 3.6764185428619385,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.04363907873630524,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 2.2967724800109863,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 0.614627480506897,
+ "eval_runtime": 59.3089,
+ "eval_samples_per_second": 41.174,
+ "eval_steps_per_second": 0.658,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01396160558464,
+ "grad_norm": 0.05742518976330757,
+ "learning_rate": 0.0001349604788260823,
+ "loss": 3.6706724166870117,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02792321116929,
+ "grad_norm": 0.05394700914621353,
+ "learning_rate": 0.00013474073268599152,
+ "loss": 3.6419858932495117,
+ "step": 2522
+ },
+ {
+ "epoch": 35.04188481675393,
+ "grad_norm": 0.05702907219529152,
+ "learning_rate": 0.00013452112210935339,
+ "loss": 3.6621429920196533,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05584642233857,
+ "grad_norm": 0.05404812470078468,
+ "learning_rate": 0.00013430164727310399,
+ "loss": 3.6589713096618652,
+ "step": 2524
+ },
+ {
+ "epoch": 35.06980802792321,
+ "grad_norm": 0.05607686936855316,
+ "learning_rate": 0.00013408230835407027,
+ "loss": 3.634030342102051,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083769633507856,
+ "grad_norm": 0.05493795499205589,
+ "learning_rate": 0.00013386310552896977,
+ "loss": 3.678600788116455,
+ "step": 2526
+ },
+ {
+ "epoch": 35.0977312390925,
+ "grad_norm": 0.0575743205845356,
+ "learning_rate": 0.00013364403897441016,
+ "loss": 3.6530723571777344,
+ "step": 2527
+ },
+ {
+ "epoch": 35.11169284467714,
+ "grad_norm": 0.05405096337199211,
+ "learning_rate": 0.00013342510886688944,
+ "loss": 3.633359432220459,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12565445026178,
+ "grad_norm": 0.055748239159584045,
+ "learning_rate": 0.00013320631538279548,
+ "loss": 3.647953987121582,
+ "step": 2529
+ },
+ {
+ "epoch": 35.139616055846425,
+ "grad_norm": 0.05661927908658981,
+ "learning_rate": 0.00013298765869840658,
+ "loss": 3.6305713653564453,
+ "step": 2530
+ },
+ {
+ "epoch": 35.153577661431065,
+ "grad_norm": 0.05812063813209534,
+ "learning_rate": 0.00013276913898989013,
+ "loss": 3.656938076019287,
+ "step": 2531
+ },
+ {
+ "epoch": 35.167539267015705,
+ "grad_norm": 0.057856228202581406,
+ "learning_rate": 0.00013255075643330369,
+ "loss": 3.657390594482422,
+ "step": 2532
+ },
+ {
+ "epoch": 35.181500872600346,
+ "grad_norm": 0.057047586888074875,
+ "learning_rate": 0.00013233251120459407,
+ "loss": 3.6266355514526367,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19546247818499,
+ "grad_norm": 0.05664423853158951,
+ "learning_rate": 0.0001321144034795976,
+ "loss": 3.6741762161254883,
+ "step": 2534
+ },
+ {
+ "epoch": 35.20942408376963,
+ "grad_norm": 0.056682784110307693,
+ "learning_rate": 0.00013189643343403977,
+ "loss": 3.6106696128845215,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223385689354274,
+ "grad_norm": 0.05825677514076233,
+ "learning_rate": 0.00013167860124353476,
+ "loss": 3.6535234451293945,
+ "step": 2536
+ },
+ {
+ "epoch": 35.23734729493892,
+ "grad_norm": 0.05540613457560539,
+ "learning_rate": 0.00013146090708358657,
+ "loss": 3.6676511764526367,
+ "step": 2537
+ },
+ {
+ "epoch": 35.25130890052356,
+ "grad_norm": 0.05705772340297699,
+ "learning_rate": 0.00013124335112958704,
+ "loss": 3.630685567855835,
+ "step": 2538
+ },
+ {
+ "epoch": 35.2652705061082,
+ "grad_norm": 0.056012462824583054,
+ "learning_rate": 0.0001310259335568172,
+ "loss": 3.6648166179656982,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27923211169284,
+ "grad_norm": 0.05761443451046944,
+ "learning_rate": 0.00013080865454044647,
+ "loss": 3.6521716117858887,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29319371727749,
+ "grad_norm": 0.05589524284005165,
+ "learning_rate": 0.00013059151425553263,
+ "loss": 3.663506031036377,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30715532286213,
+ "grad_norm": 0.056549277156591415,
+ "learning_rate": 0.00013037451287702183,
+ "loss": 3.653916835784912,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32111692844677,
+ "grad_norm": 0.056388892233371735,
+ "learning_rate": 0.0001301576505797479,
+ "loss": 3.6622371673583984,
+ "step": 2543
+ },
+ {
+ "epoch": 35.33507853403141,
+ "grad_norm": 0.05646223947405815,
+ "learning_rate": 0.00012994092753843293,
+ "loss": 3.697643756866455,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34904013961606,
+ "grad_norm": 0.055597055703401566,
+ "learning_rate": 0.00012972434392768687,
+ "loss": 3.657986640930176,
+ "step": 2545
+ },
+ {
+ "epoch": 35.3630017452007,
+ "grad_norm": 0.055099040269851685,
+ "learning_rate": 0.00012950789992200714,
+ "loss": 3.6413917541503906,
+ "step": 2546
+ },
+ {
+ "epoch": 35.37696335078534,
+ "grad_norm": 0.0568387471139431,
+ "learning_rate": 0.00012929159569577886,
+ "loss": 3.640183210372925,
+ "step": 2547
+ },
+ {
+ "epoch": 35.390924956369986,
+ "grad_norm": 0.05552329868078232,
+ "learning_rate": 0.0001290754314232743,
+ "loss": 3.6469502449035645,
+ "step": 2548
+ },
+ {
+ "epoch": 35.404886561954626,
+ "grad_norm": 0.05681881681084633,
+ "learning_rate": 0.00012885940727865334,
+ "loss": 3.6672420501708984,
+ "step": 2549
+ },
+ {
+ "epoch": 35.41884816753927,
+ "grad_norm": 0.05603218451142311,
+ "learning_rate": 0.00012864352343596247,
+ "loss": 3.6844213008880615,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43280977312391,
+ "grad_norm": 0.05711217224597931,
+ "learning_rate": 0.0001284277800691355,
+ "loss": 3.6611440181732178,
+ "step": 2551
+ },
+ {
+ "epoch": 35.446771378708554,
+ "grad_norm": 0.05441901832818985,
+ "learning_rate": 0.00012821217735199298,
+ "loss": 3.6248745918273926,
+ "step": 2552
+ },
+ {
+ "epoch": 35.460732984293195,
+ "grad_norm": 0.05754980444908142,
+ "learning_rate": 0.00012799671545824212,
+ "loss": 3.660444736480713,
+ "step": 2553
+ },
+ {
+ "epoch": 35.474694589877835,
+ "grad_norm": 0.055383119732141495,
+ "learning_rate": 0.0001277813945614768,
+ "loss": 3.6592612266540527,
+ "step": 2554
+ },
+ {
+ "epoch": 35.488656195462475,
+ "grad_norm": 0.05781600996851921,
+ "learning_rate": 0.00012756621483517682,
+ "loss": 3.65926456451416,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50261780104712,
+ "grad_norm": 0.05716165155172348,
+ "learning_rate": 0.00012735117645270905,
+ "loss": 3.6506943702697754,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51657940663176,
+ "grad_norm": 0.05843847990036011,
+ "learning_rate": 0.00012713627958732567,
+ "loss": 3.6542019844055176,
+ "step": 2557
+ },
+ {
+ "epoch": 35.5305410122164,
+ "grad_norm": 0.057625964283943176,
+ "learning_rate": 0.00012692152441216539,
+ "loss": 3.6501665115356445,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544502617801044,
+ "grad_norm": 0.0576319545507431,
+ "learning_rate": 0.0001267069111002525,
+ "loss": 3.647266387939453,
+ "step": 2559
+ },
+ {
+ "epoch": 35.55846422338569,
+ "grad_norm": 0.05821964517235756,
+ "learning_rate": 0.00012649243982449718,
+ "loss": 3.6559696197509766,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57242582897033,
+ "grad_norm": 0.05659124627709389,
+ "learning_rate": 0.00012627811075769502,
+ "loss": 3.6653411388397217,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58638743455497,
+ "grad_norm": 0.05749163031578064,
+ "learning_rate": 0.00012606392407252687,
+ "loss": 3.6690258979797363,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60034904013962,
+ "grad_norm": 0.056859105825424194,
+ "learning_rate": 0.00012584987994155943,
+ "loss": 3.639690637588501,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61431064572426,
+ "grad_norm": 0.0584794282913208,
+ "learning_rate": 0.00012563597853724388,
+ "loss": 3.6673288345336914,
+ "step": 2564
+ },
+ {
+ "epoch": 35.6282722513089,
+ "grad_norm": 0.057991158217191696,
+ "learning_rate": 0.000125422220031917,
+ "loss": 3.6846871376037598,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64223385689354,
+ "grad_norm": 0.05805192142724991,
+ "learning_rate": 0.00012520860459779972,
+ "loss": 3.639286518096924,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65619546247819,
+ "grad_norm": 0.05739975348114967,
+ "learning_rate": 0.0001249951324069986,
+ "loss": 3.6255669593811035,
+ "step": 2567
+ },
+ {
+ "epoch": 35.67015706806283,
+ "grad_norm": 0.057306256145238876,
+ "learning_rate": 0.0001247818036315042,
+ "loss": 3.6530256271362305,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68411867364747,
+ "grad_norm": 0.056031618267297745,
+ "learning_rate": 0.00012456861844319155,
+ "loss": 3.6340036392211914,
+ "step": 2569
+ },
+ {
+ "epoch": 35.69808027923211,
+ "grad_norm": 0.05747546628117561,
+ "learning_rate": 0.00012435557701382012,
+ "loss": 3.670499086380005,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712041884816756,
+ "grad_norm": 0.056736692786216736,
+ "learning_rate": 0.0001241426795150336,
+ "loss": 3.6485583782196045,
+ "step": 2571
+ },
+ {
+ "epoch": 35.726003490401396,
+ "grad_norm": 0.05779654160141945,
+ "learning_rate": 0.00012392992611835973,
+ "loss": 3.6740036010742188,
+ "step": 2572
+ },
+ {
+ "epoch": 35.73996509598604,
+ "grad_norm": 0.05827609449625015,
+ "learning_rate": 0.0001237173169952098,
+ "loss": 3.687256336212158,
+ "step": 2573
+ },
+ {
+ "epoch": 35.753926701570684,
+ "grad_norm": 0.0587199442088604,
+ "learning_rate": 0.00012350485231687954,
+ "loss": 3.660586357116699,
+ "step": 2574
+ },
+ {
+ "epoch": 35.767888307155324,
+ "grad_norm": 0.05669032037258148,
+ "learning_rate": 0.0001232925322545476,
+ "loss": 3.6783642768859863,
+ "step": 2575
+ },
+ {
+ "epoch": 35.781849912739965,
+ "grad_norm": 0.059275999665260315,
+ "learning_rate": 0.0001230803569792765,
+ "loss": 3.660548448562622,
+ "step": 2576
+ },
+ {
+ "epoch": 35.795811518324605,
+ "grad_norm": 0.056993551552295685,
+ "learning_rate": 0.0001228683266620121,
+ "loss": 3.6559674739837646,
+ "step": 2577
+ },
+ {
+ "epoch": 35.80977312390925,
+ "grad_norm": 0.058626480400562286,
+ "learning_rate": 0.00012265644147358326,
+ "loss": 3.679635763168335,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82373472949389,
+ "grad_norm": 0.05863586813211441,
+ "learning_rate": 0.00012244470158470226,
+ "loss": 3.657258987426758,
+ "step": 2579
+ },
+ {
+ "epoch": 35.83769633507853,
+ "grad_norm": 0.05809088423848152,
+ "learning_rate": 0.00012223310716596387,
+ "loss": 3.672903299331665,
+ "step": 2580
+ },
+ {
+ "epoch": 35.85165794066317,
+ "grad_norm": 0.05710308998823166,
+ "learning_rate": 0.00012202165838784601,
+ "loss": 3.679863929748535,
+ "step": 2581
+ },
+ {
+ "epoch": 35.86561954624782,
+ "grad_norm": 0.05867332965135574,
+ "learning_rate": 0.00012181035542070913,
+ "loss": 3.689939498901367,
+ "step": 2582
+ },
+ {
+ "epoch": 35.87958115183246,
+ "grad_norm": 0.057553596794605255,
+ "learning_rate": 0.00012159919843479617,
+ "loss": 3.6570992469787598,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8935427574171,
+ "grad_norm": 0.05948316305875778,
+ "learning_rate": 0.00012138818760023257,
+ "loss": 3.6697707176208496,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90750436300174,
+ "grad_norm": 0.058080170303583145,
+ "learning_rate": 0.00012117732308702592,
+ "loss": 3.667323589324951,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92146596858639,
+ "grad_norm": 0.05842271074652672,
+ "learning_rate": 0.00012096660506506605,
+ "loss": 3.688253164291382,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93542757417103,
+ "grad_norm": 0.058192040771245956,
+ "learning_rate": 0.00012075603370412443,
+ "loss": 3.6728200912475586,
+ "step": 2587
+ },
+ {
+ "epoch": 35.94938917975567,
+ "grad_norm": 0.0602075457572937,
+ "learning_rate": 0.00012054560917385476,
+ "loss": 3.6644749641418457,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96335078534032,
+ "grad_norm": 0.05764734745025635,
+ "learning_rate": 0.00012033533164379225,
+ "loss": 3.711341381072998,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97731239092496,
+ "grad_norm": 0.05787983164191246,
+ "learning_rate": 0.0001201252012833537,
+ "loss": 3.7075908184051514,
+ "step": 2590
+ },
+ {
+ "epoch": 35.9912739965096,
+ "grad_norm": 0.05705810338258743,
+ "learning_rate": 0.00011991521826183747,
+ "loss": 3.664104461669922,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.04269756004214287,
+ "learning_rate": 0.00011970538274842282,
+ "loss": 2.2993788719177246,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 0.6160264611244202,
+ "eval_runtime": 59.3623,
+ "eval_samples_per_second": 41.137,
+ "eval_steps_per_second": 0.657,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01396160558464,
+ "grad_norm": 0.05477261170744896,
+ "learning_rate": 0.00011949569491217073,
+ "loss": 3.6223397254943848,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02792321116929,
+ "grad_norm": 0.05240058898925781,
+ "learning_rate": 0.00011928615492202269,
+ "loss": 3.6275014877319336,
+ "step": 2594
+ },
+ {
+ "epoch": 36.04188481675393,
+ "grad_norm": 0.05488358810544014,
+ "learning_rate": 0.0001190767629468014,
+ "loss": 3.6421685218811035,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05584642233857,
+ "grad_norm": 0.05403005704283714,
+ "learning_rate": 0.00011886751915521007,
+ "loss": 3.6416306495666504,
+ "step": 2596
+ },
+ {
+ "epoch": 36.06980802792321,
+ "grad_norm": 0.05383783578872681,
+ "learning_rate": 0.00011865842371583278,
+ "loss": 3.6026291847229004,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083769633507856,
+ "grad_norm": 0.05379350110888481,
+ "learning_rate": 0.00011844947679713396,
+ "loss": 3.6012558937072754,
+ "step": 2598
+ },
+ {
+ "epoch": 36.0977312390925,
+ "grad_norm": 0.054601337760686874,
+ "learning_rate": 0.00011824067856745812,
+ "loss": 3.6351709365844727,
+ "step": 2599
+ },
+ {
+ "epoch": 36.11169284467714,
+ "grad_norm": 0.0523032620549202,
+ "learning_rate": 0.00011803202919503064,
+ "loss": 3.6155266761779785,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12565445026178,
+ "grad_norm": 0.05588298290967941,
+ "learning_rate": 0.00011782352884795615,
+ "loss": 3.6355879306793213,
+ "step": 2601
+ },
+ {
+ "epoch": 36.139616055846425,
+ "grad_norm": 0.05497356131672859,
+ "learning_rate": 0.00011761517769421983,
+ "loss": 3.5998620986938477,
+ "step": 2602
+ },
+ {
+ "epoch": 36.153577661431065,
+ "grad_norm": 0.055099356919527054,
+ "learning_rate": 0.00011740697590168635,
+ "loss": 3.6135497093200684,
+ "step": 2603
+ },
+ {
+ "epoch": 36.167539267015705,
+ "grad_norm": 0.05480733513832092,
+ "learning_rate": 0.00011719892363810018,
+ "loss": 3.631089687347412,
+ "step": 2604
+ },
+ {
+ "epoch": 36.181500872600346,
+ "grad_norm": 0.05424918979406357,
+ "learning_rate": 0.00011699102107108539,
+ "loss": 3.610616683959961,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19546247818499,
+ "grad_norm": 0.05589015781879425,
+ "learning_rate": 0.00011678326836814507,
+ "loss": 3.6014404296875,
+ "step": 2606
+ },
+ {
+ "epoch": 36.20942408376963,
+ "grad_norm": 0.05365624278783798,
+ "learning_rate": 0.00011657566569666198,
+ "loss": 3.6203207969665527,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223385689354274,
+ "grad_norm": 0.05549495667219162,
+ "learning_rate": 0.00011636821322389777,
+ "loss": 3.639944076538086,
+ "step": 2608
+ },
+ {
+ "epoch": 36.23734729493892,
+ "grad_norm": 0.05381424352526665,
+ "learning_rate": 0.00011616091111699333,
+ "loss": 3.6630518436431885,
+ "step": 2609
+ },
+ {
+ "epoch": 36.25130890052356,
+ "grad_norm": 0.05610237643122673,
+ "learning_rate": 0.00011595375954296786,
+ "loss": 3.634977102279663,
+ "step": 2610
+ },
+ {
+ "epoch": 36.2652705061082,
+ "grad_norm": 0.055497750639915466,
+ "learning_rate": 0.00011574675866871997,
+ "loss": 3.64717960357666,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27923211169284,
+ "grad_norm": 0.0543544627726078,
+ "learning_rate": 0.00011553990866102651,
+ "loss": 3.6224818229675293,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29319371727749,
+ "grad_norm": 0.056236982345581055,
+ "learning_rate": 0.00011533320968654265,
+ "loss": 3.6230814456939697,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30715532286213,
+ "grad_norm": 0.053707852959632874,
+ "learning_rate": 0.00011512666191180204,
+ "loss": 3.6355342864990234,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32111692844677,
+ "grad_norm": 0.057036347687244415,
+ "learning_rate": 0.00011492026550321653,
+ "loss": 3.6099886894226074,
+ "step": 2615
+ },
+ {
+ "epoch": 36.33507853403141,
+ "grad_norm": 0.05581443011760712,
+ "learning_rate": 0.00011471402062707607,
+ "loss": 3.6411590576171875,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34904013961606,
+ "grad_norm": 0.05495850369334221,
+ "learning_rate": 0.00011450792744954827,
+ "loss": 3.6358461380004883,
+ "step": 2617
+ },
+ {
+ "epoch": 36.3630017452007,
+ "grad_norm": 0.0549640990793705,
+ "learning_rate": 0.00011430198613667861,
+ "loss": 3.633671283721924,
+ "step": 2618
+ },
+ {
+ "epoch": 36.37696335078534,
+ "grad_norm": 0.05502061918377876,
+ "learning_rate": 0.00011409619685439064,
+ "loss": 3.632413864135742,
+ "step": 2619
+ },
+ {
+ "epoch": 36.390924956369986,
+ "grad_norm": 0.05518985912203789,
+ "learning_rate": 0.00011389055976848477,
+ "loss": 3.646486759185791,
+ "step": 2620
+ },
+ {
+ "epoch": 36.404886561954626,
+ "grad_norm": 0.053396400064229965,
+ "learning_rate": 0.00011368507504463914,
+ "loss": 3.6610465049743652,
+ "step": 2621
+ },
+ {
+ "epoch": 36.41884816753927,
+ "grad_norm": 0.05592766031622887,
+ "learning_rate": 0.00011347974284840914,
+ "loss": 3.6322021484375,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43280977312391,
+ "grad_norm": 0.054665833711624146,
+ "learning_rate": 0.00011327456334522725,
+ "loss": 3.6100504398345947,
+ "step": 2623
+ },
+ {
+ "epoch": 36.446771378708554,
+ "grad_norm": 0.05556786060333252,
+ "learning_rate": 0.00011306953670040275,
+ "loss": 3.647045135498047,
+ "step": 2624
+ },
+ {
+ "epoch": 36.460732984293195,
+ "grad_norm": 0.05619419366121292,
+ "learning_rate": 0.00011286466307912184,
+ "loss": 3.6307947635650635,
+ "step": 2625
+ },
+ {
+ "epoch": 36.474694589877835,
+ "grad_norm": 0.0551043376326561,
+ "learning_rate": 0.00011265994264644783,
+ "loss": 3.6331896781921387,
+ "step": 2626
+ },
+ {
+ "epoch": 36.488656195462475,
+ "grad_norm": 0.0548650398850441,
+ "learning_rate": 0.00011245537556732001,
+ "loss": 3.633991241455078,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50261780104712,
+ "grad_norm": 0.056781429797410965,
+ "learning_rate": 0.00011225096200655442,
+ "loss": 3.653165340423584,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51657940663176,
+ "grad_norm": 0.05489041283726692,
+ "learning_rate": 0.00011204670212884342,
+ "loss": 3.6232171058654785,
+ "step": 2629
+ },
+ {
+ "epoch": 36.5305410122164,
+ "grad_norm": 0.055730190128088,
+ "learning_rate": 0.00011184259609875557,
+ "loss": 3.644775390625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544502617801044,
+ "grad_norm": 0.058052800595760345,
+ "learning_rate": 0.00011163864408073522,
+ "loss": 3.6141133308410645,
+ "step": 2631
+ },
+ {
+ "epoch": 36.55846422338569,
+ "grad_norm": 0.05508120357990265,
+ "learning_rate": 0.00011143484623910293,
+ "loss": 3.6275734901428223,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57242582897033,
+ "grad_norm": 0.0572693794965744,
+ "learning_rate": 0.00011123120273805496,
+ "loss": 3.631286859512329,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58638743455497,
+ "grad_norm": 0.05745089426636696,
+ "learning_rate": 0.0001110277137416632,
+ "loss": 3.6396098136901855,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60034904013962,
+ "grad_norm": 0.05624939128756523,
+ "learning_rate": 0.00011082437941387512,
+ "loss": 3.655447006225586,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61431064572426,
+ "grad_norm": 0.057037804275751114,
+ "learning_rate": 0.00011062119991851322,
+ "loss": 3.6295433044433594,
+ "step": 2636
+ },
+ {
+ "epoch": 36.6282722513089,
+ "grad_norm": 0.05740547552704811,
+ "learning_rate": 0.000110418175419276,
+ "loss": 3.6350302696228027,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64223385689354,
+ "grad_norm": 0.05586782470345497,
+ "learning_rate": 0.00011021530607973626,
+ "loss": 3.6298985481262207,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65619546247819,
+ "grad_norm": 0.05766303837299347,
+ "learning_rate": 0.00011001259206334235,
+ "loss": 3.646199941635132,
+ "step": 2639
+ },
+ {
+ "epoch": 36.67015706806283,
+ "grad_norm": 0.05768761411309242,
+ "learning_rate": 0.00010981003353341721,
+ "loss": 3.638676881790161,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68411867364747,
+ "grad_norm": 0.05806223303079605,
+ "learning_rate": 0.00010960763065315864,
+ "loss": 3.61555814743042,
+ "step": 2641
+ },
+ {
+ "epoch": 36.69808027923211,
+ "grad_norm": 0.05654353275895119,
+ "learning_rate": 0.00010940538358563906,
+ "loss": 3.6435892581939697,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712041884816756,
+ "grad_norm": 0.0572567842900753,
+ "learning_rate": 0.00010920329249380509,
+ "loss": 3.638545513153076,
+ "step": 2643
+ },
+ {
+ "epoch": 36.726003490401396,
+ "grad_norm": 0.05685669183731079,
+ "learning_rate": 0.00010900135754047799,
+ "loss": 3.6248223781585693,
+ "step": 2644
+ },
+ {
+ "epoch": 36.73996509598604,
+ "grad_norm": 0.05836780369281769,
+ "learning_rate": 0.00010879957888835307,
+ "loss": 3.6728572845458984,
+ "step": 2645
+ },
+ {
+ "epoch": 36.753926701570684,
+ "grad_norm": 0.05652812868356705,
+ "learning_rate": 0.00010859795669999981,
+ "loss": 3.6360411643981934,
+ "step": 2646
+ },
+ {
+ "epoch": 36.767888307155324,
+ "grad_norm": 0.05731404572725296,
+ "learning_rate": 0.00010839649113786137,
+ "loss": 3.624966621398926,
+ "step": 2647
+ },
+ {
+ "epoch": 36.781849912739965,
+ "grad_norm": 0.05664033442735672,
+ "learning_rate": 0.00010819518236425514,
+ "loss": 3.6407511234283447,
+ "step": 2648
+ },
+ {
+ "epoch": 36.795811518324605,
+ "grad_norm": 0.056350719183683395,
+ "learning_rate": 0.00010799403054137197,
+ "loss": 3.6436259746551514,
+ "step": 2649
+ },
+ {
+ "epoch": 36.80977312390925,
+ "grad_norm": 0.05680667608976364,
+ "learning_rate": 0.00010779303583127609,
+ "loss": 3.6506614685058594,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82373472949389,
+ "grad_norm": 0.05701072886586189,
+ "learning_rate": 0.00010759219839590535,
+ "loss": 3.643155574798584,
+ "step": 2651
+ },
+ {
+ "epoch": 36.83769633507853,
+ "grad_norm": 0.05728844553232193,
+ "learning_rate": 0.00010739151839707089,
+ "loss": 3.6734180450439453,
+ "step": 2652
+ },
+ {
+ "epoch": 36.85165794066317,
+ "grad_norm": 0.05713675543665886,
+ "learning_rate": 0.00010719099599645706,
+ "loss": 3.665465831756592,
+ "step": 2653
+ },
+ {
+ "epoch": 36.86561954624782,
+ "grad_norm": 0.05626962333917618,
+ "learning_rate": 0.0001069906313556208,
+ "loss": 3.6559338569641113,
+ "step": 2654
+ },
+ {
+ "epoch": 36.87958115183246,
+ "grad_norm": 0.05827031657099724,
+ "learning_rate": 0.00010679042463599263,
+ "loss": 3.64331316947937,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8935427574171,
+ "grad_norm": 0.05828213319182396,
+ "learning_rate": 0.00010659037599887545,
+ "loss": 3.631319046020508,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90750436300174,
+ "grad_norm": 0.057474542409181595,
+ "learning_rate": 0.0001063904856054446,
+ "loss": 3.6365761756896973,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92146596858639,
+ "grad_norm": 0.056101586669683456,
+ "learning_rate": 0.00010619075361674836,
+ "loss": 3.62568998336792,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93542757417103,
+ "grad_norm": 0.058322884142398834,
+ "learning_rate": 0.0001059911801937071,
+ "loss": 3.6452713012695312,
+ "step": 2659
+ },
+ {
+ "epoch": 36.94938917975567,
+ "grad_norm": 0.05735790356993675,
+ "learning_rate": 0.00010579176549711365,
+ "loss": 3.644145965576172,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96335078534032,
+ "grad_norm": 0.05761740356683731,
+ "learning_rate": 0.00010559250968763265,
+ "loss": 3.652322292327881,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97731239092496,
+ "grad_norm": 0.05765356495976448,
+ "learning_rate": 0.00010539341292580086,
+ "loss": 3.6571314334869385,
+ "step": 2662
+ },
+ {
+ "epoch": 36.9912739965096,
+ "grad_norm": 0.05606544017791748,
+ "learning_rate": 0.00010519447537202729,
+ "loss": 3.6158785820007324,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.042550574988126755,
+ "learning_rate": 0.00010499569718659201,
+ "loss": 2.262657403945923,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 0.6172080039978027,
+ "eval_runtime": 59.1561,
+ "eval_samples_per_second": 41.281,
+ "eval_steps_per_second": 0.659,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01396160558464,
+ "grad_norm": 0.054308392107486725,
+ "learning_rate": 0.00010479707852964713,
+ "loss": 3.6250295639038086,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02792321116929,
+ "grad_norm": 0.05218822881579399,
+ "learning_rate": 0.00010459861956121611,
+ "loss": 3.598968982696533,
+ "step": 2666
+ },
+ {
+ "epoch": 37.04188481675393,
+ "grad_norm": 0.05366450920701027,
+ "learning_rate": 0.00010440032044119383,
+ "loss": 3.5767178535461426,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05584642233857,
+ "grad_norm": 0.05169176682829857,
+ "learning_rate": 0.00010420218132934614,
+ "loss": 3.5973262786865234,
+ "step": 2668
+ },
+ {
+ "epoch": 37.06980802792321,
+ "grad_norm": 0.05438115820288658,
+ "learning_rate": 0.00010400420238531023,
+ "loss": 3.584498167037964,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083769633507856,
+ "grad_norm": 0.05290330946445465,
+ "learning_rate": 0.00010380638376859416,
+ "loss": 3.614943504333496,
+ "step": 2670
+ },
+ {
+ "epoch": 37.0977312390925,
+ "grad_norm": 0.052781153470277786,
+ "learning_rate": 0.00010360872563857682,
+ "loss": 3.5990896224975586,
+ "step": 2671
+ },
+ {
+ "epoch": 37.11169284467714,
+ "grad_norm": 0.05197159945964813,
+ "learning_rate": 0.0001034112281545079,
+ "loss": 3.579005241394043,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12565445026178,
+ "grad_norm": 0.05242403224110603,
+ "learning_rate": 0.00010321389147550725,
+ "loss": 3.608119010925293,
+ "step": 2673
+ },
+ {
+ "epoch": 37.139616055846425,
+ "grad_norm": 0.05325670540332794,
+ "learning_rate": 0.00010301671576056588,
+ "loss": 3.6093316078186035,
+ "step": 2674
+ },
+ {
+ "epoch": 37.153577661431065,
+ "grad_norm": 0.05325893685221672,
+ "learning_rate": 0.00010281970116854436,
+ "loss": 3.6037325859069824,
+ "step": 2675
+ },
+ {
+ "epoch": 37.167539267015705,
+ "grad_norm": 0.05180941894650459,
+ "learning_rate": 0.00010262284785817392,
+ "loss": 3.584935188293457,
+ "step": 2676
+ },
+ {
+ "epoch": 37.181500872600346,
+ "grad_norm": 0.05396177992224693,
+ "learning_rate": 0.00010242615598805574,
+ "loss": 3.623363494873047,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19546247818499,
+ "grad_norm": 0.05336499959230423,
+ "learning_rate": 0.00010222962571666088,
+ "loss": 3.6231229305267334,
+ "step": 2678
+ },
+ {
+ "epoch": 37.20942408376963,
+ "grad_norm": 0.052409425377845764,
+ "learning_rate": 0.00010203325720233032,
+ "loss": 3.599355697631836,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223385689354274,
+ "grad_norm": 0.0531868040561676,
+ "learning_rate": 0.00010183705060327433,
+ "loss": 3.6197028160095215,
+ "step": 2680
+ },
+ {
+ "epoch": 37.23734729493892,
+ "grad_norm": 0.053192511200904846,
+ "learning_rate": 0.00010164100607757346,
+ "loss": 3.612718105316162,
+ "step": 2681
+ },
+ {
+ "epoch": 37.25130890052356,
+ "grad_norm": 0.052938785403966904,
+ "learning_rate": 0.00010144512378317687,
+ "loss": 3.6197361946105957,
+ "step": 2682
+ },
+ {
+ "epoch": 37.2652705061082,
+ "grad_norm": 0.0538516603410244,
+ "learning_rate": 0.00010124940387790354,
+ "loss": 3.6179161071777344,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27923211169284,
+ "grad_norm": 0.05364573001861572,
+ "learning_rate": 0.00010105384651944148,
+ "loss": 3.5888404846191406,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29319371727749,
+ "grad_norm": 0.05395280569791794,
+ "learning_rate": 0.00010085845186534769,
+ "loss": 3.6174027919769287,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30715532286213,
+ "grad_norm": 0.05424579605460167,
+ "learning_rate": 0.00010066322007304819,
+ "loss": 3.6061956882476807,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32111692844677,
+ "grad_norm": 0.05412362515926361,
+ "learning_rate": 0.00010046815129983757,
+ "loss": 3.596550464630127,
+ "step": 2687
+ },
+ {
+ "epoch": 37.33507853403141,
+ "grad_norm": 0.05387300252914429,
+ "learning_rate": 0.00010027324570287925,
+ "loss": 3.627476215362549,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34904013961606,
+ "grad_norm": 0.05500417947769165,
+ "learning_rate": 0.00010007850343920519,
+ "loss": 3.601841926574707,
+ "step": 2689
+ },
+ {
+ "epoch": 37.3630017452007,
+ "grad_norm": 0.05346129834651947,
+ "learning_rate": 9.988392466571572e-05,
+ "loss": 3.6150965690612793,
+ "step": 2690
+ },
+ {
+ "epoch": 37.37696335078534,
+ "grad_norm": 0.054463278502225876,
+ "learning_rate": 9.968950953917945e-05,
+ "loss": 3.5890097618103027,
+ "step": 2691
+ },
+ {
+ "epoch": 37.390924956369986,
+ "grad_norm": 0.05312243849039078,
+ "learning_rate": 9.949525821623309e-05,
+ "loss": 3.597362518310547,
+ "step": 2692
+ },
+ {
+ "epoch": 37.404886561954626,
+ "grad_norm": 0.05424235016107559,
+ "learning_rate": 9.930117085338156e-05,
+ "loss": 3.5707716941833496,
+ "step": 2693
+ },
+ {
+ "epoch": 37.41884816753927,
+ "grad_norm": 0.054021209478378296,
+ "learning_rate": 9.910724760699731e-05,
+ "loss": 3.6366126537323,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43280977312391,
+ "grad_norm": 0.05502825230360031,
+ "learning_rate": 9.891348863332093e-05,
+ "loss": 3.6250295639038086,
+ "step": 2695
+ },
+ {
+ "epoch": 37.446771378708554,
+ "grad_norm": 0.05483090132474899,
+ "learning_rate": 9.871989408846051e-05,
+ "loss": 3.611407518386841,
+ "step": 2696
+ },
+ {
+ "epoch": 37.460732984293195,
+ "grad_norm": 0.05381562188267708,
+ "learning_rate": 9.852646412839178e-05,
+ "loss": 3.6136293411254883,
+ "step": 2697
+ },
+ {
+ "epoch": 37.474694589877835,
+ "grad_norm": 0.054287511855363846,
+ "learning_rate": 9.833319890895756e-05,
+ "loss": 3.6080033779144287,
+ "step": 2698
+ },
+ {
+ "epoch": 37.488656195462475,
+ "grad_norm": 0.05399102717638016,
+ "learning_rate": 9.814009858586817e-05,
+ "loss": 3.6026833057403564,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50261780104712,
+ "grad_norm": 0.053677961230278015,
+ "learning_rate": 9.794716331470131e-05,
+ "loss": 3.6080117225646973,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51657940663176,
+ "grad_norm": 0.055184055119752884,
+ "learning_rate": 9.77543932509012e-05,
+ "loss": 3.6028759479522705,
+ "step": 2701
+ },
+ {
+ "epoch": 37.5305410122164,
+ "grad_norm": 0.0554618164896965,
+ "learning_rate": 9.756178854977925e-05,
+ "loss": 3.6170661449432373,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544502617801044,
+ "grad_norm": 0.05491705983877182,
+ "learning_rate": 9.736934936651362e-05,
+ "loss": 3.6152079105377197,
+ "step": 2703
+ },
+ {
+ "epoch": 37.55846422338569,
+ "grad_norm": 0.055647511035203934,
+ "learning_rate": 9.717707585614916e-05,
+ "loss": 3.610424757003784,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57242582897033,
+ "grad_norm": 0.0561630018055439,
+ "learning_rate": 9.698496817359698e-05,
+ "loss": 3.6318728923797607,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58638743455497,
+ "grad_norm": 0.05447172373533249,
+ "learning_rate": 9.679302647363476e-05,
+ "loss": 3.6186347007751465,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60034904013962,
+ "grad_norm": 0.05509922280907631,
+ "learning_rate": 9.660125091090675e-05,
+ "loss": 3.6077895164489746,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61431064572426,
+ "grad_norm": 0.055178701877593994,
+ "learning_rate": 9.64096416399228e-05,
+ "loss": 3.637204647064209,
+ "step": 2708
+ },
+ {
+ "epoch": 37.6282722513089,
+ "grad_norm": 0.05637604370713234,
+ "learning_rate": 9.621819881505918e-05,
+ "loss": 3.6154603958129883,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64223385689354,
+ "grad_norm": 0.056297753006219864,
+ "learning_rate": 9.602692259055767e-05,
+ "loss": 3.6120991706848145,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65619546247819,
+ "grad_norm": 0.054915741086006165,
+ "learning_rate": 9.583581312052646e-05,
+ "loss": 3.626859188079834,
+ "step": 2711
+ },
+ {
+ "epoch": 37.67015706806283,
+ "grad_norm": 0.05664214491844177,
+ "learning_rate": 9.564487055893867e-05,
+ "loss": 3.6137616634368896,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68411867364747,
+ "grad_norm": 0.05572838336229324,
+ "learning_rate": 9.545409505963338e-05,
+ "loss": 3.622990131378174,
+ "step": 2713
+ },
+ {
+ "epoch": 37.69808027923211,
+ "grad_norm": 0.056440044194459915,
+ "learning_rate": 9.526348677631499e-05,
+ "loss": 3.6483852863311768,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712041884816756,
+ "grad_norm": 0.0582021027803421,
+ "learning_rate": 9.507304586255311e-05,
+ "loss": 3.6085047721862793,
+ "step": 2715
+ },
+ {
+ "epoch": 37.726003490401396,
+ "grad_norm": 0.055895350873470306,
+ "learning_rate": 9.488277247178267e-05,
+ "loss": 3.625001907348633,
+ "step": 2716
+ },
+ {
+ "epoch": 37.73996509598604,
+ "grad_norm": 0.055818382650613785,
+ "learning_rate": 9.469266675730319e-05,
+ "loss": 3.627051830291748,
+ "step": 2717
+ },
+ {
+ "epoch": 37.753926701570684,
+ "grad_norm": 0.05565737560391426,
+ "learning_rate": 9.450272887227983e-05,
+ "loss": 3.607292890548706,
+ "step": 2718
+ },
+ {
+ "epoch": 37.767888307155324,
+ "grad_norm": 0.05473797023296356,
+ "learning_rate": 9.431295896974182e-05,
+ "loss": 3.613346815109253,
+ "step": 2719
+ },
+ {
+ "epoch": 37.781849912739965,
+ "grad_norm": 0.056405726820230484,
+ "learning_rate": 9.412335720258341e-05,
+ "loss": 3.6148681640625,
+ "step": 2720
+ },
+ {
+ "epoch": 37.795811518324605,
+ "grad_norm": 0.0563163161277771,
+ "learning_rate": 9.393392372356335e-05,
+ "loss": 3.6225497722625732,
+ "step": 2721
+ },
+ {
+ "epoch": 37.80977312390925,
+ "grad_norm": 0.05682912468910217,
+ "learning_rate": 9.374465868530477e-05,
+ "loss": 3.620290756225586,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82373472949389,
+ "grad_norm": 0.05696539953351021,
+ "learning_rate": 9.355556224029515e-05,
+ "loss": 3.617171287536621,
+ "step": 2723
+ },
+ {
+ "epoch": 37.83769633507853,
+ "grad_norm": 0.05691574513912201,
+ "learning_rate": 9.336663454088593e-05,
+ "loss": 3.6293022632598877,
+ "step": 2724
+ },
+ {
+ "epoch": 37.85165794066317,
+ "grad_norm": 0.05566944554448128,
+ "learning_rate": 9.317787573929282e-05,
+ "loss": 3.6245670318603516,
+ "step": 2725
+ },
+ {
+ "epoch": 37.86561954624782,
+ "grad_norm": 0.05736027657985687,
+ "learning_rate": 9.298928598759541e-05,
+ "loss": 3.625857353210449,
+ "step": 2726
+ },
+ {
+ "epoch": 37.87958115183246,
+ "grad_norm": 0.055642999708652496,
+ "learning_rate": 9.280086543773698e-05,
+ "loss": 3.610138416290283,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8935427574171,
+ "grad_norm": 0.057330258190631866,
+ "learning_rate": 9.26126142415247e-05,
+ "loss": 3.593010902404785,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90750436300174,
+ "grad_norm": 0.05633636936545372,
+ "learning_rate": 9.2424532550629e-05,
+ "loss": 3.6144258975982666,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92146596858639,
+ "grad_norm": 0.05787842720746994,
+ "learning_rate": 9.223662051658408e-05,
+ "loss": 3.609016180038452,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93542757417103,
+ "grad_norm": 0.05679284781217575,
+ "learning_rate": 9.204887829078709e-05,
+ "loss": 3.5973129272460938,
+ "step": 2731
+ },
+ {
+ "epoch": 37.94938917975567,
+ "grad_norm": 0.056735891848802567,
+ "learning_rate": 9.186130602449861e-05,
+ "loss": 3.6157774925231934,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96335078534032,
+ "grad_norm": 0.05625693500041962,
+ "learning_rate": 9.167390386884224e-05,
+ "loss": 3.617924928665161,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97731239092496,
+ "grad_norm": 0.05621540546417236,
+ "learning_rate": 9.148667197480455e-05,
+ "loss": 3.597618341445923,
+ "step": 2734
+ },
+ {
+ "epoch": 37.9912739965096,
+ "grad_norm": 0.0570329986512661,
+ "learning_rate": 9.129961049323494e-05,
+ "loss": 3.6188273429870605,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.04125213250517845,
+ "learning_rate": 9.111271957484519e-05,
+ "loss": 2.2358860969543457,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 0.6186589002609253,
+ "eval_runtime": 58.8494,
+ "eval_samples_per_second": 41.496,
+ "eval_steps_per_second": 0.663,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01396160558464,
+ "grad_norm": 0.05371171608567238,
+ "learning_rate": 9.092599937021031e-05,
+ "loss": 3.5614218711853027,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02792321116929,
+ "grad_norm": 0.05166265740990639,
+ "learning_rate": 9.073945002976715e-05,
+ "loss": 3.5690364837646484,
+ "step": 2738
+ },
+ {
+ "epoch": 38.04188481675393,
+ "grad_norm": 0.053279295563697815,
+ "learning_rate": 9.055307170381522e-05,
+ "loss": 3.575930595397949,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05584642233857,
+ "grad_norm": 0.051418207585811615,
+ "learning_rate": 9.036686454251613e-05,
+ "loss": 3.568121910095215,
+ "step": 2740
+ },
+ {
+ "epoch": 38.06980802792321,
+ "grad_norm": 0.051363058388233185,
+ "learning_rate": 9.018082869589369e-05,
+ "loss": 3.5916590690612793,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083769633507856,
+ "grad_norm": 0.05149349942803383,
+ "learning_rate": 8.999496431383368e-05,
+ "loss": 3.5861239433288574,
+ "step": 2742
+ },
+ {
+ "epoch": 38.0977312390925,
+ "grad_norm": 0.05174670368432999,
+ "learning_rate": 8.980927154608341e-05,
+ "loss": 3.5812082290649414,
+ "step": 2743
+ },
+ {
+ "epoch": 38.11169284467714,
+ "grad_norm": 0.05128023028373718,
+ "learning_rate": 8.962375054225255e-05,
+ "loss": 3.5962631702423096,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12565445026178,
+ "grad_norm": 0.05153957009315491,
+ "learning_rate": 8.943840145181177e-05,
+ "loss": 3.5739307403564453,
+ "step": 2745
+ },
+ {
+ "epoch": 38.139616055846425,
+ "grad_norm": 0.0521809458732605,
+ "learning_rate": 8.925322442409361e-05,
+ "loss": 3.5964818000793457,
+ "step": 2746
+ },
+ {
+ "epoch": 38.153577661431065,
+ "grad_norm": 0.05092332884669304,
+ "learning_rate": 8.906821960829163e-05,
+ "loss": 3.5737056732177734,
+ "step": 2747
+ },
+ {
+ "epoch": 38.167539267015705,
+ "grad_norm": 0.052281491458415985,
+ "learning_rate": 8.888338715346113e-05,
+ "loss": 3.5943212509155273,
+ "step": 2748
+ },
+ {
+ "epoch": 38.181500872600346,
+ "grad_norm": 0.05170704424381256,
+ "learning_rate": 8.869872720851831e-05,
+ "loss": 3.585753917694092,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19546247818499,
+ "grad_norm": 0.05175631120800972,
+ "learning_rate": 8.851423992224012e-05,
+ "loss": 3.606010913848877,
+ "step": 2750
+ },
+ {
+ "epoch": 38.20942408376963,
+ "grad_norm": 0.0525258332490921,
+ "learning_rate": 8.832992544326473e-05,
+ "loss": 3.591367244720459,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223385689354274,
+ "grad_norm": 0.051923174411058426,
+ "learning_rate": 8.814578392009104e-05,
+ "loss": 3.563871383666992,
+ "step": 2752
+ },
+ {
+ "epoch": 38.23734729493892,
+ "grad_norm": 0.053746938705444336,
+ "learning_rate": 8.796181550107857e-05,
+ "loss": 3.576528787612915,
+ "step": 2753
+ },
+ {
+ "epoch": 38.25130890052356,
+ "grad_norm": 0.05283728986978531,
+ "learning_rate": 8.777802033444712e-05,
+ "loss": 3.600759983062744,
+ "step": 2754
+ },
+ {
+ "epoch": 38.2652705061082,
+ "grad_norm": 0.05214889720082283,
+ "learning_rate": 8.75943985682774e-05,
+ "loss": 3.582427501678467,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27923211169284,
+ "grad_norm": 0.05283285677433014,
+ "learning_rate": 8.741095035051017e-05,
+ "loss": 3.579399585723877,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29319371727749,
+ "grad_norm": 0.053677286952733994,
+ "learning_rate": 8.722767582894613e-05,
+ "loss": 3.5718324184417725,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30715532286213,
+ "grad_norm": 0.05384226143360138,
+ "learning_rate": 8.704457515124636e-05,
+ "loss": 3.593702793121338,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32111692844677,
+ "grad_norm": 0.05237982049584389,
+ "learning_rate": 8.686164846493176e-05,
+ "loss": 3.580988883972168,
+ "step": 2759
+ },
+ {
+ "epoch": 38.33507853403141,
+ "grad_norm": 0.05386205390095711,
+ "learning_rate": 8.667889591738317e-05,
+ "loss": 3.5706114768981934,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34904013961606,
+ "grad_norm": 0.05325917899608612,
+ "learning_rate": 8.649631765584083e-05,
+ "loss": 3.5707662105560303,
+ "step": 2761
+ },
+ {
+ "epoch": 38.3630017452007,
+ "grad_norm": 0.053492557257413864,
+ "learning_rate": 8.631391382740482e-05,
+ "loss": 3.618515968322754,
+ "step": 2762
+ },
+ {
+ "epoch": 38.37696335078534,
+ "grad_norm": 0.05370897799730301,
+ "learning_rate": 8.613168457903459e-05,
+ "loss": 3.5731120109558105,
+ "step": 2763
+ },
+ {
+ "epoch": 38.390924956369986,
+ "grad_norm": 0.05325158312916756,
+ "learning_rate": 8.594963005754901e-05,
+ "loss": 3.591339588165283,
+ "step": 2764
+ },
+ {
+ "epoch": 38.404886561954626,
+ "grad_norm": 0.05411994829773903,
+ "learning_rate": 8.57677504096261e-05,
+ "loss": 3.5969743728637695,
+ "step": 2765
+ },
+ {
+ "epoch": 38.41884816753927,
+ "grad_norm": 0.05363132059574127,
+ "learning_rate": 8.558604578180301e-05,
+ "loss": 3.620718479156494,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43280977312391,
+ "grad_norm": 0.0541871078312397,
+ "learning_rate": 8.540451632047593e-05,
+ "loss": 3.589287281036377,
+ "step": 2767
+ },
+ {
+ "epoch": 38.446771378708554,
+ "grad_norm": 0.05317665636539459,
+ "learning_rate": 8.522316217189972e-05,
+ "loss": 3.5677599906921387,
+ "step": 2768
+ },
+ {
+ "epoch": 38.460732984293195,
+ "grad_norm": 0.05288826301693916,
+ "learning_rate": 8.504198348218821e-05,
+ "loss": 3.5985312461853027,
+ "step": 2769
+ },
+ {
+ "epoch": 38.474694589877835,
+ "grad_norm": 0.05443241819739342,
+ "learning_rate": 8.486098039731384e-05,
+ "loss": 3.5577774047851562,
+ "step": 2770
+ },
+ {
+ "epoch": 38.488656195462475,
+ "grad_norm": 0.054173294454813004,
+ "learning_rate": 8.46801530631075e-05,
+ "loss": 3.5860495567321777,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50261780104712,
+ "grad_norm": 0.05407063290476799,
+ "learning_rate": 8.449950162525859e-05,
+ "loss": 3.604748249053955,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51657940663176,
+ "grad_norm": 0.05461187660694122,
+ "learning_rate": 8.431902622931443e-05,
+ "loss": 3.563236713409424,
+ "step": 2773
+ },
+ {
+ "epoch": 38.5305410122164,
+ "grad_norm": 0.05529208108782768,
+ "learning_rate": 8.413872702068119e-05,
+ "loss": 3.586937427520752,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544502617801044,
+ "grad_norm": 0.05481037497520447,
+ "learning_rate": 8.395860414462238e-05,
+ "loss": 3.6134698390960693,
+ "step": 2775
+ },
+ {
+ "epoch": 38.55846422338569,
+ "grad_norm": 0.054424822330474854,
+ "learning_rate": 8.377865774625985e-05,
+ "loss": 3.575363874435425,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57242582897033,
+ "grad_norm": 0.0553407222032547,
+ "learning_rate": 8.359888797057316e-05,
+ "loss": 3.5970382690429688,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58638743455497,
+ "grad_norm": 0.05625150352716446,
+ "learning_rate": 8.341929496239958e-05,
+ "loss": 3.6032121181488037,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60034904013962,
+ "grad_norm": 0.05586005747318268,
+ "learning_rate": 8.323987886643402e-05,
+ "loss": 3.602691888809204,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61431064572426,
+ "grad_norm": 0.054208606481552124,
+ "learning_rate": 8.306063982722855e-05,
+ "loss": 3.601468563079834,
+ "step": 2780
+ },
+ {
+ "epoch": 38.6282722513089,
+ "grad_norm": 0.056410644203424454,
+ "learning_rate": 8.288157798919318e-05,
+ "loss": 3.585646867752075,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64223385689354,
+ "grad_norm": 0.054060712456703186,
+ "learning_rate": 8.27026934965945e-05,
+ "loss": 3.566913604736328,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65619546247819,
+ "grad_norm": 0.0546812042593956,
+ "learning_rate": 8.25239864935566e-05,
+ "loss": 3.581869602203369,
+ "step": 2783
+ },
+ {
+ "epoch": 38.67015706806283,
+ "grad_norm": 0.05502549931406975,
+ "learning_rate": 8.23454571240605e-05,
+ "loss": 3.584887981414795,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68411867364747,
+ "grad_norm": 0.056419193744659424,
+ "learning_rate": 8.216710553194416e-05,
+ "loss": 3.5797278881073,
+ "step": 2785
+ },
+ {
+ "epoch": 38.69808027923211,
+ "grad_norm": 0.05669495090842247,
+ "learning_rate": 8.198893186090222e-05,
+ "loss": 3.604696273803711,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712041884816756,
+ "grad_norm": 0.05554523691534996,
+ "learning_rate": 8.181093625448585e-05,
+ "loss": 3.591275215148926,
+ "step": 2787
+ },
+ {
+ "epoch": 38.726003490401396,
+ "grad_norm": 0.05579648166894913,
+ "learning_rate": 8.163311885610299e-05,
+ "loss": 3.5911900997161865,
+ "step": 2788
+ },
+ {
+ "epoch": 38.73996509598604,
+ "grad_norm": 0.05564593896269798,
+ "learning_rate": 8.145547980901791e-05,
+ "loss": 3.6048314571380615,
+ "step": 2789
+ },
+ {
+ "epoch": 38.753926701570684,
+ "grad_norm": 0.05567439645528793,
+ "learning_rate": 8.127801925635126e-05,
+ "loss": 3.564915180206299,
+ "step": 2790
+ },
+ {
+ "epoch": 38.767888307155324,
+ "grad_norm": 0.055503327399492264,
+ "learning_rate": 8.110073734107954e-05,
+ "loss": 3.574024200439453,
+ "step": 2791
+ },
+ {
+ "epoch": 38.781849912739965,
+ "grad_norm": 0.05515526980161667,
+ "learning_rate": 8.092363420603584e-05,
+ "loss": 3.6017959117889404,
+ "step": 2792
+ },
+ {
+ "epoch": 38.795811518324605,
+ "grad_norm": 0.05573352053761482,
+ "learning_rate": 8.074670999390896e-05,
+ "loss": 3.5869383811950684,
+ "step": 2793
+ },
+ {
+ "epoch": 38.80977312390925,
+ "grad_norm": 0.05688846483826637,
+ "learning_rate": 8.056996484724342e-05,
+ "loss": 3.612272024154663,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82373472949389,
+ "grad_norm": 0.05551573261618614,
+ "learning_rate": 8.039339890843958e-05,
+ "loss": 3.5818700790405273,
+ "step": 2795
+ },
+ {
+ "epoch": 38.83769633507853,
+ "grad_norm": 0.05531308427453041,
+ "learning_rate": 8.02170123197535e-05,
+ "loss": 3.558943748474121,
+ "step": 2796
+ },
+ {
+ "epoch": 38.85165794066317,
+ "grad_norm": 0.05705123022198677,
+ "learning_rate": 8.004080522329674e-05,
+ "loss": 3.603858709335327,
+ "step": 2797
+ },
+ {
+ "epoch": 38.86561954624782,
+ "grad_norm": 0.055374935269355774,
+ "learning_rate": 7.986477776103601e-05,
+ "loss": 3.5933685302734375,
+ "step": 2798
+ },
+ {
+ "epoch": 38.87958115183246,
+ "grad_norm": 0.056353647261857986,
+ "learning_rate": 7.968893007479343e-05,
+ "loss": 3.598839282989502,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8935427574171,
+ "grad_norm": 0.05722062662243843,
+ "learning_rate": 7.951326230624658e-05,
+ "loss": 3.645443916320801,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90750436300174,
+ "grad_norm": 0.055990997701883316,
+ "learning_rate": 7.933777459692756e-05,
+ "loss": 3.5926828384399414,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92146596858639,
+ "grad_norm": 0.0561661571264267,
+ "learning_rate": 7.916246708822373e-05,
+ "loss": 3.6137099266052246,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93542757417103,
+ "grad_norm": 0.05624176934361458,
+ "learning_rate": 7.898733992137715e-05,
+ "loss": 3.597254991531372,
+ "step": 2803
+ },
+ {
+ "epoch": 38.94938917975567,
+ "grad_norm": 0.05730769410729408,
+ "learning_rate": 7.881239323748475e-05,
+ "loss": 3.595649242401123,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96335078534032,
+ "grad_norm": 0.05613584816455841,
+ "learning_rate": 7.863762717749771e-05,
+ "loss": 3.5786027908325195,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97731239092496,
+ "grad_norm": 0.05605073273181915,
+ "learning_rate": 7.846304188222189e-05,
+ "loss": 3.5973892211914062,
+ "step": 2806
+ },
+ {
+ "epoch": 38.9912739965096,
+ "grad_norm": 0.05703037977218628,
+ "learning_rate": 7.828863749231777e-05,
+ "loss": 3.614419460296631,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.042099807411432266,
+ "learning_rate": 7.811441414829958e-05,
+ "loss": 2.2424778938293457,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 0.620037853717804,
+ "eval_runtime": 59.1566,
+ "eval_samples_per_second": 41.28,
+ "eval_steps_per_second": 0.659,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01396160558464,
+ "grad_norm": 0.05103587359189987,
+ "learning_rate": 7.794037199053598e-05,
+ "loss": 3.556558132171631,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02792321116929,
+ "grad_norm": 0.049501027911901474,
+ "learning_rate": 7.776651115924959e-05,
+ "loss": 3.5505282878875732,
+ "step": 2810
+ },
+ {
+ "epoch": 39.04188481675393,
+ "grad_norm": 0.05127207562327385,
+ "learning_rate": 7.759283179451704e-05,
+ "loss": 3.5475456714630127,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05584642233857,
+ "grad_norm": 0.05074001103639603,
+ "learning_rate": 7.741933403626848e-05,
+ "loss": 3.5623722076416016,
+ "step": 2812
+ },
+ {
+ "epoch": 39.06980802792321,
+ "grad_norm": 0.05134766176342964,
+ "learning_rate": 7.724601802428805e-05,
+ "loss": 3.540435791015625,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083769633507856,
+ "grad_norm": 0.05144895985722542,
+ "learning_rate": 7.707288389821329e-05,
+ "loss": 3.6154632568359375,
+ "step": 2814
+ },
+ {
+ "epoch": 39.0977312390925,
+ "grad_norm": 0.0506402924656868,
+ "learning_rate": 7.689993179753519e-05,
+ "loss": 3.559809684753418,
+ "step": 2815
+ },
+ {
+ "epoch": 39.11169284467714,
+ "grad_norm": 0.05063097923994064,
+ "learning_rate": 7.672716186159829e-05,
+ "loss": 3.562814474105835,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12565445026178,
+ "grad_norm": 0.051375292241573334,
+ "learning_rate": 7.655457422959993e-05,
+ "loss": 3.5684101581573486,
+ "step": 2817
+ },
+ {
+ "epoch": 39.139616055846425,
+ "grad_norm": 0.0503469854593277,
+ "learning_rate": 7.638216904059122e-05,
+ "loss": 3.556791305541992,
+ "step": 2818
+ },
+ {
+ "epoch": 39.153577661431065,
+ "grad_norm": 0.050908755511045456,
+ "learning_rate": 7.620994643347559e-05,
+ "loss": 3.575982093811035,
+ "step": 2819
+ },
+ {
+ "epoch": 39.167539267015705,
+ "grad_norm": 0.049920931458473206,
+ "learning_rate": 7.60379065470098e-05,
+ "loss": 3.5394392013549805,
+ "step": 2820
+ },
+ {
+ "epoch": 39.181500872600346,
+ "grad_norm": 0.052037835121154785,
+ "learning_rate": 7.586604951980326e-05,
+ "loss": 3.5330114364624023,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19546247818499,
+ "grad_norm": 0.05024134740233421,
+ "learning_rate": 7.569437549031813e-05,
+ "loss": 3.555091142654419,
+ "step": 2822
+ },
+ {
+ "epoch": 39.20942408376963,
+ "grad_norm": 0.05144693702459335,
+ "learning_rate": 7.55228845968691e-05,
+ "loss": 3.5827114582061768,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223385689354274,
+ "grad_norm": 0.052199963480234146,
+ "learning_rate": 7.535157697762313e-05,
+ "loss": 3.559880256652832,
+ "step": 2824
+ },
+ {
+ "epoch": 39.23734729493892,
+ "grad_norm": 0.05210355296730995,
+ "learning_rate": 7.518045277059979e-05,
+ "loss": 3.5808815956115723,
+ "step": 2825
+ },
+ {
+ "epoch": 39.25130890052356,
+ "grad_norm": 0.05106557160615921,
+ "learning_rate": 7.500951211367068e-05,
+ "loss": 3.5711770057678223,
+ "step": 2826
+ },
+ {
+ "epoch": 39.2652705061082,
+ "grad_norm": 0.051606904715299606,
+ "learning_rate": 7.483875514455979e-05,
+ "loss": 3.5610320568084717,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27923211169284,
+ "grad_norm": 0.05146080255508423,
+ "learning_rate": 7.466818200084264e-05,
+ "loss": 3.558434247970581,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29319371727749,
+ "grad_norm": 0.05288682132959366,
+ "learning_rate": 7.449779281994712e-05,
+ "loss": 3.6013455390930176,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30715532286213,
+ "grad_norm": 0.05138479545712471,
+ "learning_rate": 7.43275877391528e-05,
+ "loss": 3.5789742469787598,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32111692844677,
+ "grad_norm": 0.052338190376758575,
+ "learning_rate": 7.415756689559061e-05,
+ "loss": 3.5699806213378906,
+ "step": 2831
+ },
+ {
+ "epoch": 39.33507853403141,
+ "grad_norm": 0.051882997155189514,
+ "learning_rate": 7.398773042624341e-05,
+ "loss": 3.54583740234375,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34904013961606,
+ "grad_norm": 0.05100737139582634,
+ "learning_rate": 7.381807846794532e-05,
+ "loss": 3.5269007682800293,
+ "step": 2833
+ },
+ {
+ "epoch": 39.3630017452007,
+ "grad_norm": 0.05167330801486969,
+ "learning_rate": 7.3648611157382e-05,
+ "loss": 3.547883987426758,
+ "step": 2834
+ },
+ {
+ "epoch": 39.37696335078534,
+ "grad_norm": 0.052119553089141846,
+ "learning_rate": 7.34793286310899e-05,
+ "loss": 3.570747137069702,
+ "step": 2835
+ },
+ {
+ "epoch": 39.390924956369986,
+ "grad_norm": 0.05193249508738518,
+ "learning_rate": 7.331023102545716e-05,
+ "loss": 3.5619454383850098,
+ "step": 2836
+ },
+ {
+ "epoch": 39.404886561954626,
+ "grad_norm": 0.05142345651984215,
+ "learning_rate": 7.314131847672268e-05,
+ "loss": 3.568817138671875,
+ "step": 2837
+ },
+ {
+ "epoch": 39.41884816753927,
+ "grad_norm": 0.05171925202012062,
+ "learning_rate": 7.297259112097608e-05,
+ "loss": 3.5891952514648438,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43280977312391,
+ "grad_norm": 0.051284465938806534,
+ "learning_rate": 7.280404909415801e-05,
+ "loss": 3.56705379486084,
+ "step": 2839
+ },
+ {
+ "epoch": 39.446771378708554,
+ "grad_norm": 0.05284653604030609,
+ "learning_rate": 7.263569253205973e-05,
+ "loss": 3.5296974182128906,
+ "step": 2840
+ },
+ {
+ "epoch": 39.460732984293195,
+ "grad_norm": 0.05274311453104019,
+ "learning_rate": 7.24675215703232e-05,
+ "loss": 3.5653305053710938,
+ "step": 2841
+ },
+ {
+ "epoch": 39.474694589877835,
+ "grad_norm": 0.05247433856129646,
+ "learning_rate": 7.229953634444055e-05,
+ "loss": 3.584923267364502,
+ "step": 2842
+ },
+ {
+ "epoch": 39.488656195462475,
+ "grad_norm": 0.05246812105178833,
+ "learning_rate": 7.213173698975444e-05,
+ "loss": 3.604335069656372,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50261780104712,
+ "grad_norm": 0.052823301404714584,
+ "learning_rate": 7.1964123641458e-05,
+ "loss": 3.5415382385253906,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51657940663176,
+ "grad_norm": 0.05378150939941406,
+ "learning_rate": 7.179669643459405e-05,
+ "loss": 3.5816025733947754,
+ "step": 2845
+ },
+ {
+ "epoch": 39.5305410122164,
+ "grad_norm": 0.053023822605609894,
+ "learning_rate": 7.162945550405575e-05,
+ "loss": 3.542984962463379,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544502617801044,
+ "grad_norm": 0.0532502755522728,
+ "learning_rate": 7.146240098458613e-05,
+ "loss": 3.5629453659057617,
+ "step": 2847
+ },
+ {
+ "epoch": 39.55846422338569,
+ "grad_norm": 0.05303101986646652,
+ "learning_rate": 7.129553301077801e-05,
+ "loss": 3.5725502967834473,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57242582897033,
+ "grad_norm": 0.052930813282728195,
+ "learning_rate": 7.112885171707378e-05,
+ "loss": 3.5748348236083984,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58638743455497,
+ "grad_norm": 0.0558759830892086,
+ "learning_rate": 7.096235723776562e-05,
+ "loss": 3.569337844848633,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60034904013962,
+ "grad_norm": 0.05266093835234642,
+ "learning_rate": 7.079604970699514e-05,
+ "loss": 3.579071044921875,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61431064572426,
+ "grad_norm": 0.0530242957174778,
+ "learning_rate": 7.062992925875332e-05,
+ "loss": 3.569580078125,
+ "step": 2852
+ },
+ {
+ "epoch": 39.6282722513089,
+ "grad_norm": 0.05434603616595268,
+ "learning_rate": 7.046399602688045e-05,
+ "loss": 3.590569496154785,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64223385689354,
+ "grad_norm": 0.05442731827497482,
+ "learning_rate": 7.029825014506572e-05,
+ "loss": 3.584219455718994,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65619546247819,
+ "grad_norm": 0.05413390323519707,
+ "learning_rate": 7.013269174684795e-05,
+ "loss": 3.5826475620269775,
+ "step": 2855
+ },
+ {
+ "epoch": 39.67015706806283,
+ "grad_norm": 0.053916264325380325,
+ "learning_rate": 6.996732096561422e-05,
+ "loss": 3.590533971786499,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68411867364747,
+ "grad_norm": 0.05374522507190704,
+ "learning_rate": 6.980213793460092e-05,
+ "loss": 3.5486674308776855,
+ "step": 2857
+ },
+ {
+ "epoch": 39.69808027923211,
+ "grad_norm": 0.05496302619576454,
+ "learning_rate": 6.963714278689304e-05,
+ "loss": 3.585171699523926,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712041884816756,
+ "grad_norm": 0.05460911989212036,
+ "learning_rate": 6.947233565542417e-05,
+ "loss": 3.545654296875,
+ "step": 2859
+ },
+ {
+ "epoch": 39.726003490401396,
+ "grad_norm": 0.055129971355199814,
+ "learning_rate": 6.930771667297653e-05,
+ "loss": 3.5461995601654053,
+ "step": 2860
+ },
+ {
+ "epoch": 39.73996509598604,
+ "grad_norm": 0.05326374992728233,
+ "learning_rate": 6.914328597218042e-05,
+ "loss": 3.5750136375427246,
+ "step": 2861
+ },
+ {
+ "epoch": 39.753926701570684,
+ "grad_norm": 0.05515550076961517,
+ "learning_rate": 6.8979043685515e-05,
+ "loss": 3.6057279109954834,
+ "step": 2862
+ },
+ {
+ "epoch": 39.767888307155324,
+ "grad_norm": 0.05500713735818863,
+ "learning_rate": 6.881498994530708e-05,
+ "loss": 3.5784642696380615,
+ "step": 2863
+ },
+ {
+ "epoch": 39.781849912739965,
+ "grad_norm": 0.05474477633833885,
+ "learning_rate": 6.865112488373186e-05,
+ "loss": 3.562279224395752,
+ "step": 2864
+ },
+ {
+ "epoch": 39.795811518324605,
+ "grad_norm": 0.05467807129025459,
+ "learning_rate": 6.84874486328125e-05,
+ "loss": 3.582568645477295,
+ "step": 2865
+ },
+ {
+ "epoch": 39.80977312390925,
+ "grad_norm": 0.05313905328512192,
+ "learning_rate": 6.832396132441993e-05,
+ "loss": 3.5913379192352295,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82373472949389,
+ "grad_norm": 0.05476069450378418,
+ "learning_rate": 6.816066309027309e-05,
+ "loss": 3.588045835494995,
+ "step": 2867
+ },
+ {
+ "epoch": 39.83769633507853,
+ "grad_norm": 0.05372064188122749,
+ "learning_rate": 6.799755406193815e-05,
+ "loss": 3.5555458068847656,
+ "step": 2868
+ },
+ {
+ "epoch": 39.85165794066317,
+ "grad_norm": 0.05371057987213135,
+ "learning_rate": 6.783463437082929e-05,
+ "loss": 3.5642948150634766,
+ "step": 2869
+ },
+ {
+ "epoch": 39.86561954624782,
+ "grad_norm": 0.05320272222161293,
+ "learning_rate": 6.767190414820792e-05,
+ "loss": 3.5728759765625,
+ "step": 2870
+ },
+ {
+ "epoch": 39.87958115183246,
+ "grad_norm": 0.054637160152196884,
+ "learning_rate": 6.750936352518284e-05,
+ "loss": 3.5528564453125,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8935427574171,
+ "grad_norm": 0.05465191975235939,
+ "learning_rate": 6.734701263271011e-05,
+ "loss": 3.5658421516418457,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90750436300174,
+ "grad_norm": 0.05452556163072586,
+ "learning_rate": 6.718485160159289e-05,
+ "loss": 3.5749800205230713,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92146596858639,
+ "grad_norm": 0.05434390529990196,
+ "learning_rate": 6.702288056248145e-05,
+ "loss": 3.5784506797790527,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93542757417103,
+ "grad_norm": 0.05458179861307144,
+ "learning_rate": 6.686109964587285e-05,
+ "loss": 3.5742459297180176,
+ "step": 2875
+ },
+ {
+ "epoch": 39.94938917975567,
+ "grad_norm": 0.05434086546301842,
+ "learning_rate": 6.669950898211106e-05,
+ "loss": 3.5479142665863037,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96335078534032,
+ "grad_norm": 0.054262660443782806,
+ "learning_rate": 6.653810870138676e-05,
+ "loss": 3.5872840881347656,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97731239092496,
+ "grad_norm": 0.05524756386876106,
+ "learning_rate": 6.637689893373729e-05,
+ "loss": 3.569241523742676,
+ "step": 2878
+ },
+ {
+ "epoch": 39.9912739965096,
+ "grad_norm": 0.05461305379867554,
+ "learning_rate": 6.621587980904651e-05,
+ "loss": 3.579705238342285,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.04095758497714996,
+ "learning_rate": 6.605505145704437e-05,
+ "loss": 2.2304811477661133,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 0.6212776303291321,
+ "eval_runtime": 58.55,
+ "eval_samples_per_second": 41.708,
+ "eval_steps_per_second": 0.666,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01396160558464,
+ "grad_norm": 0.0490451343357563,
+ "learning_rate": 6.58944140073077e-05,
+ "loss": 3.56355357170105,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02792321116929,
+ "grad_norm": 0.04885198175907135,
+ "learning_rate": 6.573396758925891e-05,
+ "loss": 3.563727855682373,
+ "step": 2882
+ },
+ {
+ "epoch": 40.04188481675393,
+ "grad_norm": 0.04925692081451416,
+ "learning_rate": 6.557371233216692e-05,
+ "loss": 3.5260658264160156,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05584642233857,
+ "grad_norm": 0.048543550074100494,
+ "learning_rate": 6.541364836514646e-05,
+ "loss": 3.520245313644409,
+ "step": 2884
+ },
+ {
+ "epoch": 40.06980802792321,
+ "grad_norm": 0.04830699786543846,
+ "learning_rate": 6.525377581715829e-05,
+ "loss": 3.5621728897094727,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083769633507856,
+ "grad_norm": 0.050245992839336395,
+ "learning_rate": 6.509409481700864e-05,
+ "loss": 3.5443291664123535,
+ "step": 2886
+ },
+ {
+ "epoch": 40.0977312390925,
+ "grad_norm": 0.050024352967739105,
+ "learning_rate": 6.49346054933496e-05,
+ "loss": 3.5327181816101074,
+ "step": 2887
+ },
+ {
+ "epoch": 40.11169284467714,
+ "grad_norm": 0.05024973675608635,
+ "learning_rate": 6.477530797467911e-05,
+ "loss": 3.547147035598755,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12565445026178,
+ "grad_norm": 0.049933336675167084,
+ "learning_rate": 6.461620238934006e-05,
+ "loss": 3.544022560119629,
+ "step": 2889
+ },
+ {
+ "epoch": 40.139616055846425,
+ "grad_norm": 0.0494471937417984,
+ "learning_rate": 6.445728886552109e-05,
+ "loss": 3.5606048107147217,
+ "step": 2890
+ },
+ {
+ "epoch": 40.153577661431065,
+ "grad_norm": 0.05088911950588226,
+ "learning_rate": 6.429856753125573e-05,
+ "loss": 3.5984342098236084,
+ "step": 2891
+ },
+ {
+ "epoch": 40.167539267015705,
+ "grad_norm": 0.05072201415896416,
+ "learning_rate": 6.414003851442318e-05,
+ "loss": 3.546821117401123,
+ "step": 2892
+ },
+ {
+ "epoch": 40.181500872600346,
+ "grad_norm": 0.04955511912703514,
+ "learning_rate": 6.398170194274722e-05,
+ "loss": 3.531683921813965,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19546247818499,
+ "grad_norm": 0.05047255754470825,
+ "learning_rate": 6.38235579437968e-05,
+ "loss": 3.540769100189209,
+ "step": 2894
+ },
+ {
+ "epoch": 40.20942408376963,
+ "grad_norm": 0.04991615191102028,
+ "learning_rate": 6.366560664498572e-05,
+ "loss": 3.5645358562469482,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223385689354274,
+ "grad_norm": 0.04939914494752884,
+ "learning_rate": 6.350784817357242e-05,
+ "loss": 3.500366687774658,
+ "step": 2896
+ },
+ {
+ "epoch": 40.23734729493892,
+ "grad_norm": 0.05038188025355339,
+ "learning_rate": 6.335028265666023e-05,
+ "loss": 3.547734022140503,
+ "step": 2897
+ },
+ {
+ "epoch": 40.25130890052356,
+ "grad_norm": 0.04952634498476982,
+ "learning_rate": 6.319291022119653e-05,
+ "loss": 3.525299549102783,
+ "step": 2898
+ },
+ {
+ "epoch": 40.2652705061082,
+ "grad_norm": 0.050515759736299515,
+ "learning_rate": 6.303573099397378e-05,
+ "loss": 3.5263519287109375,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27923211169284,
+ "grad_norm": 0.05072543025016785,
+ "learning_rate": 6.287874510162821e-05,
+ "loss": 3.5518910884857178,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29319371727749,
+ "grad_norm": 0.04986647143959999,
+ "learning_rate": 6.272195267064063e-05,
+ "loss": 3.5375378131866455,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30715532286213,
+ "grad_norm": 0.05132763460278511,
+ "learning_rate": 6.25653538273358e-05,
+ "loss": 3.561056613922119,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32111692844677,
+ "grad_norm": 0.050235629081726074,
+ "learning_rate": 6.240894869788267e-05,
+ "loss": 3.5394344329833984,
+ "step": 2903
+ },
+ {
+ "epoch": 40.33507853403141,
+ "grad_norm": 0.0514477975666523,
+ "learning_rate": 6.225273740829404e-05,
+ "loss": 3.5648441314697266,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34904013961606,
+ "grad_norm": 0.05128472298383713,
+ "learning_rate": 6.209672008442635e-05,
+ "loss": 3.5410401821136475,
+ "step": 2905
+ },
+ {
+ "epoch": 40.3630017452007,
+ "grad_norm": 0.05111101642251015,
+ "learning_rate": 6.19408968519801e-05,
+ "loss": 3.549811840057373,
+ "step": 2906
+ },
+ {
+ "epoch": 40.37696335078534,
+ "grad_norm": 0.05159356817603111,
+ "learning_rate": 6.178526783649916e-05,
+ "loss": 3.5218734741210938,
+ "step": 2907
+ },
+ {
+ "epoch": 40.390924956369986,
+ "grad_norm": 0.049884531646966934,
+ "learning_rate": 6.162983316337109e-05,
+ "loss": 3.5478882789611816,
+ "step": 2908
+ },
+ {
+ "epoch": 40.404886561954626,
+ "grad_norm": 0.051223963499069214,
+ "learning_rate": 6.147459295782676e-05,
+ "loss": 3.5509533882141113,
+ "step": 2909
+ },
+ {
+ "epoch": 40.41884816753927,
+ "grad_norm": 0.05251265689730644,
+ "learning_rate": 6.131954734494045e-05,
+ "loss": 3.5500032901763916,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43280977312391,
+ "grad_norm": 0.05150502547621727,
+ "learning_rate": 6.11646964496296e-05,
+ "loss": 3.551039218902588,
+ "step": 2911
+ },
+ {
+ "epoch": 40.446771378708554,
+ "grad_norm": 0.050975680351257324,
+ "learning_rate": 6.101004039665471e-05,
+ "loss": 3.5218753814697266,
+ "step": 2912
+ },
+ {
+ "epoch": 40.460732984293195,
+ "grad_norm": 0.052832040935754776,
+ "learning_rate": 6.085557931061937e-05,
+ "loss": 3.5624842643737793,
+ "step": 2913
+ },
+ {
+ "epoch": 40.474694589877835,
+ "grad_norm": 0.051763877272605896,
+ "learning_rate": 6.070131331597015e-05,
+ "loss": 3.572378158569336,
+ "step": 2914
+ },
+ {
+ "epoch": 40.488656195462475,
+ "grad_norm": 0.04998628795146942,
+ "learning_rate": 6.054724253699636e-05,
+ "loss": 3.5298023223876953,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50261780104712,
+ "grad_norm": 0.05136828497052193,
+ "learning_rate": 6.03933670978301e-05,
+ "loss": 3.541879653930664,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51657940663176,
+ "grad_norm": 0.05399318039417267,
+ "learning_rate": 6.0239687122445826e-05,
+ "loss": 3.5703022480010986,
+ "step": 2917
+ },
+ {
+ "epoch": 40.5305410122164,
+ "grad_norm": 0.050659872591495514,
+ "learning_rate": 6.0086202734661e-05,
+ "loss": 3.555227518081665,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544502617801044,
+ "grad_norm": 0.051574256271123886,
+ "learning_rate": 5.9932914058135004e-05,
+ "loss": 3.549437999725342,
+ "step": 2919
+ },
+ {
+ "epoch": 40.55846422338569,
+ "grad_norm": 0.05199757218360901,
+ "learning_rate": 5.977982121636984e-05,
+ "loss": 3.564350128173828,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57242582897033,
+ "grad_norm": 0.05227932706475258,
+ "learning_rate": 5.962692433270962e-05,
+ "loss": 3.538482666015625,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58638743455497,
+ "grad_norm": 0.05216033384203911,
+ "learning_rate": 5.9474223530340634e-05,
+ "loss": 3.5592684745788574,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60034904013962,
+ "grad_norm": 0.051728200167417526,
+ "learning_rate": 5.932171893229124e-05,
+ "loss": 3.516632556915283,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61431064572426,
+ "grad_norm": 0.05290914699435234,
+ "learning_rate": 5.916941066143137e-05,
+ "loss": 3.5496959686279297,
+ "step": 2924
+ },
+ {
+ "epoch": 40.6282722513089,
+ "grad_norm": 0.053392428904771805,
+ "learning_rate": 5.9017298840473364e-05,
+ "loss": 3.548393964767456,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64223385689354,
+ "grad_norm": 0.052064698189496994,
+ "learning_rate": 5.8865383591970775e-05,
+ "loss": 3.5496573448181152,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65619546247819,
+ "grad_norm": 0.05260622501373291,
+ "learning_rate": 5.8713665038319e-05,
+ "loss": 3.556190252304077,
+ "step": 2927
+ },
+ {
+ "epoch": 40.67015706806283,
+ "grad_norm": 0.052858006209135056,
+ "learning_rate": 5.856214330175498e-05,
+ "loss": 3.550811290740967,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68411867364747,
+ "grad_norm": 0.052420660853385925,
+ "learning_rate": 5.841081850435704e-05,
+ "loss": 3.536261796951294,
+ "step": 2929
+ },
+ {
+ "epoch": 40.69808027923211,
+ "grad_norm": 0.053714022040367126,
+ "learning_rate": 5.825969076804488e-05,
+ "loss": 3.587533473968506,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712041884816756,
+ "grad_norm": 0.05208775773644447,
+ "learning_rate": 5.810876021457925e-05,
+ "loss": 3.540727376937866,
+ "step": 2931
+ },
+ {
+ "epoch": 40.726003490401396,
+ "grad_norm": 0.05275597423315048,
+ "learning_rate": 5.795802696556222e-05,
+ "loss": 3.574899196624756,
+ "step": 2932
+ },
+ {
+ "epoch": 40.73996509598604,
+ "grad_norm": 0.05299631133675575,
+ "learning_rate": 5.78074911424369e-05,
+ "loss": 3.530336380004883,
+ "step": 2933
+ },
+ {
+ "epoch": 40.753926701570684,
+ "grad_norm": 0.054256338626146317,
+ "learning_rate": 5.765715286648734e-05,
+ "loss": 3.5379605293273926,
+ "step": 2934
+ },
+ {
+ "epoch": 40.767888307155324,
+ "grad_norm": 0.05158054083585739,
+ "learning_rate": 5.7507012258838154e-05,
+ "loss": 3.562894821166992,
+ "step": 2935
+ },
+ {
+ "epoch": 40.781849912739965,
+ "grad_norm": 0.05369063839316368,
+ "learning_rate": 5.7357069440455095e-05,
+ "loss": 3.529047966003418,
+ "step": 2936
+ },
+ {
+ "epoch": 40.795811518324605,
+ "grad_norm": 0.05377575755119324,
+ "learning_rate": 5.7207324532144454e-05,
+ "loss": 3.5563199520111084,
+ "step": 2937
+ },
+ {
+ "epoch": 40.80977312390925,
+ "grad_norm": 0.05389560014009476,
+ "learning_rate": 5.7057777654552785e-05,
+ "loss": 3.5682358741760254,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82373472949389,
+ "grad_norm": 0.053750719875097275,
+ "learning_rate": 5.690842892816741e-05,
+ "loss": 3.5546717643737793,
+ "step": 2939
+ },
+ {
+ "epoch": 40.83769633507853,
+ "grad_norm": 0.053277388215065,
+ "learning_rate": 5.675927847331593e-05,
+ "loss": 3.5211615562438965,
+ "step": 2940
+ },
+ {
+ "epoch": 40.85165794066317,
+ "grad_norm": 0.05360420420765877,
+ "learning_rate": 5.661032641016619e-05,
+ "loss": 3.5586609840393066,
+ "step": 2941
+ },
+ {
+ "epoch": 40.86561954624782,
+ "grad_norm": 0.05353206396102905,
+ "learning_rate": 5.6461572858726086e-05,
+ "loss": 3.5891380310058594,
+ "step": 2942
+ },
+ {
+ "epoch": 40.87958115183246,
+ "grad_norm": 0.05361996591091156,
+ "learning_rate": 5.6313017938843595e-05,
+ "loss": 3.5477938652038574,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8935427574171,
+ "grad_norm": 0.053566232323646545,
+ "learning_rate": 5.6164661770207026e-05,
+ "loss": 3.560500144958496,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90750436300174,
+ "grad_norm": 0.05357549339532852,
+ "learning_rate": 5.6016504472344e-05,
+ "loss": 3.557918071746826,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92146596858639,
+ "grad_norm": 0.05396264046430588,
+ "learning_rate": 5.5868546164622284e-05,
+ "loss": 3.5511131286621094,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93542757417103,
+ "grad_norm": 0.05372125282883644,
+ "learning_rate": 5.572078696624917e-05,
+ "loss": 3.529221534729004,
+ "step": 2947
+ },
+ {
+ "epoch": 40.94938917975567,
+ "grad_norm": 0.05311017110943794,
+ "learning_rate": 5.557322699627168e-05,
+ "loss": 3.5863747596740723,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96335078534032,
+ "grad_norm": 0.05271926149725914,
+ "learning_rate": 5.542586637357607e-05,
+ "loss": 3.5291237831115723,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97731239092496,
+ "grad_norm": 0.05377885699272156,
+ "learning_rate": 5.527870521688817e-05,
+ "loss": 3.5661916732788086,
+ "step": 2950
+ },
+ {
+ "epoch": 40.9912739965096,
+ "grad_norm": 0.05412346124649048,
+ "learning_rate": 5.513174364477311e-05,
+ "loss": 3.5414018630981445,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.041567444801330566,
+ "learning_rate": 5.498498177563518e-05,
+ "loss": 2.210968494415283,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 0.6227477192878723,
+ "eval_runtime": 59.639,
+ "eval_samples_per_second": 40.946,
+ "eval_steps_per_second": 0.654,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01396160558464,
+ "grad_norm": 0.04824776574969292,
+ "learning_rate": 5.4838419727717716e-05,
+ "loss": 3.5330162048339844,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02792321116929,
+ "grad_norm": 0.04834216460585594,
+ "learning_rate": 5.4692057619103144e-05,
+ "loss": 3.506112575531006,
+ "step": 2954
+ },
+ {
+ "epoch": 41.04188481675393,
+ "grad_norm": 0.048684537410736084,
+ "learning_rate": 5.454589556771286e-05,
+ "loss": 3.552931785583496,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05584642233857,
+ "grad_norm": 0.04952389374375343,
+ "learning_rate": 5.439993369130681e-05,
+ "loss": 3.522508144378662,
+ "step": 2956
+ },
+ {
+ "epoch": 41.06980802792321,
+ "grad_norm": 0.047509536147117615,
+ "learning_rate": 5.4254172107483924e-05,
+ "loss": 3.521543502807617,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083769633507856,
+ "grad_norm": 0.04766285791993141,
+ "learning_rate": 5.410861093368169e-05,
+ "loss": 3.5325450897216797,
+ "step": 2958
+ },
+ {
+ "epoch": 41.0977312390925,
+ "grad_norm": 0.04850609600543976,
+ "learning_rate": 5.3963250287176094e-05,
+ "loss": 3.5616042613983154,
+ "step": 2959
+ },
+ {
+ "epoch": 41.11169284467714,
+ "grad_norm": 0.04834820702672005,
+ "learning_rate": 5.381809028508168e-05,
+ "loss": 3.5167665481567383,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12565445026178,
+ "grad_norm": 0.04798751696944237,
+ "learning_rate": 5.367313104435104e-05,
+ "loss": 3.5004749298095703,
+ "step": 2961
+ },
+ {
+ "epoch": 41.139616055846425,
+ "grad_norm": 0.047878291457891464,
+ "learning_rate": 5.352837268177544e-05,
+ "loss": 3.5392799377441406,
+ "step": 2962
+ },
+ {
+ "epoch": 41.153577661431065,
+ "grad_norm": 0.048691265285015106,
+ "learning_rate": 5.3383815313983916e-05,
+ "loss": 3.525770664215088,
+ "step": 2963
+ },
+ {
+ "epoch": 41.167539267015705,
+ "grad_norm": 0.048624299466609955,
+ "learning_rate": 5.323945905744379e-05,
+ "loss": 3.5122787952423096,
+ "step": 2964
+ },
+ {
+ "epoch": 41.181500872600346,
+ "grad_norm": 0.0488092415034771,
+ "learning_rate": 5.3095304028460284e-05,
+ "loss": 3.50136661529541,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19546247818499,
+ "grad_norm": 0.04929783567786217,
+ "learning_rate": 5.2951350343176495e-05,
+ "loss": 3.51039981842041,
+ "step": 2966
+ },
+ {
+ "epoch": 41.20942408376963,
+ "grad_norm": 0.049221813678741455,
+ "learning_rate": 5.2807598117573384e-05,
+ "loss": 3.530416965484619,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223385689354274,
+ "grad_norm": 0.04767163097858429,
+ "learning_rate": 5.266404746746939e-05,
+ "loss": 3.5467047691345215,
+ "step": 2968
+ },
+ {
+ "epoch": 41.23734729493892,
+ "grad_norm": 0.048979729413986206,
+ "learning_rate": 5.252069850852072e-05,
+ "loss": 3.557248115539551,
+ "step": 2969
+ },
+ {
+ "epoch": 41.25130890052356,
+ "grad_norm": 0.048910439014434814,
+ "learning_rate": 5.237755135622107e-05,
+ "loss": 3.5098860263824463,
+ "step": 2970
+ },
+ {
+ "epoch": 41.2652705061082,
+ "grad_norm": 0.048609018325805664,
+ "learning_rate": 5.223460612590159e-05,
+ "loss": 3.5389151573181152,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27923211169284,
+ "grad_norm": 0.04871147871017456,
+ "learning_rate": 5.209186293273044e-05,
+ "loss": 3.5385420322418213,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29319371727749,
+ "grad_norm": 0.04857845976948738,
+ "learning_rate": 5.1949321891713446e-05,
+ "loss": 3.507431983947754,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30715532286213,
+ "grad_norm": 0.05005535110831261,
+ "learning_rate": 5.180698311769338e-05,
+ "loss": 3.5485854148864746,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32111692844677,
+ "grad_norm": 0.05051431804895401,
+ "learning_rate": 5.166484672534983e-05,
+ "loss": 3.5140175819396973,
+ "step": 2975
+ },
+ {
+ "epoch": 41.33507853403141,
+ "grad_norm": 0.04972303658723831,
+ "learning_rate": 5.152291282919961e-05,
+ "loss": 3.541937828063965,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34904013961606,
+ "grad_norm": 0.04871612414717674,
+ "learning_rate": 5.138118154359632e-05,
+ "loss": 3.543015956878662,
+ "step": 2977
+ },
+ {
+ "epoch": 41.3630017452007,
+ "grad_norm": 0.05032794922590256,
+ "learning_rate": 5.1239652982730335e-05,
+ "loss": 3.4958419799804688,
+ "step": 2978
+ },
+ {
+ "epoch": 41.37696335078534,
+ "grad_norm": 0.05026672035455704,
+ "learning_rate": 5.1098327260628466e-05,
+ "loss": 3.519773244857788,
+ "step": 2979
+ },
+ {
+ "epoch": 41.390924956369986,
+ "grad_norm": 0.04959705099463463,
+ "learning_rate": 5.09572044911545e-05,
+ "loss": 3.508209466934204,
+ "step": 2980
+ },
+ {
+ "epoch": 41.404886561954626,
+ "grad_norm": 0.04974748566746712,
+ "learning_rate": 5.0816284788008527e-05,
+ "loss": 3.5200767517089844,
+ "step": 2981
+ },
+ {
+ "epoch": 41.41884816753927,
+ "grad_norm": 0.050430137664079666,
+ "learning_rate": 5.067556826472684e-05,
+ "loss": 3.528503656387329,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43280977312391,
+ "grad_norm": 0.05002054572105408,
+ "learning_rate": 5.053505503468228e-05,
+ "loss": 3.544261932373047,
+ "step": 2983
+ },
+ {
+ "epoch": 41.446771378708554,
+ "grad_norm": 0.050833702087402344,
+ "learning_rate": 5.0394745211083785e-05,
+ "loss": 3.528820276260376,
+ "step": 2984
+ },
+ {
+ "epoch": 41.460732984293195,
+ "grad_norm": 0.05177270993590355,
+ "learning_rate": 5.025463890697655e-05,
+ "loss": 3.5320916175842285,
+ "step": 2985
+ },
+ {
+ "epoch": 41.474694589877835,
+ "grad_norm": 0.05068309232592583,
+ "learning_rate": 5.011473623524159e-05,
+ "loss": 3.535667896270752,
+ "step": 2986
+ },
+ {
+ "epoch": 41.488656195462475,
+ "grad_norm": 0.05038442835211754,
+ "learning_rate": 4.9975037308595864e-05,
+ "loss": 3.544569969177246,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50261780104712,
+ "grad_norm": 0.051365263760089874,
+ "learning_rate": 4.983554223959257e-05,
+ "loss": 3.5355050563812256,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51657940663176,
+ "grad_norm": 0.05044831335544586,
+ "learning_rate": 4.9696251140620134e-05,
+ "loss": 3.511427879333496,
+ "step": 2989
+ },
+ {
+ "epoch": 41.5305410122164,
+ "grad_norm": 0.050490397959947586,
+ "learning_rate": 4.9557164123902924e-05,
+ "loss": 3.544853687286377,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544502617801044,
+ "grad_norm": 0.05034386366605759,
+ "learning_rate": 4.9418281301500844e-05,
+ "loss": 3.5631227493286133,
+ "step": 2991
+ },
+ {
+ "epoch": 41.55846422338569,
+ "grad_norm": 0.0505383126437664,
+ "learning_rate": 4.9279602785309365e-05,
+ "loss": 3.5400047302246094,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57242582897033,
+ "grad_norm": 0.05098932981491089,
+ "learning_rate": 4.914112868705911e-05,
+ "loss": 3.5390548706054688,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58638743455497,
+ "grad_norm": 0.05151797831058502,
+ "learning_rate": 4.900285911831624e-05,
+ "loss": 3.535597324371338,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60034904013962,
+ "grad_norm": 0.05120394378900528,
+ "learning_rate": 4.886479419048201e-05,
+ "loss": 3.5225820541381836,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61431064572426,
+ "grad_norm": 0.05033888295292854,
+ "learning_rate": 4.872693401479292e-05,
+ "loss": 3.4912185668945312,
+ "step": 2996
+ },
+ {
+ "epoch": 41.6282722513089,
+ "grad_norm": 0.05037662014365196,
+ "learning_rate": 4.858927870232039e-05,
+ "loss": 3.5338728427886963,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64223385689354,
+ "grad_norm": 0.05047066509723663,
+ "learning_rate": 4.845182836397068e-05,
+ "loss": 3.537113666534424,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65619546247819,
+ "grad_norm": 0.051437340676784515,
+ "learning_rate": 4.831458311048532e-05,
+ "loss": 3.537376880645752,
+ "step": 2999
+ },
+ {
+ "epoch": 41.67015706806283,
+ "grad_norm": 0.05054743215441704,
+ "learning_rate": 4.817754305244008e-05,
+ "loss": 3.558231830596924,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68411867364747,
+ "grad_norm": 0.050861239433288574,
+ "learning_rate": 4.804070830024577e-05,
+ "loss": 3.5442352294921875,
+ "step": 3001
+ },
+ {
+ "epoch": 41.69808027923211,
+ "grad_norm": 0.05110339820384979,
+ "learning_rate": 4.7904078964147654e-05,
+ "loss": 3.537139892578125,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712041884816756,
+ "grad_norm": 0.05177166685461998,
+ "learning_rate": 4.776765515422557e-05,
+ "loss": 3.5539965629577637,
+ "step": 3003
+ },
+ {
+ "epoch": 41.726003490401396,
+ "grad_norm": 0.051353856921195984,
+ "learning_rate": 4.763143698039371e-05,
+ "loss": 3.5276026725769043,
+ "step": 3004
+ },
+ {
+ "epoch": 41.73996509598604,
+ "grad_norm": 0.05237841606140137,
+ "learning_rate": 4.749542455240043e-05,
+ "loss": 3.5436954498291016,
+ "step": 3005
+ },
+ {
+ "epoch": 41.753926701570684,
+ "grad_norm": 0.051556285470724106,
+ "learning_rate": 4.735961797982876e-05,
+ "loss": 3.525407552719116,
+ "step": 3006
+ },
+ {
+ "epoch": 41.767888307155324,
+ "grad_norm": 0.0522376224398613,
+ "learning_rate": 4.7224017372095414e-05,
+ "loss": 3.5100879669189453,
+ "step": 3007
+ },
+ {
+ "epoch": 41.781849912739965,
+ "grad_norm": 0.05480589717626572,
+ "learning_rate": 4.708862283845143e-05,
+ "loss": 3.5373239517211914,
+ "step": 3008
+ },
+ {
+ "epoch": 41.795811518324605,
+ "grad_norm": 0.05243408679962158,
+ "learning_rate": 4.695343448798163e-05,
+ "loss": 3.564474105834961,
+ "step": 3009
+ },
+ {
+ "epoch": 41.80977312390925,
+ "grad_norm": 0.05195145308971405,
+ "learning_rate": 4.681845242960492e-05,
+ "loss": 3.5219478607177734,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82373472949389,
+ "grad_norm": 0.051901329308748245,
+ "learning_rate": 4.668367677207398e-05,
+ "loss": 3.5563879013061523,
+ "step": 3011
+ },
+ {
+ "epoch": 41.83769633507853,
+ "grad_norm": 0.05103657394647598,
+ "learning_rate": 4.654910762397499e-05,
+ "loss": 3.5318007469177246,
+ "step": 3012
+ },
+ {
+ "epoch": 41.85165794066317,
+ "grad_norm": 0.05273298919200897,
+ "learning_rate": 4.6414745093727913e-05,
+ "loss": 3.5337777137756348,
+ "step": 3013
+ },
+ {
+ "epoch": 41.86561954624782,
+ "grad_norm": 0.05128652974963188,
+ "learning_rate": 4.6280589289586255e-05,
+ "loss": 3.523869037628174,
+ "step": 3014
+ },
+ {
+ "epoch": 41.87958115183246,
+ "grad_norm": 0.051192507147789,
+ "learning_rate": 4.614664031963692e-05,
+ "loss": 3.5141286849975586,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8935427574171,
+ "grad_norm": 0.052560918033123016,
+ "learning_rate": 4.601289829180004e-05,
+ "loss": 3.542026996612549,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90750436300174,
+ "grad_norm": 0.05215048789978027,
+ "learning_rate": 4.587936331382934e-05,
+ "loss": 3.5189595222473145,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92146596858639,
+ "grad_norm": 0.05149101838469505,
+ "learning_rate": 4.574603549331151e-05,
+ "loss": 3.559493064880371,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93542757417103,
+ "grad_norm": 0.05164817348122597,
+ "learning_rate": 4.561291493766625e-05,
+ "loss": 3.553772211074829,
+ "step": 3019
+ },
+ {
+ "epoch": 41.94938917975567,
+ "grad_norm": 0.05225958675146103,
+ "learning_rate": 4.5480001754146455e-05,
+ "loss": 3.5521295070648193,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96335078534032,
+ "grad_norm": 0.051618315279483795,
+ "learning_rate": 4.5347296049837875e-05,
+ "loss": 3.5303773880004883,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97731239092496,
+ "grad_norm": 0.051267798990011215,
+ "learning_rate": 4.52147979316592e-05,
+ "loss": 3.5165226459503174,
+ "step": 3022
+ },
+ {
+ "epoch": 41.9912739965096,
+ "grad_norm": 0.05264551192522049,
+ "learning_rate": 4.5082507506361574e-05,
+ "loss": 3.5303573608398438,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.03939991816878319,
+ "learning_rate": 4.495042488052904e-05,
+ "loss": 2.205373764038086,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 0.6238990426063538,
+ "eval_runtime": 59.3761,
+ "eval_samples_per_second": 41.128,
+ "eval_steps_per_second": 0.657,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01396160558464,
+ "grad_norm": 0.04582864046096802,
+ "learning_rate": 4.481855016057839e-05,
+ "loss": 3.55916690826416,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02792321116929,
+ "grad_norm": 0.045121654868125916,
+ "learning_rate": 4.468688345275848e-05,
+ "loss": 3.525993585586548,
+ "step": 3026
+ },
+ {
+ "epoch": 42.04188481675393,
+ "grad_norm": 0.045851755887269974,
+ "learning_rate": 4.455542486315086e-05,
+ "loss": 3.536099433898926,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05584642233857,
+ "grad_norm": 0.04772236570715904,
+ "learning_rate": 4.442417449766938e-05,
+ "loss": 3.4842875003814697,
+ "step": 3028
+ },
+ {
+ "epoch": 42.06980802792321,
+ "grad_norm": 0.04591905325651169,
+ "learning_rate": 4.429313246206014e-05,
+ "loss": 3.4977431297302246,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083769633507856,
+ "grad_norm": 0.04562445729970932,
+ "learning_rate": 4.416229886190119e-05,
+ "loss": 3.5109801292419434,
+ "step": 3030
+ },
+ {
+ "epoch": 42.0977312390925,
+ "grad_norm": 0.046725720167160034,
+ "learning_rate": 4.403167380260281e-05,
+ "loss": 3.507308006286621,
+ "step": 3031
+ },
+ {
+ "epoch": 42.11169284467714,
+ "grad_norm": 0.047091979533433914,
+ "learning_rate": 4.3901257389407475e-05,
+ "loss": 3.5205769538879395,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12565445026178,
+ "grad_norm": 0.04644174128770828,
+ "learning_rate": 4.3771049727389075e-05,
+ "loss": 3.519360065460205,
+ "step": 3033
+ },
+ {
+ "epoch": 42.139616055846425,
+ "grad_norm": 0.0471586212515831,
+ "learning_rate": 4.364105092145377e-05,
+ "loss": 3.511544704437256,
+ "step": 3034
+ },
+ {
+ "epoch": 42.153577661431065,
+ "grad_norm": 0.04756023734807968,
+ "learning_rate": 4.3511261076339e-05,
+ "loss": 3.5082921981811523,
+ "step": 3035
+ },
+ {
+ "epoch": 42.167539267015705,
+ "grad_norm": 0.04774578660726547,
+ "learning_rate": 4.3381680296614394e-05,
+ "loss": 3.514894485473633,
+ "step": 3036
+ },
+ {
+ "epoch": 42.181500872600346,
+ "grad_norm": 0.047490909695625305,
+ "learning_rate": 4.3252308686680626e-05,
+ "loss": 3.5040950775146484,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19546247818499,
+ "grad_norm": 0.04813152179121971,
+ "learning_rate": 4.3123146350770146e-05,
+ "loss": 3.5205535888671875,
+ "step": 3038
+ },
+ {
+ "epoch": 42.20942408376963,
+ "grad_norm": 0.047403670847415924,
+ "learning_rate": 4.2994193392946726e-05,
+ "loss": 3.4987292289733887,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223385689354274,
+ "grad_norm": 0.047732383012771606,
+ "learning_rate": 4.286544991710541e-05,
+ "loss": 3.5132713317871094,
+ "step": 3040
+ },
+ {
+ "epoch": 42.23734729493892,
+ "grad_norm": 0.0471004955470562,
+ "learning_rate": 4.2736916026972576e-05,
+ "loss": 3.4943442344665527,
+ "step": 3041
+ },
+ {
+ "epoch": 42.25130890052356,
+ "grad_norm": 0.04769458621740341,
+ "learning_rate": 4.260859182610542e-05,
+ "loss": 3.488661527633667,
+ "step": 3042
+ },
+ {
+ "epoch": 42.2652705061082,
+ "grad_norm": 0.04852530360221863,
+ "learning_rate": 4.2480477417892776e-05,
+ "loss": 3.502333641052246,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27923211169284,
+ "grad_norm": 0.047681331634521484,
+ "learning_rate": 4.235257290555381e-05,
+ "loss": 3.5339419841766357,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29319371727749,
+ "grad_norm": 0.04790053516626358,
+ "learning_rate": 4.222487839213903e-05,
+ "loss": 3.531029462814331,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30715532286213,
+ "grad_norm": 0.04872751981019974,
+ "learning_rate": 4.209739398052956e-05,
+ "loss": 3.5269131660461426,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32111692844677,
+ "grad_norm": 0.0475989505648613,
+ "learning_rate": 4.1970119773437316e-05,
+ "loss": 3.520803213119507,
+ "step": 3047
+ },
+ {
+ "epoch": 42.33507853403141,
+ "grad_norm": 0.047820597887039185,
+ "learning_rate": 4.184305587340483e-05,
+ "loss": 3.516744613647461,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34904013961606,
+ "grad_norm": 0.046507153660058975,
+ "learning_rate": 4.171620238280511e-05,
+ "loss": 3.4996590614318848,
+ "step": 3049
+ },
+ {
+ "epoch": 42.3630017452007,
+ "grad_norm": 0.047835614532232285,
+ "learning_rate": 4.158955940384179e-05,
+ "loss": 3.5125174522399902,
+ "step": 3050
+ },
+ {
+ "epoch": 42.37696335078534,
+ "grad_norm": 0.048144467175006866,
+ "learning_rate": 4.146312703854881e-05,
+ "loss": 3.5353198051452637,
+ "step": 3051
+ },
+ {
+ "epoch": 42.390924956369986,
+ "grad_norm": 0.04883087798953056,
+ "learning_rate": 4.133690538879046e-05,
+ "loss": 3.518767833709717,
+ "step": 3052
+ },
+ {
+ "epoch": 42.404886561954626,
+ "grad_norm": 0.047930024564266205,
+ "learning_rate": 4.1210894556261226e-05,
+ "loss": 3.5251386165618896,
+ "step": 3053
+ },
+ {
+ "epoch": 42.41884816753927,
+ "grad_norm": 0.04800357297062874,
+ "learning_rate": 4.108509464248578e-05,
+ "loss": 3.5164432525634766,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43280977312391,
+ "grad_norm": 0.048447951674461365,
+ "learning_rate": 4.095950574881891e-05,
+ "loss": 3.4852211475372314,
+ "step": 3055
+ },
+ {
+ "epoch": 42.446771378708554,
+ "grad_norm": 0.04907943308353424,
+ "learning_rate": 4.083412797644514e-05,
+ "loss": 3.513625383377075,
+ "step": 3056
+ },
+ {
+ "epoch": 42.460732984293195,
+ "grad_norm": 0.04915059357881546,
+ "learning_rate": 4.0708961426379214e-05,
+ "loss": 3.5324220657348633,
+ "step": 3057
+ },
+ {
+ "epoch": 42.474694589877835,
+ "grad_norm": 0.04837046191096306,
+ "learning_rate": 4.0584006199465484e-05,
+ "loss": 3.513761520385742,
+ "step": 3058
+ },
+ {
+ "epoch": 42.488656195462475,
+ "grad_norm": 0.04837717115879059,
+ "learning_rate": 4.0459262396378165e-05,
+ "loss": 3.4778947830200195,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50261780104712,
+ "grad_norm": 0.048395294696092606,
+ "learning_rate": 4.033473011762116e-05,
+ "loss": 3.5122289657592773,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51657940663176,
+ "grad_norm": 0.04952006787061691,
+ "learning_rate": 4.0210409463527656e-05,
+ "loss": 3.5174784660339355,
+ "step": 3061
+ },
+ {
+ "epoch": 42.5305410122164,
+ "grad_norm": 0.050250094383955,
+ "learning_rate": 4.008630053426082e-05,
+ "loss": 3.481564521789551,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544502617801044,
+ "grad_norm": 0.04852081835269928,
+ "learning_rate": 3.996240342981279e-05,
+ "loss": 3.5288257598876953,
+ "step": 3063
+ },
+ {
+ "epoch": 42.55846422338569,
+ "grad_norm": 0.048682864755392075,
+ "learning_rate": 3.983871825000529e-05,
+ "loss": 3.517641067504883,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57242582897033,
+ "grad_norm": 0.049304574728012085,
+ "learning_rate": 3.971524509448925e-05,
+ "loss": 3.5146350860595703,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58638743455497,
+ "grad_norm": 0.049619562923908234,
+ "learning_rate": 3.959198406274472e-05,
+ "loss": 3.5137906074523926,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60034904013962,
+ "grad_norm": 0.04923011362552643,
+ "learning_rate": 3.946893525408097e-05,
+ "loss": 3.5143535137176514,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61431064572426,
+ "grad_norm": 0.0497874915599823,
+ "learning_rate": 3.934609876763604e-05,
+ "loss": 3.5229015350341797,
+ "step": 3068
+ },
+ {
+ "epoch": 42.6282722513089,
+ "grad_norm": 0.050324805080890656,
+ "learning_rate": 3.922347470237728e-05,
+ "loss": 3.5317938327789307,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64223385689354,
+ "grad_norm": 0.049399957060813904,
+ "learning_rate": 3.910106315710053e-05,
+ "loss": 3.535451650619507,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65619546247819,
+ "grad_norm": 0.04937320575118065,
+ "learning_rate": 3.897886423043061e-05,
+ "loss": 3.510342836380005,
+ "step": 3071
+ },
+ {
+ "epoch": 42.67015706806283,
+ "grad_norm": 0.04943464323878288,
+ "learning_rate": 3.8856878020820975e-05,
+ "loss": 3.53397798538208,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68411867364747,
+ "grad_norm": 0.04982450604438782,
+ "learning_rate": 3.873510462655375e-05,
+ "loss": 3.5372724533081055,
+ "step": 3073
+ },
+ {
+ "epoch": 42.69808027923211,
+ "grad_norm": 0.04879437014460564,
+ "learning_rate": 3.861354414573954e-05,
+ "loss": 3.5235369205474854,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712041884816756,
+ "grad_norm": 0.049403633922338486,
+ "learning_rate": 3.849219667631735e-05,
+ "loss": 3.523263931274414,
+ "step": 3075
+ },
+ {
+ "epoch": 42.726003490401396,
+ "grad_norm": 0.04999256134033203,
+ "learning_rate": 3.8371062316054764e-05,
+ "loss": 3.5183444023132324,
+ "step": 3076
+ },
+ {
+ "epoch": 42.73996509598604,
+ "grad_norm": 0.04949156567454338,
+ "learning_rate": 3.8250141162547466e-05,
+ "loss": 3.5372426509857178,
+ "step": 3077
+ },
+ {
+ "epoch": 42.753926701570684,
+ "grad_norm": 0.05010758712887764,
+ "learning_rate": 3.812943331321956e-05,
+ "loss": 3.494075298309326,
+ "step": 3078
+ },
+ {
+ "epoch": 42.767888307155324,
+ "grad_norm": 0.04966307431459427,
+ "learning_rate": 3.8008938865322955e-05,
+ "loss": 3.518535614013672,
+ "step": 3079
+ },
+ {
+ "epoch": 42.781849912739965,
+ "grad_norm": 0.050056494772434235,
+ "learning_rate": 3.788865791593813e-05,
+ "loss": 3.5020852088928223,
+ "step": 3080
+ },
+ {
+ "epoch": 42.795811518324605,
+ "grad_norm": 0.04934854060411453,
+ "learning_rate": 3.7768590561973064e-05,
+ "loss": 3.5209922790527344,
+ "step": 3081
+ },
+ {
+ "epoch": 42.80977312390925,
+ "grad_norm": 0.049946680665016174,
+ "learning_rate": 3.764873690016394e-05,
+ "loss": 3.517817974090576,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82373472949389,
+ "grad_norm": 0.04958394914865494,
+ "learning_rate": 3.7529097027074654e-05,
+ "loss": 3.5093369483947754,
+ "step": 3083
+ },
+ {
+ "epoch": 42.83769633507853,
+ "grad_norm": 0.04957697167992592,
+ "learning_rate": 3.7409671039096936e-05,
+ "loss": 3.4855117797851562,
+ "step": 3084
+ },
+ {
+ "epoch": 42.85165794066317,
+ "grad_norm": 0.05044497177004814,
+ "learning_rate": 3.7290459032450184e-05,
+ "loss": 3.518657684326172,
+ "step": 3085
+ },
+ {
+ "epoch": 42.86561954624782,
+ "grad_norm": 0.049479082226753235,
+ "learning_rate": 3.717146110318123e-05,
+ "loss": 3.5081276893615723,
+ "step": 3086
+ },
+ {
+ "epoch": 42.87958115183246,
+ "grad_norm": 0.0499165914952755,
+ "learning_rate": 3.7052677347164614e-05,
+ "loss": 3.525416374206543,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8935427574171,
+ "grad_norm": 0.05002507194876671,
+ "learning_rate": 3.693410786010231e-05,
+ "loss": 3.5287156105041504,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90750436300174,
+ "grad_norm": 0.04990137368440628,
+ "learning_rate": 3.6815752737523537e-05,
+ "loss": 3.511509895324707,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92146596858639,
+ "grad_norm": 0.05140728875994682,
+ "learning_rate": 3.6697612074784944e-05,
+ "loss": 3.529341220855713,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93542757417103,
+ "grad_norm": 0.049925290048122406,
+ "learning_rate": 3.657968596707029e-05,
+ "loss": 3.538410186767578,
+ "step": 3091
+ },
+ {
+ "epoch": 42.94938917975567,
+ "grad_norm": 0.04992802441120148,
+ "learning_rate": 3.64619745093906e-05,
+ "loss": 3.502620220184326,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96335078534032,
+ "grad_norm": 0.051207154989242554,
+ "learning_rate": 3.6344477796583714e-05,
+ "loss": 3.50933575630188,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97731239092496,
+ "grad_norm": 0.05062158778309822,
+ "learning_rate": 3.622719592331472e-05,
+ "loss": 3.5215516090393066,
+ "step": 3094
+ },
+ {
+ "epoch": 42.9912739965096,
+ "grad_norm": 0.05091328173875809,
+ "learning_rate": 3.611012898407544e-05,
+ "loss": 3.523796796798706,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.03939380124211311,
+ "learning_rate": 3.599327707318464e-05,
+ "loss": 2.227039337158203,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 0.6250734925270081,
+ "eval_runtime": 59.919,
+ "eval_samples_per_second": 40.755,
+ "eval_steps_per_second": 0.651,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01396160558464,
+ "grad_norm": 0.045106735080480576,
+ "learning_rate": 3.587664028478779e-05,
+ "loss": 3.484560489654541,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02792321116929,
+ "grad_norm": 0.04480452463030815,
+ "learning_rate": 3.5760218712857046e-05,
+ "loss": 3.4960923194885254,
+ "step": 3098
+ },
+ {
+ "epoch": 43.04188481675393,
+ "grad_norm": 0.04462869092822075,
+ "learning_rate": 3.564401245119119e-05,
+ "loss": 3.509406089782715,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05584642233857,
+ "grad_norm": 0.045312751084566116,
+ "learning_rate": 3.552802159341543e-05,
+ "loss": 3.508052349090576,
+ "step": 3100
+ },
+ {
+ "epoch": 43.06980802792321,
+ "grad_norm": 0.046693965792655945,
+ "learning_rate": 3.541224623298155e-05,
+ "loss": 3.482706069946289,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083769633507856,
+ "grad_norm": 0.04617954418063164,
+ "learning_rate": 3.529668646316765e-05,
+ "loss": 3.505126953125,
+ "step": 3102
+ },
+ {
+ "epoch": 43.0977312390925,
+ "grad_norm": 0.045126017183065414,
+ "learning_rate": 3.5181342377078166e-05,
+ "loss": 3.511366367340088,
+ "step": 3103
+ },
+ {
+ "epoch": 43.11169284467714,
+ "grad_norm": 0.045670486986637115,
+ "learning_rate": 3.5066214067643785e-05,
+ "loss": 3.5024423599243164,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12565445026178,
+ "grad_norm": 0.04518337547779083,
+ "learning_rate": 3.495130162762119e-05,
+ "loss": 3.4837448596954346,
+ "step": 3105
+ },
+ {
+ "epoch": 43.139616055846425,
+ "grad_norm": 0.04559924453496933,
+ "learning_rate": 3.483660514959343e-05,
+ "loss": 3.505150556564331,
+ "step": 3106
+ },
+ {
+ "epoch": 43.153577661431065,
+ "grad_norm": 0.04574387148022652,
+ "learning_rate": 3.4722124725969235e-05,
+ "loss": 3.4935598373413086,
+ "step": 3107
+ },
+ {
+ "epoch": 43.167539267015705,
+ "grad_norm": 0.045334119349718094,
+ "learning_rate": 3.460786044898346e-05,
+ "loss": 3.498568534851074,
+ "step": 3108
+ },
+ {
+ "epoch": 43.181500872600346,
+ "grad_norm": 0.04662643373012543,
+ "learning_rate": 3.4493812410696784e-05,
+ "loss": 3.5291733741760254,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19546247818499,
+ "grad_norm": 0.04649493470788002,
+ "learning_rate": 3.437998070299564e-05,
+ "loss": 3.5123510360717773,
+ "step": 3110
+ },
+ {
+ "epoch": 43.20942408376963,
+ "grad_norm": 0.047649726271629333,
+ "learning_rate": 3.426636541759225e-05,
+ "loss": 3.4946069717407227,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223385689354274,
+ "grad_norm": 0.045752767473459244,
+ "learning_rate": 3.415296664602425e-05,
+ "loss": 3.5374412536621094,
+ "step": 3112
+ },
+ {
+ "epoch": 43.23734729493892,
+ "grad_norm": 0.04584295302629471,
+ "learning_rate": 3.403978447965507e-05,
+ "loss": 3.5236153602600098,
+ "step": 3113
+ },
+ {
+ "epoch": 43.25130890052356,
+ "grad_norm": 0.04679722711443901,
+ "learning_rate": 3.392681900967355e-05,
+ "loss": 3.4972286224365234,
+ "step": 3114
+ },
+ {
+ "epoch": 43.2652705061082,
+ "grad_norm": 0.04629402980208397,
+ "learning_rate": 3.381407032709395e-05,
+ "loss": 3.5219178199768066,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27923211169284,
+ "grad_norm": 0.046149227768182755,
+ "learning_rate": 3.3701538522755714e-05,
+ "loss": 3.4896912574768066,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29319371727749,
+ "grad_norm": 0.04659546539187431,
+ "learning_rate": 3.35892236873238e-05,
+ "loss": 3.4911556243896484,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30715532286213,
+ "grad_norm": 0.0461997464299202,
+ "learning_rate": 3.3477125911288284e-05,
+ "loss": 3.499098777770996,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32111692844677,
+ "grad_norm": 0.0467575341463089,
+ "learning_rate": 3.3365245284964225e-05,
+ "loss": 3.4635555744171143,
+ "step": 3119
+ },
+ {
+ "epoch": 43.33507853403141,
+ "grad_norm": 0.04900452122092247,
+ "learning_rate": 3.3253581898491836e-05,
+ "loss": 3.50935697555542,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34904013961606,
+ "grad_norm": 0.0469595268368721,
+ "learning_rate": 3.3142135841836276e-05,
+ "loss": 3.508852958679199,
+ "step": 3121
+ },
+ {
+ "epoch": 43.3630017452007,
+ "grad_norm": 0.048511046916246414,
+ "learning_rate": 3.303090720478773e-05,
+ "loss": 3.5232086181640625,
+ "step": 3122
+ },
+ {
+ "epoch": 43.37696335078534,
+ "grad_norm": 0.047680724412202835,
+ "learning_rate": 3.291989607696092e-05,
+ "loss": 3.510446786880493,
+ "step": 3123
+ },
+ {
+ "epoch": 43.390924956369986,
+ "grad_norm": 0.045223165303468704,
+ "learning_rate": 3.280910254779565e-05,
+ "loss": 3.477311611175537,
+ "step": 3124
+ },
+ {
+ "epoch": 43.404886561954626,
+ "grad_norm": 0.047112707048654556,
+ "learning_rate": 3.269852670655629e-05,
+ "loss": 3.488718032836914,
+ "step": 3125
+ },
+ {
+ "epoch": 43.41884816753927,
+ "grad_norm": 0.04957739636301994,
+ "learning_rate": 3.258816864233169e-05,
+ "loss": 3.511599540710449,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43280977312391,
+ "grad_norm": 0.04708652198314667,
+ "learning_rate": 3.247802844403536e-05,
+ "loss": 3.5076985359191895,
+ "step": 3127
+ },
+ {
+ "epoch": 43.446771378708554,
+ "grad_norm": 0.04767759144306183,
+ "learning_rate": 3.236810620040532e-05,
+ "loss": 3.5122997760772705,
+ "step": 3128
+ },
+ {
+ "epoch": 43.460732984293195,
+ "grad_norm": 0.04700035974383354,
+ "learning_rate": 3.225840200000396e-05,
+ "loss": 3.479099988937378,
+ "step": 3129
+ },
+ {
+ "epoch": 43.474694589877835,
+ "grad_norm": 0.046844471246004105,
+ "learning_rate": 3.214891593121788e-05,
+ "loss": 3.4828433990478516,
+ "step": 3130
+ },
+ {
+ "epoch": 43.488656195462475,
+ "grad_norm": 0.04689041152596474,
+ "learning_rate": 3.203964808225801e-05,
+ "loss": 3.49288010597229,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50261780104712,
+ "grad_norm": 0.04734606295824051,
+ "learning_rate": 3.193059854115968e-05,
+ "loss": 3.50508975982666,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51657940663176,
+ "grad_norm": 0.04855719581246376,
+ "learning_rate": 3.182176739578195e-05,
+ "loss": 3.5019211769104004,
+ "step": 3133
+ },
+ {
+ "epoch": 43.5305410122164,
+ "grad_norm": 0.047114770859479904,
+ "learning_rate": 3.171315473380815e-05,
+ "loss": 3.522956371307373,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544502617801044,
+ "grad_norm": 0.04759597033262253,
+ "learning_rate": 3.160476064274555e-05,
+ "loss": 3.488766670227051,
+ "step": 3135
+ },
+ {
+ "epoch": 43.55846422338569,
+ "grad_norm": 0.04906241595745087,
+ "learning_rate": 3.149658520992537e-05,
+ "loss": 3.507312536239624,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57242582897033,
+ "grad_norm": 0.047610968351364136,
+ "learning_rate": 3.138862852250249e-05,
+ "loss": 3.5200676918029785,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58638743455497,
+ "grad_norm": 0.048370856791734695,
+ "learning_rate": 3.1280890667455694e-05,
+ "loss": 3.488644599914551,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60034904013962,
+ "grad_norm": 0.04862125962972641,
+ "learning_rate": 3.1173371731587496e-05,
+ "loss": 3.5156755447387695,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61431064572426,
+ "grad_norm": 0.04806689918041229,
+ "learning_rate": 3.1066071801523905e-05,
+ "loss": 3.500375747680664,
+ "step": 3140
+ },
+ {
+ "epoch": 43.6282722513089,
+ "grad_norm": 0.04813141003251076,
+ "learning_rate": 3.095899096371456e-05,
+ "loss": 3.489372730255127,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64223385689354,
+ "grad_norm": 0.047211259603500366,
+ "learning_rate": 3.0852129304432506e-05,
+ "loss": 3.51399564743042,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65619546247819,
+ "grad_norm": 0.04881766065955162,
+ "learning_rate": 3.074548690977434e-05,
+ "loss": 3.5099027156829834,
+ "step": 3143
+ },
+ {
+ "epoch": 43.67015706806283,
+ "grad_norm": 0.04750249162316322,
+ "learning_rate": 3.0639063865659834e-05,
+ "loss": 3.5211377143859863,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68411867364747,
+ "grad_norm": 0.04968058690428734,
+ "learning_rate": 3.0532860257832144e-05,
+ "loss": 3.4967284202575684,
+ "step": 3145
+ },
+ {
+ "epoch": 43.69808027923211,
+ "grad_norm": 0.04763082042336464,
+ "learning_rate": 3.04268761718576e-05,
+ "loss": 3.510603666305542,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712041884816756,
+ "grad_norm": 0.04788519814610481,
+ "learning_rate": 3.0321111693125648e-05,
+ "loss": 3.486903190612793,
+ "step": 3147
+ },
+ {
+ "epoch": 43.726003490401396,
+ "grad_norm": 0.04858475551009178,
+ "learning_rate": 3.0215566906848906e-05,
+ "loss": 3.5200934410095215,
+ "step": 3148
+ },
+ {
+ "epoch": 43.73996509598604,
+ "grad_norm": 0.04838474094867706,
+ "learning_rate": 3.0110241898062726e-05,
+ "loss": 3.4970016479492188,
+ "step": 3149
+ },
+ {
+ "epoch": 43.753926701570684,
+ "grad_norm": 0.0477188378572464,
+ "learning_rate": 3.0005136751625748e-05,
+ "loss": 3.5032782554626465,
+ "step": 3150
+ },
+ {
+ "epoch": 43.767888307155324,
+ "grad_norm": 0.048580750823020935,
+ "learning_rate": 2.9900251552219178e-05,
+ "loss": 3.5062713623046875,
+ "step": 3151
+ },
+ {
+ "epoch": 43.781849912739965,
+ "grad_norm": 0.048856496810913086,
+ "learning_rate": 2.9795586384347202e-05,
+ "loss": 3.5173544883728027,
+ "step": 3152
+ },
+ {
+ "epoch": 43.795811518324605,
+ "grad_norm": 0.04810957610607147,
+ "learning_rate": 2.9691141332336494e-05,
+ "loss": 3.5213584899902344,
+ "step": 3153
+ },
+ {
+ "epoch": 43.80977312390925,
+ "grad_norm": 0.048595767468214035,
+ "learning_rate": 2.9586916480336735e-05,
+ "loss": 3.5045084953308105,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82373472949389,
+ "grad_norm": 0.048057105392217636,
+ "learning_rate": 2.9482911912320006e-05,
+ "loss": 3.4990477561950684,
+ "step": 3155
+ },
+ {
+ "epoch": 43.83769633507853,
+ "grad_norm": 0.047996897250413895,
+ "learning_rate": 2.9379127712080756e-05,
+ "loss": 3.5060601234436035,
+ "step": 3156
+ },
+ {
+ "epoch": 43.85165794066317,
+ "grad_norm": 0.0492832325398922,
+ "learning_rate": 2.9275563963236107e-05,
+ "loss": 3.4959583282470703,
+ "step": 3157
+ },
+ {
+ "epoch": 43.86561954624782,
+ "grad_norm": 0.047811295837163925,
+ "learning_rate": 2.9172220749225548e-05,
+ "loss": 3.5086965560913086,
+ "step": 3158
+ },
+ {
+ "epoch": 43.87958115183246,
+ "grad_norm": 0.04762179031968117,
+ "learning_rate": 2.906909815331085e-05,
+ "loss": 3.5065364837646484,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8935427574171,
+ "grad_norm": 0.04817652702331543,
+ "learning_rate": 2.896619625857588e-05,
+ "loss": 3.4908013343811035,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90750436300174,
+ "grad_norm": 0.04766019433736801,
+ "learning_rate": 2.8863515147927003e-05,
+ "loss": 3.4984240531921387,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92146596858639,
+ "grad_norm": 0.04822462424635887,
+ "learning_rate": 2.8761054904092514e-05,
+ "loss": 3.4717109203338623,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93542757417103,
+ "grad_norm": 0.048152536153793335,
+ "learning_rate": 2.8658815609622665e-05,
+ "loss": 3.4566991329193115,
+ "step": 3163
+ },
+ {
+ "epoch": 43.94938917975567,
+ "grad_norm": 0.04917222261428833,
+ "learning_rate": 2.855679734688993e-05,
+ "loss": 3.5119752883911133,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96335078534032,
+ "grad_norm": 0.04876438155770302,
+ "learning_rate": 2.8455000198088502e-05,
+ "loss": 3.500932216644287,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97731239092496,
+ "grad_norm": 0.048159267753362656,
+ "learning_rate": 2.8353424245234582e-05,
+ "loss": 3.4993038177490234,
+ "step": 3166
+ },
+ {
+ "epoch": 43.9912739965096,
+ "grad_norm": 0.04845968633890152,
+ "learning_rate": 2.825206957016599e-05,
+ "loss": 3.5007855892181396,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.03740447014570236,
+ "learning_rate": 2.8150936254542376e-05,
+ "loss": 2.204704523086548,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 0.6261638402938843,
+ "eval_runtime": 60.2,
+ "eval_samples_per_second": 40.565,
+ "eval_steps_per_second": 0.648,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01396160558464,
+ "grad_norm": 0.042757242918014526,
+ "learning_rate": 2.8050024379845118e-05,
+ "loss": 3.4896841049194336,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02792321116929,
+ "grad_norm": 0.04291922226548195,
+ "learning_rate": 2.7949334027376986e-05,
+ "loss": 3.519742965698242,
+ "step": 3170
+ },
+ {
+ "epoch": 44.04188481675393,
+ "grad_norm": 0.042688220739364624,
+ "learning_rate": 2.7848865278262427e-05,
+ "loss": 3.468482255935669,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05584642233857,
+ "grad_norm": 0.044173792004585266,
+ "learning_rate": 2.7748618213447268e-05,
+ "loss": 3.50119686126709,
+ "step": 3172
+ },
+ {
+ "epoch": 44.06980802792321,
+ "grad_norm": 0.04376482963562012,
+ "learning_rate": 2.764859291369885e-05,
+ "loss": 3.466043472290039,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083769633507856,
+ "grad_norm": 0.044028062373399734,
+ "learning_rate": 2.7548789459605603e-05,
+ "loss": 3.505507469177246,
+ "step": 3174
+ },
+ {
+ "epoch": 44.0977312390925,
+ "grad_norm": 0.04578253999352455,
+ "learning_rate": 2.744920793157743e-05,
+ "loss": 3.498349189758301,
+ "step": 3175
+ },
+ {
+ "epoch": 44.11169284467714,
+ "grad_norm": 0.044063616544008255,
+ "learning_rate": 2.7349848409845462e-05,
+ "loss": 3.4928536415100098,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12565445026178,
+ "grad_norm": 0.04500304162502289,
+ "learning_rate": 2.725071097446174e-05,
+ "loss": 3.5163066387176514,
+ "step": 3177
+ },
+ {
+ "epoch": 44.139616055846425,
+ "grad_norm": 0.0449041984975338,
+ "learning_rate": 2.715179570529959e-05,
+ "loss": 3.512826442718506,
+ "step": 3178
+ },
+ {
+ "epoch": 44.153577661431065,
+ "grad_norm": 0.0450286865234375,
+ "learning_rate": 2.7053102682053134e-05,
+ "loss": 3.508207082748413,
+ "step": 3179
+ },
+ {
+ "epoch": 44.167539267015705,
+ "grad_norm": 0.04426885396242142,
+ "learning_rate": 2.6954631984237766e-05,
+ "loss": 3.4822616577148438,
+ "step": 3180
+ },
+ {
+ "epoch": 44.181500872600346,
+ "grad_norm": 0.04422073811292648,
+ "learning_rate": 2.6856383691189347e-05,
+ "loss": 3.490192413330078,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19546247818499,
+ "grad_norm": 0.04403862729668617,
+ "learning_rate": 2.675835788206485e-05,
+ "loss": 3.477935552597046,
+ "step": 3182
+ },
+ {
+ "epoch": 44.20942408376963,
+ "grad_norm": 0.0447356291115284,
+ "learning_rate": 2.6660554635841912e-05,
+ "loss": 3.4865713119506836,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223385689354274,
+ "grad_norm": 0.04489988833665848,
+ "learning_rate": 2.6562974031318788e-05,
+ "loss": 3.4577198028564453,
+ "step": 3184
+ },
+ {
+ "epoch": 44.23734729493892,
+ "grad_norm": 0.046116411685943604,
+ "learning_rate": 2.6465616147114485e-05,
+ "loss": 3.4821128845214844,
+ "step": 3185
+ },
+ {
+ "epoch": 44.25130890052356,
+ "grad_norm": 0.04555881768465042,
+ "learning_rate": 2.6368481061668344e-05,
+ "loss": 3.5149850845336914,
+ "step": 3186
+ },
+ {
+ "epoch": 44.2652705061082,
+ "grad_norm": 0.04373980686068535,
+ "learning_rate": 2.6271568853240557e-05,
+ "loss": 3.49794602394104,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27923211169284,
+ "grad_norm": 0.04471353814005852,
+ "learning_rate": 2.6174879599911347e-05,
+ "loss": 3.4806714057922363,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29319371727749,
+ "grad_norm": 0.044246856123209,
+ "learning_rate": 2.6078413379581627e-05,
+ "loss": 3.481912851333618,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30715532286213,
+ "grad_norm": 0.04349849745631218,
+ "learning_rate": 2.5982170269972397e-05,
+ "loss": 3.478154182434082,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32111692844677,
+ "grad_norm": 0.044867731630802155,
+ "learning_rate": 2.588615034862507e-05,
+ "loss": 3.497056007385254,
+ "step": 3191
+ },
+ {
+ "epoch": 44.33507853403141,
+ "grad_norm": 0.04483356326818466,
+ "learning_rate": 2.5790353692901158e-05,
+ "loss": 3.4905591011047363,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34904013961606,
+ "grad_norm": 0.044564153999090195,
+ "learning_rate": 2.5694780379982238e-05,
+ "loss": 3.488785743713379,
+ "step": 3193
+ },
+ {
+ "epoch": 44.3630017452007,
+ "grad_norm": 0.044888898730278015,
+ "learning_rate": 2.5599430486870055e-05,
+ "loss": 3.5219860076904297,
+ "step": 3194
+ },
+ {
+ "epoch": 44.37696335078534,
+ "grad_norm": 0.04566695913672447,
+ "learning_rate": 2.5504304090386225e-05,
+ "loss": 3.5113654136657715,
+ "step": 3195
+ },
+ {
+ "epoch": 44.390924956369986,
+ "grad_norm": 0.045218855142593384,
+ "learning_rate": 2.5409401267172474e-05,
+ "loss": 3.4437203407287598,
+ "step": 3196
+ },
+ {
+ "epoch": 44.404886561954626,
+ "grad_norm": 0.04407033324241638,
+ "learning_rate": 2.5314722093690176e-05,
+ "loss": 3.4498770236968994,
+ "step": 3197
+ },
+ {
+ "epoch": 44.41884816753927,
+ "grad_norm": 0.04537077248096466,
+ "learning_rate": 2.522026664622071e-05,
+ "loss": 3.492107391357422,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43280977312391,
+ "grad_norm": 0.04648765176534653,
+ "learning_rate": 2.5126035000865124e-05,
+ "loss": 3.4614241123199463,
+ "step": 3199
+ },
+ {
+ "epoch": 44.446771378708554,
+ "grad_norm": 0.044741127640008926,
+ "learning_rate": 2.503202723354413e-05,
+ "loss": 3.488483428955078,
+ "step": 3200
+ },
+ {
+ "epoch": 44.460732984293195,
+ "grad_norm": 0.04496583715081215,
+ "learning_rate": 2.4938243419998047e-05,
+ "loss": 3.5051095485687256,
+ "step": 3201
+ },
+ {
+ "epoch": 44.474694589877835,
+ "grad_norm": 0.044309262186288834,
+ "learning_rate": 2.484468363578689e-05,
+ "loss": 3.503817081451416,
+ "step": 3202
+ },
+ {
+ "epoch": 44.488656195462475,
+ "grad_norm": 0.04608091711997986,
+ "learning_rate": 2.475134795629004e-05,
+ "loss": 3.4899988174438477,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50261780104712,
+ "grad_norm": 0.044740043580532074,
+ "learning_rate": 2.4658236456706357e-05,
+ "loss": 3.5216774940490723,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51657940663176,
+ "grad_norm": 0.04505334049463272,
+ "learning_rate": 2.456534921205403e-05,
+ "loss": 3.4898223876953125,
+ "step": 3205
+ },
+ {
+ "epoch": 44.5305410122164,
+ "grad_norm": 0.04470878466963768,
+ "learning_rate": 2.4472686297170822e-05,
+ "loss": 3.5085020065307617,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544502617801044,
+ "grad_norm": 0.04671593755483627,
+ "learning_rate": 2.4380247786713403e-05,
+ "loss": 3.4924004077911377,
+ "step": 3207
+ },
+ {
+ "epoch": 44.55846422338569,
+ "grad_norm": 0.04502643644809723,
+ "learning_rate": 2.428803375515788e-05,
+ "loss": 3.4717538356781006,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57242582897033,
+ "grad_norm": 0.0448530837893486,
+ "learning_rate": 2.4196044276799397e-05,
+ "loss": 3.489748954772949,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58638743455497,
+ "grad_norm": 0.04592438042163849,
+ "learning_rate": 2.410427942575228e-05,
+ "loss": 3.49373722076416,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60034904013962,
+ "grad_norm": 0.04584595933556557,
+ "learning_rate": 2.4012739275949746e-05,
+ "loss": 3.5074329376220703,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61431064572426,
+ "grad_norm": 0.04531612619757652,
+ "learning_rate": 2.3921423901144008e-05,
+ "loss": 3.4626989364624023,
+ "step": 3212
+ },
+ {
+ "epoch": 44.6282722513089,
+ "grad_norm": 0.045012522488832474,
+ "learning_rate": 2.3830333374906342e-05,
+ "loss": 3.4816949367523193,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64223385689354,
+ "grad_norm": 0.04681449756026268,
+ "learning_rate": 2.3739467770626637e-05,
+ "loss": 3.5316720008850098,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65619546247819,
+ "grad_norm": 0.04523782432079315,
+ "learning_rate": 2.3648827161513686e-05,
+ "loss": 3.5114080905914307,
+ "step": 3215
+ },
+ {
+ "epoch": 44.67015706806283,
+ "grad_norm": 0.046586692333221436,
+ "learning_rate": 2.3558411620595038e-05,
+ "loss": 3.5085229873657227,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68411867364747,
+ "grad_norm": 0.046556103974580765,
+ "learning_rate": 2.3468221220716844e-05,
+ "loss": 3.496047019958496,
+ "step": 3217
+ },
+ {
+ "epoch": 44.69808027923211,
+ "grad_norm": 0.046004518866539,
+ "learning_rate": 2.33782560345438e-05,
+ "loss": 3.498654365539551,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712041884816756,
+ "grad_norm": 0.046531420201063156,
+ "learning_rate": 2.328851613455932e-05,
+ "loss": 3.4796559810638428,
+ "step": 3219
+ },
+ {
+ "epoch": 44.726003490401396,
+ "grad_norm": 0.04683070257306099,
+ "learning_rate": 2.3199001593065204e-05,
+ "loss": 3.494418144226074,
+ "step": 3220
+ },
+ {
+ "epoch": 44.73996509598604,
+ "grad_norm": 0.04620381072163582,
+ "learning_rate": 2.3109712482181684e-05,
+ "loss": 3.488778591156006,
+ "step": 3221
+ },
+ {
+ "epoch": 44.753926701570684,
+ "grad_norm": 0.04523707926273346,
+ "learning_rate": 2.3020648873847414e-05,
+ "loss": 3.48232364654541,
+ "step": 3222
+ },
+ {
+ "epoch": 44.767888307155324,
+ "grad_norm": 0.04752018675208092,
+ "learning_rate": 2.2931810839819235e-05,
+ "loss": 3.5018577575683594,
+ "step": 3223
+ },
+ {
+ "epoch": 44.781849912739965,
+ "grad_norm": 0.04529130086302757,
+ "learning_rate": 2.284319845167253e-05,
+ "loss": 3.478651523590088,
+ "step": 3224
+ },
+ {
+ "epoch": 44.795811518324605,
+ "grad_norm": 0.04641016200184822,
+ "learning_rate": 2.2754811780800532e-05,
+ "loss": 3.5006914138793945,
+ "step": 3225
+ },
+ {
+ "epoch": 44.80977312390925,
+ "grad_norm": 0.0468735434114933,
+ "learning_rate": 2.2666650898414827e-05,
+ "loss": 3.481553554534912,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82373472949389,
+ "grad_norm": 0.04699835553765297,
+ "learning_rate": 2.2578715875545095e-05,
+ "loss": 3.496619939804077,
+ "step": 3227
+ },
+ {
+ "epoch": 44.83769633507853,
+ "grad_norm": 0.045977186411619186,
+ "learning_rate": 2.2491006783038976e-05,
+ "loss": 3.4861462116241455,
+ "step": 3228
+ },
+ {
+ "epoch": 44.85165794066317,
+ "grad_norm": 0.04640476405620575,
+ "learning_rate": 2.2403523691562092e-05,
+ "loss": 3.518702983856201,
+ "step": 3229
+ },
+ {
+ "epoch": 44.86561954624782,
+ "grad_norm": 0.04592562094330788,
+ "learning_rate": 2.2316266671598003e-05,
+ "loss": 3.47056245803833,
+ "step": 3230
+ },
+ {
+ "epoch": 44.87958115183246,
+ "grad_norm": 0.04678195342421532,
+ "learning_rate": 2.222923579344805e-05,
+ "loss": 3.4597339630126953,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8935427574171,
+ "grad_norm": 0.04635465890169144,
+ "learning_rate": 2.2142431127231502e-05,
+ "loss": 3.4888925552368164,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90750436300174,
+ "grad_norm": 0.04632442072033882,
+ "learning_rate": 2.2055852742885322e-05,
+ "loss": 3.5017807483673096,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92146596858639,
+ "grad_norm": 0.04697349667549133,
+ "learning_rate": 2.196950071016409e-05,
+ "loss": 3.4915904998779297,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93542757417103,
+ "grad_norm": 0.04719756916165352,
+ "learning_rate": 2.1883375098640134e-05,
+ "loss": 3.4640939235687256,
+ "step": 3235
+ },
+ {
+ "epoch": 44.94938917975567,
+ "grad_norm": 0.046704139560461044,
+ "learning_rate": 2.1797475977703352e-05,
+ "loss": 3.4803948402404785,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96335078534032,
+ "grad_norm": 0.04726865515112877,
+ "learning_rate": 2.1711803416560964e-05,
+ "loss": 3.4580843448638916,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97731239092496,
+ "grad_norm": 0.04642590507864952,
+ "learning_rate": 2.16263574842379e-05,
+ "loss": 3.457815170288086,
+ "step": 3238
+ },
+ {
+ "epoch": 44.9912739965096,
+ "grad_norm": 0.047316454350948334,
+ "learning_rate": 2.15411382495764e-05,
+ "loss": 3.5036163330078125,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.035441044718027115,
+ "learning_rate": 2.1456145781236057e-05,
+ "loss": 2.2000532150268555,
+ "step": 3240
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.383068587870126e+18,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-3240/training_args.bin b/runs/l2r90-ssl/checkpoint-3240/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..103048067ef8897d0386013ed89ce6764bad5521
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3240/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f320fb44ea3a4a93198f1a6c2405447428148ce142cd94ee283a3c077a327c2d
+size 4792
diff --git a/runs/l2r90-ssl/checkpoint-360/chat_template.jinja b/runs/l2r90-ssl/checkpoint-360/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-360/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-ssl/checkpoint-360/config.json b/runs/l2r90-ssl/checkpoint-360/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-360/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-ssl/checkpoint-360/generation_config.json b/runs/l2r90-ssl/checkpoint-360/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-360/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-ssl/checkpoint-360/model.safetensors b/runs/l2r90-ssl/checkpoint-360/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..b3ef194217a64be74ed602c47a5b79c348f3e487
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-360/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d86ba69945e41e28c0313407b5b3147e6302379eea3de6dbff18cfe26ffc3501
+size 583356232
diff --git a/runs/l2r90-ssl/checkpoint-360/optimizer.pt b/runs/l2r90-ssl/checkpoint-360/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..8e5bfaebfc6223bfbc5b2a025b28a1219d76eb2d
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-360/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:99f770e866710d9b34cacbc0575ba43866d3dc44a6e5166488d193be42564d0d
+size 1166825338
diff --git a/runs/l2r90-ssl/checkpoint-360/rng_state_0.pth b/runs/l2r90-ssl/checkpoint-360/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..c3dcf55387a71ee2a6afda87f94800657089f6cc
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-360/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e5cf3bac380addaa516f941e6a276bb95c92ecfb3c08371ff9fba4b2faac6df0
+size 15024
diff --git a/runs/l2r90-ssl/checkpoint-360/rng_state_1.pth b/runs/l2r90-ssl/checkpoint-360/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..26b693a57bc20ed89f333203ee3df5b86d2cc75b
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-360/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:219db464e75bbf2b6c41a71470b6182f4de4159f65fd71e3323fd50c658eea3a
+size 15024
diff --git a/runs/l2r90-ssl/checkpoint-360/rng_state_2.pth b/runs/l2r90-ssl/checkpoint-360/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..9f43ebb920eaa7f2483f205604373ffec78981af
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-360/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:03b8316374c36bf601b0d74a53807a0d460d535e42771ed5f0f6db088f66ab63
+size 15024
diff --git a/runs/l2r90-ssl/checkpoint-360/rng_state_3.pth b/runs/l2r90-ssl/checkpoint-360/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..c89f029f8f7c32ec2bc7c6471885e82ebfe5322e
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-360/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6d6c7bdfadb70cba54ce8abee8b9fbcd1d5870419c1ccbd2e7362a47c09ba74e
+size 15024
diff --git a/runs/l2r90-ssl/checkpoint-360/scheduler.pt b/runs/l2r90-ssl/checkpoint-360/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..d63b22e047ab13a6c29adc08f78969e1d6e569ff
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-360/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2802ceffe363a15cca2d2192020ec6c680a5afa8cc5d08bb488000d79d1a8eb0
+size 1064
diff --git a/runs/l2r90-ssl/checkpoint-360/tokenizer.json b/runs/l2r90-ssl/checkpoint-360/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-360/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-ssl/checkpoint-360/tokenizer_config.json b/runs/l2r90-ssl/checkpoint-360/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-360/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-360/trainer_state.json b/runs/l2r90-ssl/checkpoint-360/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..8a6a250b500bfbe1a148cdd7781c9b3fe261ed30
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-360/trainer_state.json
@@ -0,0 +1,2586 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 5.0,
+ "eval_steps": 500,
+ "global_step": 360,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.2588357925415039,
+ "learning_rate": 0.0,
+ "loss": 3.9982247352600098,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.25455471873283386,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.998683452606201,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.2555367052555084,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.9821574687957764,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.2383691370487213,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.9531917572021484,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2023179680109024,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.9148876667022705,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.17691479623317719,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.875638961791992,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.14372293651103973,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.836796283721924,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1074419692158699,
+ "learning_rate": 4.2e-05,
+ "loss": 3.8036489486694336,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.08204353600740433,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.774808406829834,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.0632256343960762,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.7503671646118164,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.047985248267650604,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.7345473766326904,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.03675895184278488,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.721525192260742,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.02908993512392044,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.713454246520996,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.02443614974617958,
+ "learning_rate": 7.8e-05,
+ "loss": 3.7078704833984375,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02109951712191105,
+ "learning_rate": 8.4e-05,
+ "loss": 3.7034401893615723,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.018616363406181335,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.7009854316711426,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.016399752348661423,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.697768211364746,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.014256286434829235,
+ "learning_rate": 0.000102,
+ "loss": 3.6974306106567383,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.013020666316151619,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.694406509399414,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.011879238300025463,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.691807270050049,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.010290366597473621,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.6937897205352783,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.009482803754508495,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.6902384757995605,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.008425266481935978,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.686815023422241,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.007676936220377684,
+ "learning_rate": 0.000138,
+ "loss": 3.685192823410034,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.007491032127290964,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.6839442253112793,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.007428477052599192,
+ "learning_rate": 0.00015,
+ "loss": 3.6783409118652344,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.006965892389416695,
+ "learning_rate": 0.000156,
+ "loss": 3.6813433170318604,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.008106694556772709,
+ "learning_rate": 0.000162,
+ "loss": 3.6741726398468018,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.0076780058443546295,
+ "learning_rate": 0.000168,
+ "loss": 3.67673921585083,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.007904314436018467,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.6724467277526855,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.008422375656664371,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.673013925552368,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00811714306473732,
+ "learning_rate": 0.000186,
+ "loss": 3.6698389053344727,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.009063836187124252,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.668990135192871,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.01878332532942295,
+ "learning_rate": 0.000198,
+ "loss": 3.6658849716186523,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.01812782697379589,
+ "learning_rate": 0.000204,
+ "loss": 3.6697864532470703,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.020587686449289322,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.662757158279419,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.01515690516680479,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.6558732986450195,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.02510508894920349,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.6495113372802734,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.020487826317548752,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.638803243637085,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.017446046695113182,
+ "learning_rate": 0.000234,
+ "loss": 3.633307456970215,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.030667584389448166,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.6277506351470947,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.023789944127202034,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.6189804077148438,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.02660694345831871,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.610522747039795,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.023814303800463676,
+ "learning_rate": 0.000258,
+ "loss": 3.615910291671753,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.021863561123609543,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.612008571624756,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03705216199159622,
+ "learning_rate": 0.00027,
+ "loss": 3.6066887378692627,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.024143919348716736,
+ "learning_rate": 0.000276,
+ "loss": 3.5987331867218018,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.03251276910305023,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.595367431640625,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06619632989168167,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.59018611907959,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.042714253067970276,
+ "learning_rate": 0.000294,
+ "loss": 3.58981990814209,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.018893254920840263,
+ "learning_rate": 0.0003,
+ "loss": 3.578472375869751,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0749685987830162,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.5828654766082764,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.12142857164144516,
+ "learning_rate": 0.000312,
+ "loss": 3.5840892791748047,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.06944665312767029,
+ "learning_rate": 0.000318,
+ "loss": 3.580961227416992,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.04545000568032265,
+ "learning_rate": 0.000324,
+ "loss": 3.5674569606781006,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.061073049902915955,
+ "learning_rate": 0.00033,
+ "loss": 3.569991111755371,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.04133498668670654,
+ "learning_rate": 0.000336,
+ "loss": 3.5657706260681152,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.055982090532779694,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.559979200363159,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.05259208381175995,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.558039426803589,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.029976001009345055,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.5491275787353516,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.04998327046632767,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.5406670570373535,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.03230690956115723,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.5391530990600586,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.06102178618311882,
+ "learning_rate": 0.000372,
+ "loss": 3.5311684608459473,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.031715717166662216,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.518899440765381,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.06841139495372772,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.514200210571289,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.10323675721883774,
+ "learning_rate": 0.00039,
+ "loss": 3.513646125793457,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.1448228508234024,
+ "learning_rate": 0.000396,
+ "loss": 3.5110087394714355,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.12129563093185425,
+ "learning_rate": 0.000402,
+ "loss": 3.501497507095337,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.030224842950701714,
+ "learning_rate": 0.000408,
+ "loss": 3.4870352745056152,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.07492101192474365,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.482234001159668,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.03942760080099106,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.4746525287628174,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.04967200756072998,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.604069232940674,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8671882152557373,
+ "eval_runtime": 42.6602,
+ "eval_samples_per_second": 57.243,
+ "eval_steps_per_second": 0.469,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.05385325849056244,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.462052822113037,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.04870929941534996,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.459104537963867,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.08701689541339874,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.4533209800720215,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.12219952791929245,
+ "learning_rate": 0.00045,
+ "loss": 3.4558796882629395,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.0823654904961586,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.443049192428589,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.12765446305274963,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.4373486042022705,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.14215047657489777,
+ "learning_rate": 0.000468,
+ "loss": 3.449429988861084,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.03649887070059776,
+ "learning_rate": 0.000474,
+ "loss": 3.42877197265625,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09551949054002762,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.4352874755859375,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.030483486130833626,
+ "learning_rate": 0.000486,
+ "loss": 3.4125874042510986,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.0745147094130516,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.4153835773468018,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.06620422005653381,
+ "learning_rate": 0.000498,
+ "loss": 3.4053854942321777,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.06098725646734238,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.404346466064453,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.047781601548194885,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.3991761207580566,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.04409855231642723,
+ "learning_rate": 0.000516,
+ "loss": 3.396144390106201,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.04355994611978531,
+ "learning_rate": 0.000522,
+ "loss": 3.3928911685943604,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.0670008510351181,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.3831443786621094,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.0812622532248497,
+ "learning_rate": 0.000534,
+ "loss": 3.382413148880005,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.0632157102227211,
+ "learning_rate": 0.00054,
+ "loss": 3.3784892559051514,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.028686825186014175,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.3673207759857178,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.034818850457668304,
+ "learning_rate": 0.000552,
+ "loss": 3.3648972511291504,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.04509204626083374,
+ "learning_rate": 0.000558,
+ "loss": 3.35561203956604,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.054211195558309555,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.3612167835235596,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.03752049803733826,
+ "learning_rate": 0.00057,
+ "loss": 3.357103109359741,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.03620241954922676,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.3469700813293457,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.02454194985330105,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.3429036140441895,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.041030462831258774,
+ "learning_rate": 0.000588,
+ "loss": 3.338247299194336,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.03266424685716629,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.332688331604004,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.04730091616511345,
+ "learning_rate": 0.0006,
+ "loss": 3.335902452468872,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.065738245844841,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.3289852142333984,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.09062394499778748,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.3278567790985107,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.1482805460691452,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.332685947418213,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.20066532492637634,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.363180637359619,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.11500677466392517,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.3308753967285156,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.11754138022661209,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.3383703231811523,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.09672239422798157,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.3267624378204346,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.07190799713134766,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.317737579345703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.0815449133515358,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.3198161125183105,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09612327069044113,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.315814971923828,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.14795289933681488,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.317605495452881,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06667589396238327,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.3127357959747314,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.058537885546684265,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.319807767868042,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.06428155303001404,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.297541618347168,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.03881775587797165,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.297539710998535,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.061801470816135406,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.290297508239746,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.07642851769924164,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.284428596496582,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08123870939016342,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.2820234298706055,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07174684852361679,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.2697296142578125,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.04030010476708412,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.264328956604004,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.05662153288722038,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.2699337005615234,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.03761111572384834,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.255704879760742,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.043027132749557495,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.2579030990600586,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.044409338384866714,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.248547077178955,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.0239730142056942,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.249307632446289,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03628705441951752,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.2451741695404053,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.029564054682850838,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.238445997238159,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.032036975026130676,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.2320892810821533,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.030085794627666473,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.231468915939331,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.025232303887605667,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.232199192047119,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.027053389698266983,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.210745334625244,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.028464114293456078,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.210076332092285,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04616783559322357,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.2103490829467773,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.039081938564777374,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.202662467956543,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.03428936004638672,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.1997475624084473,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.04351917654275894,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.19844913482666,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.0840945914387703,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.2013347148895264,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.36558300256729126,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.26310396194458,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.09432770311832428,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.2881736755371094,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.061970267444849014,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.2840280532836914,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.05112256482243538,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.2632040977478027,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06797203421592712,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.254606246948242,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.051026541739702225,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.429918050765991,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.8117228746414185,
+ "eval_runtime": 42.8572,
+ "eval_samples_per_second": 56.98,
+ "eval_steps_per_second": 0.467,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.0608622208237648,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.2348179817199707,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.05581629276275635,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.22896146774292,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.09162136912345886,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.2343664169311523,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.13750384747982025,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.2507717609405518,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.0872517079114914,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.2440056800842285,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.07161762565374374,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.229494571685791,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04082738608121872,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.2183239459991455,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.04972497373819351,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.218075752258301,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.049254074692726135,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.213379383087158,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.056671254336833954,
+ "learning_rate": 0.000599663984012462,
+ "loss": 3.209568977355957,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.05541909113526344,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.207760810852051,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.03819117695093155,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 3.199732780456543,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.0529443584382534,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.1945419311523438,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.043010979890823364,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.185936212539673,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.07311078161001205,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.189894914627075,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06110258400440216,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 3.1911773681640625,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.06880149245262146,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 3.187437057495117,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.051326874643564224,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 3.1826558113098145,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.03363822400569916,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 3.175527572631836,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.10268989205360413,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 3.172367572784424,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.09723010659217834,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 3.1873984336853027,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.04129317030310631,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 3.165799617767334,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.05292229354381561,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 3.171780824661255,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.03873143345117569,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 3.1680874824523926,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.047018859535455704,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 3.161402940750122,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.03942955285310745,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 3.1530532836914062,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.039801858365535736,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 3.1465344429016113,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.03795155510306358,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 3.1364030838012695,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.0342850424349308,
+ "learning_rate": 0.000599379982511273,
+ "loss": 3.1391525268554688,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.03243280574679375,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 3.1331253051757812,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.029679352417588234,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 3.1346120834350586,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.030230293050408363,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 3.132615804672241,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.02845083735883236,
+ "learning_rate": 0.000599309205504528,
+ "loss": 3.1242547035217285,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.03208369016647339,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 3.119110584259033,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.040268607437610626,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 3.113833427429199,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.06951889395713806,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 3.1186914443969727,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.08879823237657547,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 3.1140987873077393,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07374681532382965,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 3.1147828102111816,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.038190409541130066,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 3.1161253452301025,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.03297388181090355,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 3.1006083488464355,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.05905177444219589,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 3.106564521789551,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.037567321211099625,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 3.0884456634521484,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.02725658379495144,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 3.09000825881958,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.04272424429655075,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 3.090528964996338,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.027587950229644775,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 3.085131883621216,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.03282274678349495,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 3.0880112648010254,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.04064026474952698,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 3.079259157180786,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.023374097421765327,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 3.0747299194335938,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.025003870949149132,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 3.0721259117126465,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.030445197597146034,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 3.0733323097229004,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.025935636833310127,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 3.070709228515625,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.021337421610951424,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 3.0627551078796387,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.040481697767972946,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 3.0529394149780273,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.047608256340026855,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 3.041020631790161,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.03356955572962761,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 3.0532307624816895,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.02101682312786579,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 3.0492985248565674,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.030098581686615944,
+ "learning_rate": 0.000598804365316586,
+ "loss": 3.0527801513671875,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.040166422724723816,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 3.0505762100219727,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.051579318940639496,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 3.0422110557556152,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.06485334783792496,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 3.0435101985931396,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.055590901523828506,
+ "learning_rate": 0.000598706872384039,
+ "loss": 3.0448739528656006,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.03183475881814957,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 3.0416858196258545,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.03995629772543907,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 3.0255980491638184,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.03907827287912369,
+ "learning_rate": 0.000598631251036868,
+ "loss": 3.032339096069336,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.027681387960910797,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 3.025439739227295,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.03470184653997421,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 3.02883243560791,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.031407181173563004,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 3.0289883613586426,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.0371522456407547,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 3.013608932495117,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.03651345893740654,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 3.0120253562927246,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.06004595756530762,
+ "learning_rate": 0.000598473577921366,
+ "loss": 3.017958402633667,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.09979518502950668,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 3.023041248321533,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.10138462483882904,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.275729179382324,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7571707367897034,
+ "eval_runtime": 42.8283,
+ "eval_samples_per_second": 57.018,
+ "eval_steps_per_second": 0.467,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.04982660710811615,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 3.016209602355957,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.06532539427280426,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 3.015043020248413,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.04369623214006424,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 3.0096778869628906,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.05180002748966217,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 3.0121543407440186,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.0440233014523983,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.9937305450439453,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.05645671859383583,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 3.0023865699768066,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.03990716487169266,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 3.0003232955932617,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.03818608075380325,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.994137763977051,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.035388972610235214,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.993170738220215,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.03162284567952156,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.9922614097595215,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.029497336596250534,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.9852986335754395,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.02586926706135273,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.9798672199249268,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.027355050668120384,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.985158920288086,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.0341978520154953,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.982581615447998,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.030727429315447807,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.9701411724090576,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.04429870471358299,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.9789254665374756,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.05775132030248642,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.9758143424987793,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.05635962635278702,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.981480121612549,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.03639192506670952,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.968752861022949,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.028748342767357826,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.9584972858428955,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.035653006285429,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.967545747756958,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.03697885945439339,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.9578099250793457,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.03518654406070709,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.9680328369140625,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.034762222319841385,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.9670262336730957,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03813016042113304,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.957522392272949,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.04372904822230339,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.9540514945983887,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.06646446138620377,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.951028823852539,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.07831092923879623,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.95898699760437,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.05779588967561722,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.965884208679199,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.04414296895265579,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.9510560035705566,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.04797815904021263,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.955803394317627,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.04482145980000496,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.954059362411499,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.037104446440935135,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.94792103767395,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.03616701066493988,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.9500932693481445,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.03882797807455063,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.9558305740356445,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.04356204345822334,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.9237029552459717,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.04535257816314697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.925853729248047,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.04519587755203247,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.924984931945801,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.055922407656908035,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.932561159133911,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.06330762803554535,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.9388322830200195,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.0517115443944931,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.9272942543029785,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.057369694113731384,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.9201431274414062,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.05687880516052246,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.922886371612549,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.055525410920381546,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.9211201667785645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04503652825951576,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.9164583683013916,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.03385302051901817,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.9145395755767822,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03564384952187538,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.916003704071045,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.04188712686300278,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.915839433670044,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03657370060682297,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.903120279312134,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.03126626834273338,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.895824432373047,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.04196290299296379,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.9093422889709473,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.1706087589263916,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.9183316230773926,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.05810421332716942,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.911245107650757,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.061330314725637436,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.9159231185913086,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.06062281131744385,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.9211535453796387,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.05590800568461418,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.9140496253967285,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.04496239125728607,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.9040920734405518,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.03835335746407509,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.9026548862457275,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.05137286335229874,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.889634132385254,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.04091101884841919,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.886838674545288,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.04076292738318443,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.892918109893799,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.045437734574079514,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.893305778503418,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.041114673018455505,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.88517165184021,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.03600773215293884,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.8779795169830322,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.04311220720410347,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.87247896194458,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.04818756878376007,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.889103412628174,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.053460877388715744,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.8899683952331543,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.06763825565576553,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.8823537826538086,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.07304801046848297,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.877340316772461,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.061337269842624664,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.8820111751556396,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.04660911113023758,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.868886709213257,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.053929202258586884,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.159337043762207,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.721762478351593,
+ "eval_runtime": 42.7977,
+ "eval_samples_per_second": 57.059,
+ "eval_steps_per_second": 0.467,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.04259369149804115,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.859579086303711,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.044425517320632935,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.8681139945983887,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.04301869496703148,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.853911876678467,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.04574296995997429,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.8724427223205566,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.050247322767972946,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.8528172969818115,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.046683840453624725,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.853405714035034,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.05966169387102127,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.8605127334594727,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.08935574442148209,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.862339973449707,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.08019179105758667,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.8655967712402344,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.04753892868757248,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.8657939434051514,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06868117302656174,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.861976146697998,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.07040470093488693,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.8562135696411133,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.05683179199695587,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.8576629161834717,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.04918001592159271,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.861943244934082,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.04921915754675865,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.8495378494262695,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.037651967257261276,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.8428618907928467,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.03522057458758354,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.8372769355773926,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.03111186809837818,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.8377151489257812,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.03906773030757904,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.846196174621582,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.02546556107699871,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.8288016319274902,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.03119821846485138,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.8334617614746094,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.030583377927541733,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.8209362030029297,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.029693076387047768,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.8267486095428467,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.028978517279028893,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.819244861602783,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.027871523052453995,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.8289647102355957,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.029287064447999,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.8235747814178467,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.024990517646074295,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.819016218185425,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.030980689451098442,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.828155040740967,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03721801936626434,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.813418388366699,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03738148882985115,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.808328866958618,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.03580411896109581,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.812070846557617,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.039036594331264496,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.807159423828125,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.044666968286037445,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.799992561340332,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.055589817464351654,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.804142951965332,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.06139020994305611,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.8182730674743652,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.07769318670034409,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.8166518211364746,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.06521332263946533,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.8142249584198,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.09083619713783264,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.8166933059692383,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.09737467020750046,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.8111555576324463,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.07148188352584839,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.8168413639068604,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.055516090244054794,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.8183553218841553,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.05338391661643982,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.8127028942108154,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.040785472840070724,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.8124775886535645,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.037135496735572815,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.8065357208251953,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.032455578446388245,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.798459053039551,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.03465662896633148,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.798649549484253,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.032026033848524094,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.7906997203826904,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.026019956916570663,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.791969060897827,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.030298635363578796,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.784593105316162,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.028147678822278976,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.782454013824463,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.023933131247758865,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.787851333618164,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.026789527386426926,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.788067579269409,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.024554884061217308,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.777033567428589,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.028679484501481056,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.786342144012451,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.026410218328237534,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.7790610790252686,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.027413051575422287,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.778944969177246,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.02908242493867874,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.763850212097168,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.029632994905114174,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.7672982215881348,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.027458716183900833,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.7702178955078125,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.03263958916068077,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.7581803798675537,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04427146911621094,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.768141508102417,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.07142394036054611,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.75819730758667,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.08833584934473038,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.782716751098633,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.08168833702802658,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.7763240337371826,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07253874838352203,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.7717690467834473,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.08186902105808258,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.767303466796875,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.05854064226150513,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.761544704437256,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.04602917283773422,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.7669177055358887,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.03920970857143402,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.766085624694824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.04551195353269577,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.7623450756073,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.040861159563064575,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.7471840381622314,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.026907291263341904,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 2.0710244178771973,
+ "step": 360
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.5388280082137088e+17,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-360/training_args.bin b/runs/l2r90-ssl/checkpoint-360/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5924a9264314215faa31ea62b22ba59c4573b542
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-360/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2e206f0318237cba8820486751e683a8c1e9fd84a823416f09cf3e109aba53ea
+size 4792
diff --git a/runs/l2r90-ssl/checkpoint-3600/chat_template.jinja b/runs/l2r90-ssl/checkpoint-3600/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3600/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-ssl/checkpoint-3600/config.json b/runs/l2r90-ssl/checkpoint-3600/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3600/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-ssl/checkpoint-3600/generation_config.json b/runs/l2r90-ssl/checkpoint-3600/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3600/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-ssl/checkpoint-3600/model.safetensors b/runs/l2r90-ssl/checkpoint-3600/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..5367afc536511faba4ee445b983fe71fb18ace96
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3600/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:77615e2e44a772db8bded3cc96bfdc363aa67aa3c21f038d17aa15f5dd53d6db
+size 583356232
diff --git a/runs/l2r90-ssl/checkpoint-3600/optimizer.pt b/runs/l2r90-ssl/checkpoint-3600/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..07962f2d78f5be1fc4a30ac7f8a2410d247767b4
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3600/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6f624c4779bbf5328b88a180015c1c6f0142c19c5bdee8335c682a68f3a2c349
+size 1166827898
diff --git a/runs/l2r90-ssl/checkpoint-3600/rng_state_0.pth b/runs/l2r90-ssl/checkpoint-3600/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..10cf19ced1d6fe361d9ff544bbdb548022a59618
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3600/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c7d3c90a44e89c13781b08180267f17a74b3cac676f1d241cd03526411fa5da3
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-3600/rng_state_1.pth b/runs/l2r90-ssl/checkpoint-3600/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..1c06c37cb1623f83caf5f5a109012a87ce995705
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3600/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:41d5db16f3caa3e5d010918481614cb1af2c949c60f242b127f9b971148ce197
+size 14512
diff --git a/runs/l2r90-ssl/checkpoint-3600/scheduler.pt b/runs/l2r90-ssl/checkpoint-3600/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..dff04aa3b724ae7dec72031a9c3d0ecf0b695f1b
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3600/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:187927afe10d5874e052562039160b9509612fe972489c80a97620ccae701300
+size 1064
diff --git a/runs/l2r90-ssl/checkpoint-3600/tokenizer.json b/runs/l2r90-ssl/checkpoint-3600/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3600/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-ssl/checkpoint-3600/tokenizer_config.json b/runs/l2r90-ssl/checkpoint-3600/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3600/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-3600/trainer_state.json b/runs/l2r90-ssl/checkpoint-3600/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..492881515dfe705aff934134944c5a9366c82972
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3600/trainer_state.json
@@ -0,0 +1,25618 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 50.0,
+ "eval_steps": 500,
+ "global_step": 3600,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.2588357925415039,
+ "learning_rate": 0.0,
+ "loss": 3.9982247352600098,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.25455471873283386,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.998683452606201,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.2555367052555084,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.9821574687957764,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.2383691370487213,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.9531917572021484,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2023179680109024,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.9148876667022705,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.17691479623317719,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.875638961791992,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.14372293651103973,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.836796283721924,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1074419692158699,
+ "learning_rate": 4.2e-05,
+ "loss": 3.8036489486694336,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.08204353600740433,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.774808406829834,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.0632256343960762,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.7503671646118164,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.047985248267650604,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.7345473766326904,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.03675895184278488,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.721525192260742,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.02908993512392044,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.713454246520996,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.02443614974617958,
+ "learning_rate": 7.8e-05,
+ "loss": 3.7078704833984375,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02109951712191105,
+ "learning_rate": 8.4e-05,
+ "loss": 3.7034401893615723,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.018616363406181335,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.7009854316711426,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.016399752348661423,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.697768211364746,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.014256286434829235,
+ "learning_rate": 0.000102,
+ "loss": 3.6974306106567383,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.013020666316151619,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.694406509399414,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.011879238300025463,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.691807270050049,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.010290366597473621,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.6937897205352783,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.009482803754508495,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.6902384757995605,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.008425266481935978,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.686815023422241,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.007676936220377684,
+ "learning_rate": 0.000138,
+ "loss": 3.685192823410034,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.007491032127290964,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.6839442253112793,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.007428477052599192,
+ "learning_rate": 0.00015,
+ "loss": 3.6783409118652344,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.006965892389416695,
+ "learning_rate": 0.000156,
+ "loss": 3.6813433170318604,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.008106694556772709,
+ "learning_rate": 0.000162,
+ "loss": 3.6741726398468018,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.0076780058443546295,
+ "learning_rate": 0.000168,
+ "loss": 3.67673921585083,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.007904314436018467,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.6724467277526855,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.008422375656664371,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.673013925552368,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00811714306473732,
+ "learning_rate": 0.000186,
+ "loss": 3.6698389053344727,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.009063836187124252,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.668990135192871,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.01878332532942295,
+ "learning_rate": 0.000198,
+ "loss": 3.6658849716186523,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.01812782697379589,
+ "learning_rate": 0.000204,
+ "loss": 3.6697864532470703,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.020587686449289322,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.662757158279419,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.01515690516680479,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.6558732986450195,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.02510508894920349,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.6495113372802734,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.020487826317548752,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.638803243637085,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.017446046695113182,
+ "learning_rate": 0.000234,
+ "loss": 3.633307456970215,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.030667584389448166,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.6277506351470947,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.023789944127202034,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.6189804077148438,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.02660694345831871,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.610522747039795,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.023814303800463676,
+ "learning_rate": 0.000258,
+ "loss": 3.615910291671753,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.021863561123609543,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.612008571624756,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03705216199159622,
+ "learning_rate": 0.00027,
+ "loss": 3.6066887378692627,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.024143919348716736,
+ "learning_rate": 0.000276,
+ "loss": 3.5987331867218018,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.03251276910305023,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.595367431640625,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06619632989168167,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.59018611907959,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.042714253067970276,
+ "learning_rate": 0.000294,
+ "loss": 3.58981990814209,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.018893254920840263,
+ "learning_rate": 0.0003,
+ "loss": 3.578472375869751,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0749685987830162,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.5828654766082764,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.12142857164144516,
+ "learning_rate": 0.000312,
+ "loss": 3.5840892791748047,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.06944665312767029,
+ "learning_rate": 0.000318,
+ "loss": 3.580961227416992,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.04545000568032265,
+ "learning_rate": 0.000324,
+ "loss": 3.5674569606781006,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.061073049902915955,
+ "learning_rate": 0.00033,
+ "loss": 3.569991111755371,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.04133498668670654,
+ "learning_rate": 0.000336,
+ "loss": 3.5657706260681152,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.055982090532779694,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.559979200363159,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.05259208381175995,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.558039426803589,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.029976001009345055,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.5491275787353516,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.04998327046632767,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.5406670570373535,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.03230690956115723,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.5391530990600586,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.06102178618311882,
+ "learning_rate": 0.000372,
+ "loss": 3.5311684608459473,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.031715717166662216,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.518899440765381,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.06841139495372772,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.514200210571289,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.10323675721883774,
+ "learning_rate": 0.00039,
+ "loss": 3.513646125793457,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.1448228508234024,
+ "learning_rate": 0.000396,
+ "loss": 3.5110087394714355,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.12129563093185425,
+ "learning_rate": 0.000402,
+ "loss": 3.501497507095337,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.030224842950701714,
+ "learning_rate": 0.000408,
+ "loss": 3.4870352745056152,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.07492101192474365,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.482234001159668,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.03942760080099106,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.4746525287628174,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.04967200756072998,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.604069232940674,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8671882152557373,
+ "eval_runtime": 42.6602,
+ "eval_samples_per_second": 57.243,
+ "eval_steps_per_second": 0.469,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.05385325849056244,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.462052822113037,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.04870929941534996,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.459104537963867,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.08701689541339874,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.4533209800720215,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.12219952791929245,
+ "learning_rate": 0.00045,
+ "loss": 3.4558796882629395,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.0823654904961586,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.443049192428589,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.12765446305274963,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.4373486042022705,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.14215047657489777,
+ "learning_rate": 0.000468,
+ "loss": 3.449429988861084,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.03649887070059776,
+ "learning_rate": 0.000474,
+ "loss": 3.42877197265625,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09551949054002762,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.4352874755859375,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.030483486130833626,
+ "learning_rate": 0.000486,
+ "loss": 3.4125874042510986,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.0745147094130516,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.4153835773468018,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.06620422005653381,
+ "learning_rate": 0.000498,
+ "loss": 3.4053854942321777,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.06098725646734238,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.404346466064453,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.047781601548194885,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.3991761207580566,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.04409855231642723,
+ "learning_rate": 0.000516,
+ "loss": 3.396144390106201,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.04355994611978531,
+ "learning_rate": 0.000522,
+ "loss": 3.3928911685943604,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.0670008510351181,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.3831443786621094,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.0812622532248497,
+ "learning_rate": 0.000534,
+ "loss": 3.382413148880005,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.0632157102227211,
+ "learning_rate": 0.00054,
+ "loss": 3.3784892559051514,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.028686825186014175,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.3673207759857178,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.034818850457668304,
+ "learning_rate": 0.000552,
+ "loss": 3.3648972511291504,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.04509204626083374,
+ "learning_rate": 0.000558,
+ "loss": 3.35561203956604,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.054211195558309555,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.3612167835235596,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.03752049803733826,
+ "learning_rate": 0.00057,
+ "loss": 3.357103109359741,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.03620241954922676,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.3469700813293457,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.02454194985330105,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.3429036140441895,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.041030462831258774,
+ "learning_rate": 0.000588,
+ "loss": 3.338247299194336,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.03266424685716629,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.332688331604004,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.04730091616511345,
+ "learning_rate": 0.0006,
+ "loss": 3.335902452468872,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.065738245844841,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.3289852142333984,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.09062394499778748,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.3278567790985107,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.1482805460691452,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.332685947418213,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.20066532492637634,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.363180637359619,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.11500677466392517,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.3308753967285156,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.11754138022661209,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.3383703231811523,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.09672239422798157,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.3267624378204346,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.07190799713134766,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.317737579345703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.0815449133515358,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.3198161125183105,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09612327069044113,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.315814971923828,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.14795289933681488,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.317605495452881,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06667589396238327,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.3127357959747314,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.058537885546684265,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.319807767868042,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.06428155303001404,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.297541618347168,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.03881775587797165,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.297539710998535,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.061801470816135406,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.290297508239746,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.07642851769924164,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.284428596496582,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08123870939016342,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.2820234298706055,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07174684852361679,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.2697296142578125,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.04030010476708412,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.264328956604004,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.05662153288722038,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.2699337005615234,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.03761111572384834,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.255704879760742,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.043027132749557495,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.2579030990600586,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.044409338384866714,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.248547077178955,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.0239730142056942,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.249307632446289,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03628705441951752,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.2451741695404053,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.029564054682850838,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.238445997238159,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.032036975026130676,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.2320892810821533,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.030085794627666473,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.231468915939331,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.025232303887605667,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.232199192047119,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.027053389698266983,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.210745334625244,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.028464114293456078,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.210076332092285,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04616783559322357,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.2103490829467773,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.039081938564777374,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.202662467956543,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.03428936004638672,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.1997475624084473,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.04351917654275894,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.19844913482666,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.0840945914387703,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.2013347148895264,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.36558300256729126,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.26310396194458,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.09432770311832428,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.2881736755371094,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.061970267444849014,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.2840280532836914,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.05112256482243538,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.2632040977478027,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06797203421592712,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.254606246948242,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.051026541739702225,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.429918050765991,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.8117228746414185,
+ "eval_runtime": 42.8572,
+ "eval_samples_per_second": 56.98,
+ "eval_steps_per_second": 0.467,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.0608622208237648,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.2348179817199707,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.05581629276275635,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.22896146774292,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.09162136912345886,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.2343664169311523,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.13750384747982025,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.2507717609405518,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.0872517079114914,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.2440056800842285,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.07161762565374374,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.229494571685791,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04082738608121872,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.2183239459991455,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.04972497373819351,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.218075752258301,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.049254074692726135,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.213379383087158,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.056671254336833954,
+ "learning_rate": 0.000599663984012462,
+ "loss": 3.209568977355957,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.05541909113526344,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.207760810852051,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.03819117695093155,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 3.199732780456543,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.0529443584382534,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.1945419311523438,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.043010979890823364,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.185936212539673,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.07311078161001205,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.189894914627075,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06110258400440216,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 3.1911773681640625,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.06880149245262146,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 3.187437057495117,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.051326874643564224,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 3.1826558113098145,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.03363822400569916,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 3.175527572631836,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.10268989205360413,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 3.172367572784424,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.09723010659217834,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 3.1873984336853027,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.04129317030310631,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 3.165799617767334,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.05292229354381561,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 3.171780824661255,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.03873143345117569,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 3.1680874824523926,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.047018859535455704,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 3.161402940750122,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.03942955285310745,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 3.1530532836914062,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.039801858365535736,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 3.1465344429016113,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.03795155510306358,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 3.1364030838012695,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.0342850424349308,
+ "learning_rate": 0.000599379982511273,
+ "loss": 3.1391525268554688,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.03243280574679375,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 3.1331253051757812,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.029679352417588234,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 3.1346120834350586,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.030230293050408363,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 3.132615804672241,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.02845083735883236,
+ "learning_rate": 0.000599309205504528,
+ "loss": 3.1242547035217285,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.03208369016647339,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 3.119110584259033,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.040268607437610626,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 3.113833427429199,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.06951889395713806,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 3.1186914443969727,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.08879823237657547,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 3.1140987873077393,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07374681532382965,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 3.1147828102111816,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.038190409541130066,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 3.1161253452301025,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.03297388181090355,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 3.1006083488464355,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.05905177444219589,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 3.106564521789551,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.037567321211099625,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 3.0884456634521484,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.02725658379495144,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 3.09000825881958,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.04272424429655075,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 3.090528964996338,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.027587950229644775,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 3.085131883621216,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.03282274678349495,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 3.0880112648010254,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.04064026474952698,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 3.079259157180786,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.023374097421765327,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 3.0747299194335938,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.025003870949149132,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 3.0721259117126465,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.030445197597146034,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 3.0733323097229004,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.025935636833310127,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 3.070709228515625,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.021337421610951424,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 3.0627551078796387,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.040481697767972946,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 3.0529394149780273,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.047608256340026855,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 3.041020631790161,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.03356955572962761,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 3.0532307624816895,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.02101682312786579,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 3.0492985248565674,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.030098581686615944,
+ "learning_rate": 0.000598804365316586,
+ "loss": 3.0527801513671875,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.040166422724723816,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 3.0505762100219727,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.051579318940639496,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 3.0422110557556152,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.06485334783792496,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 3.0435101985931396,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.055590901523828506,
+ "learning_rate": 0.000598706872384039,
+ "loss": 3.0448739528656006,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.03183475881814957,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 3.0416858196258545,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.03995629772543907,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 3.0255980491638184,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.03907827287912369,
+ "learning_rate": 0.000598631251036868,
+ "loss": 3.032339096069336,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.027681387960910797,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 3.025439739227295,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.03470184653997421,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 3.02883243560791,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.031407181173563004,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 3.0289883613586426,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.0371522456407547,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 3.013608932495117,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.03651345893740654,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 3.0120253562927246,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.06004595756530762,
+ "learning_rate": 0.000598473577921366,
+ "loss": 3.017958402633667,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.09979518502950668,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 3.023041248321533,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.10138462483882904,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.275729179382324,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7571707367897034,
+ "eval_runtime": 42.8283,
+ "eval_samples_per_second": 57.018,
+ "eval_steps_per_second": 0.467,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.04982660710811615,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 3.016209602355957,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.06532539427280426,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 3.015043020248413,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.04369623214006424,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 3.0096778869628906,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.05180002748966217,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 3.0121543407440186,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.0440233014523983,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.9937305450439453,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.05645671859383583,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 3.0023865699768066,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.03990716487169266,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 3.0003232955932617,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.03818608075380325,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.994137763977051,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.035388972610235214,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.993170738220215,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.03162284567952156,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.9922614097595215,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.029497336596250534,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.9852986335754395,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.02586926706135273,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.9798672199249268,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.027355050668120384,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.985158920288086,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.0341978520154953,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.982581615447998,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.030727429315447807,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.9701411724090576,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.04429870471358299,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.9789254665374756,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.05775132030248642,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.9758143424987793,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.05635962635278702,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.981480121612549,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.03639192506670952,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.968752861022949,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.028748342767357826,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.9584972858428955,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.035653006285429,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.967545747756958,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.03697885945439339,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.9578099250793457,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.03518654406070709,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.9680328369140625,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.034762222319841385,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.9670262336730957,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03813016042113304,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.957522392272949,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.04372904822230339,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.9540514945983887,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.06646446138620377,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.951028823852539,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.07831092923879623,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.95898699760437,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.05779588967561722,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.965884208679199,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.04414296895265579,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.9510560035705566,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.04797815904021263,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.955803394317627,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.04482145980000496,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.954059362411499,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.037104446440935135,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.94792103767395,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.03616701066493988,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.9500932693481445,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.03882797807455063,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.9558305740356445,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.04356204345822334,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.9237029552459717,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.04535257816314697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.925853729248047,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.04519587755203247,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.924984931945801,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.055922407656908035,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.932561159133911,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.06330762803554535,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.9388322830200195,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.0517115443944931,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.9272942543029785,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.057369694113731384,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.9201431274414062,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.05687880516052246,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.922886371612549,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.055525410920381546,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.9211201667785645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04503652825951576,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.9164583683013916,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.03385302051901817,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.9145395755767822,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03564384952187538,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.916003704071045,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.04188712686300278,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.915839433670044,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03657370060682297,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.903120279312134,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.03126626834273338,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.895824432373047,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.04196290299296379,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.9093422889709473,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.1706087589263916,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.9183316230773926,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.05810421332716942,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.911245107650757,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.061330314725637436,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.9159231185913086,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.06062281131744385,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.9211535453796387,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.05590800568461418,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.9140496253967285,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.04496239125728607,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.9040920734405518,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.03835335746407509,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.9026548862457275,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.05137286335229874,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.889634132385254,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.04091101884841919,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.886838674545288,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.04076292738318443,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.892918109893799,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.045437734574079514,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.893305778503418,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.041114673018455505,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.88517165184021,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.03600773215293884,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.8779795169830322,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.04311220720410347,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.87247896194458,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.04818756878376007,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.889103412628174,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.053460877388715744,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.8899683952331543,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.06763825565576553,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.8823537826538086,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.07304801046848297,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.877340316772461,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.061337269842624664,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.8820111751556396,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.04660911113023758,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.868886709213257,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.053929202258586884,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.159337043762207,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.721762478351593,
+ "eval_runtime": 42.7977,
+ "eval_samples_per_second": 57.059,
+ "eval_steps_per_second": 0.467,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.04259369149804115,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.859579086303711,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.044425517320632935,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.8681139945983887,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.04301869496703148,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.853911876678467,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.04574296995997429,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.8724427223205566,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.050247322767972946,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.8528172969818115,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.046683840453624725,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.853405714035034,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.05966169387102127,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.8605127334594727,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.08935574442148209,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.862339973449707,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.08019179105758667,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.8655967712402344,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.04753892868757248,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.8657939434051514,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06868117302656174,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.861976146697998,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.07040470093488693,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.8562135696411133,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.05683179199695587,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.8576629161834717,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.04918001592159271,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.861943244934082,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.04921915754675865,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.8495378494262695,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.037651967257261276,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.8428618907928467,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.03522057458758354,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.8372769355773926,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.03111186809837818,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.8377151489257812,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.03906773030757904,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.846196174621582,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.02546556107699871,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.8288016319274902,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.03119821846485138,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.8334617614746094,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.030583377927541733,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.8209362030029297,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.029693076387047768,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.8267486095428467,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.028978517279028893,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.819244861602783,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.027871523052453995,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.8289647102355957,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.029287064447999,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.8235747814178467,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.024990517646074295,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.819016218185425,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.030980689451098442,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.828155040740967,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03721801936626434,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.813418388366699,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03738148882985115,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.808328866958618,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.03580411896109581,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.812070846557617,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.039036594331264496,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.807159423828125,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.044666968286037445,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.799992561340332,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.055589817464351654,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.804142951965332,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.06139020994305611,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.8182730674743652,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.07769318670034409,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.8166518211364746,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.06521332263946533,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.8142249584198,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.09083619713783264,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.8166933059692383,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.09737467020750046,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.8111555576324463,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.07148188352584839,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.8168413639068604,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.055516090244054794,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.8183553218841553,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.05338391661643982,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.8127028942108154,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.040785472840070724,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.8124775886535645,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.037135496735572815,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.8065357208251953,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.032455578446388245,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.798459053039551,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.03465662896633148,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.798649549484253,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.032026033848524094,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.7906997203826904,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.026019956916570663,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.791969060897827,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.030298635363578796,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.784593105316162,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.028147678822278976,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.782454013824463,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.023933131247758865,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.787851333618164,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.026789527386426926,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.788067579269409,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.024554884061217308,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.777033567428589,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.028679484501481056,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.786342144012451,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.026410218328237534,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.7790610790252686,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.027413051575422287,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.778944969177246,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.02908242493867874,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.763850212097168,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.029632994905114174,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.7672982215881348,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.027458716183900833,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.7702178955078125,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.03263958916068077,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.7581803798675537,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04427146911621094,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.768141508102417,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.07142394036054611,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.75819730758667,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.08833584934473038,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.782716751098633,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.08168833702802658,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.7763240337371826,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07253874838352203,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.7717690467834473,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.08186902105808258,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.767303466796875,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.05854064226150513,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.761544704437256,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.04602917283773422,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.7669177055358887,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.03920970857143402,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.766085624694824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.04551195353269577,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.7623450756073,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.040861159563064575,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.7471840381622314,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.026907291263341904,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 2.0710244178771973,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6934371590614319,
+ "eval_runtime": 42.6015,
+ "eval_samples_per_second": 57.322,
+ "eval_steps_per_second": 0.469,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.03801899403333664,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.739086627960205,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.03707238286733627,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.741638422012329,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.0387551411986351,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.740777015686035,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.028010359033942223,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.7353010177612305,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.026474518701434135,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.73842453956604,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.02920856513082981,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.736295461654663,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.026074763387441635,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.7295751571655273,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.026862047612667084,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.723040819168091,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.026300370693206787,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.7199485301971436,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.026030950248241425,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.7276217937469482,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.028342435136437416,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.7193045616149902,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.032135091722011566,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.7143118381500244,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.03068603202700615,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.7191050052642822,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.035828713327646255,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.718085765838623,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.03343002125620842,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.707131862640381,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.035051796585321426,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.7091879844665527,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.03395913913846016,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6980862617492676,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.034198079258203506,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.699704647064209,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.0322839580476284,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.702791213989258,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.041531339287757874,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.7018051147460938,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.040806882083415985,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.7096598148345947,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.04156940430402756,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6956775188446045,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.05608797073364258,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.70127272605896,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0729481428861618,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.708817720413208,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.0905836969614029,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.718085289001465,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.08345075696706772,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.7235910892486572,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.08599896728992462,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.7178783416748047,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.07301691174507141,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.714029312133789,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.08231015503406525,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.7196826934814453,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.08286412060260773,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.7176082134246826,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.08590883016586304,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.7242164611816406,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.07531397044658661,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.7225024700164795,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.04711190238595009,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.695887327194214,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.05076327919960022,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.7029781341552734,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.04495669901371002,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.7057900428771973,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.04312257468700409,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.697537899017334,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.043580975383520126,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.67315673828125,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.03360886126756668,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.6883413791656494,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.030360296368598938,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.700927257537842,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.0310810673981905,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.6917803287506104,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.02703232690691948,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.6667580604553223,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.02756045199930668,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.6702821254730225,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.02611083723604679,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.669790744781494,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.02615324780344963,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.6672821044921875,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.026000119745731354,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.6671881675720215,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.02882794849574566,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.67586088180542,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03327213600277901,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.6618690490722656,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.032920945435762405,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.665444850921631,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.02812962979078293,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.6680707931518555,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.025886602699756622,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.652898073196411,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.025222789496183395,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.6560399532318115,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.023131893947720528,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.66127347946167,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.02513662725687027,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.6463732719421387,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.027243830263614655,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.6507630348205566,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.025437965989112854,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.6485800743103027,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.02770340070128441,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.657241106033325,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.03594067320227623,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.6281137466430664,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04486451297998428,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.6340909004211426,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.05102410912513733,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.6463394165039062,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04514491558074951,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.636554002761841,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04469453915953636,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.6453680992126465,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.05421523004770279,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.646010398864746,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04779462888836861,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.6315622329711914,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.06697724014520645,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.6416730880737305,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.07520416378974915,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.641632318496704,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.055692676454782486,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.6289877891540527,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.04539895057678223,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.6363251209259033,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.04829065501689911,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.641608476638794,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.041314542293548584,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.6269643306732178,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.04100198298692703,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.6243174076080322,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.04105962812900543,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.626585006713867,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.03172122687101364,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9550899267196655,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6614077091217041,
+ "eval_runtime": 42.9009,
+ "eval_samples_per_second": 56.922,
+ "eval_steps_per_second": 0.466,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.03258531913161278,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.612126588821411,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.040250636637210846,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.6025328636169434,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.034234337508678436,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.6085972785949707,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.03635786101222038,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.6126222610473633,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.039404451847076416,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.6035327911376953,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.038504838943481445,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5985798835754395,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.03744236007332802,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5873115062713623,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.034184787422418594,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.5899314880371094,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.03766429424285889,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.602734327316284,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.03741588443517685,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.592207193374634,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.03187853842973709,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5850372314453125,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.038447409868240356,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.592654228210449,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.04046263173222542,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.5881710052490234,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.03872602432966232,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5881471633911133,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.03960775211453438,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5826945304870605,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.033132150769233704,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.584038734436035,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.029504230245947838,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.584357261657715,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.03665253147482872,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5677168369293213,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.037208545953035355,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.560795783996582,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.038244519382715225,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.591458559036255,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.043307431042194366,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.5642588138580322,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.055119726806879044,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.57377290725708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.053323689848184586,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5767312049865723,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.0502864345908165,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.562190055847168,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.05568842589855194,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.565556049346924,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.061149828135967255,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5883288383483887,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.059933699667453766,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.5710184574127197,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.06041933596134186,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.572603225708008,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.05482695251703262,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.5667271614074707,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.04712968319654465,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.554041624069214,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.04842432215809822,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.5660672187805176,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.04273047298192978,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.554839849472046,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.05356749892234802,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.5521483421325684,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.06062676012516022,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.5551886558532715,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.05431462824344635,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5654542446136475,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.0505850687623024,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5648036003112793,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.04696856811642647,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.5559377670288086,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.042585279792547226,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.5660319328308105,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.04026484116911888,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.551206588745117,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03382421284914017,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.556361198425293,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.031177924945950508,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.5413894653320312,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.028217196464538574,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.5600497722625732,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.027062954381108284,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5447568893432617,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.03001212701201439,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.540614604949951,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.028869854286313057,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.544583797454834,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.02869918756186962,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.5471596717834473,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03359716013073921,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5416982173919678,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.04000182822346687,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.534921407699585,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.043030042201280594,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5368564128875732,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.04329848289489746,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.539764881134033,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.03740967437624931,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.5246665477752686,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.03733935207128525,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.529362916946411,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.039609868079423904,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.542544364929199,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.03937261551618576,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5366649627685547,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.03261927142739296,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.5250535011291504,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.03904294595122337,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.520721435546875,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.03913239762187004,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.5122013092041016,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.03610851988196373,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.5212767124176025,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.03944548964500427,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.537186622619629,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04762452840805054,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.5204620361328125,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04427091404795647,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5286149978637695,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04366641491651535,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.5398361682891846,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.04644155502319336,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.519331932067871,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.04372382536530495,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.531022787094116,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.04404324293136597,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.521758794784546,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.042115990072488785,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.5146074295043945,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.03796013817191124,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.5040054321289062,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.04327977076172829,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.5209312438964844,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04150385782122612,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.515871286392212,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.04465346783399582,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.5247764587402344,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.05690346285700798,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.501911163330078,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04907930642366409,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8968371152877808,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6358620524406433,
+ "eval_runtime": 42.9991,
+ "eval_samples_per_second": 56.792,
+ "eval_steps_per_second": 0.465,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.03755255416035652,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.486576795578003,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.040234558284282684,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.501087188720703,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.041416723281145096,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4942233562469482,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.039655014872550964,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.503453254699707,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.041707683354616165,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.490978717803955,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.04545537009835243,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4757118225097656,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.04328389838337898,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.504516839981079,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.043312299996614456,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.4920215606689453,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.046466976404190063,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.494760513305664,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.041762981563806534,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4875552654266357,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.03713781014084816,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.495652675628662,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.040603458881378174,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4821407794952393,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.037969645112752914,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.484981060028076,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.03652074560523033,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4736409187316895,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.03783264383673668,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.483346700668335,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.03404324874281883,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.48856782913208,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.03294725716114044,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.483706474304199,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.0347001850605011,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.474888324737549,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.0359107106924057,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.478990077972412,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.03738103806972504,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.4709420204162598,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.03722188249230385,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.465726852416992,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.034907322376966476,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.482231855392456,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.035695310682058334,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4785828590393066,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.04009179025888443,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.4795875549316406,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.038996778428554535,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.478696823120117,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04082407429814339,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.465200901031494,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04856136813759804,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4605681896209717,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.05671004578471184,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.47277569770813,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.0481606163084507,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.476151466369629,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.037689510732889175,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.466142177581787,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.05005146935582161,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.4798035621643066,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.04745829105377197,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.46163272857666,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.04638158157467842,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4728071689605713,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.04530330374836922,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.4621877670288086,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.03971488028764725,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.474224090576172,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.04521876201033592,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.46122670173645,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03923020884394646,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.4603679180145264,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03986494988203049,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.4589664936065674,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.04280020296573639,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.4624319076538086,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.03930043429136276,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.47080135345459,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.0314876064658165,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.458728790283203,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.03189660236239433,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4467148780822754,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03264275938272476,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4613699913024902,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03021484799683094,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.4535279273986816,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.030740290880203247,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.4621875286102295,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03096810355782509,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.458500385284424,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03105205111205578,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.461958646774292,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.03330068290233612,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.449777364730835,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.032831475138664246,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.4616191387176514,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.0387776717543602,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.445256471633911,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.041803423315286636,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.4545960426330566,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.04148292914032936,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.433408737182617,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.04648579657077789,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.449134349822998,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.04186409339308739,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.4449565410614014,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.038569554686546326,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.443675994873047,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03908300772309303,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4626526832580566,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.038897737860679626,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.438591957092285,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.0360480472445488,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.436077117919922,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0429830439388752,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4384970664978027,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.04465422406792641,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.458029270172119,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.037096526473760605,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.436816930770874,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.047823466360569,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.44439959526062,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.05929310992360115,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.4599215984344482,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.053512319922447205,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4387903213500977,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.042625587433576584,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4359588623046875,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.056042976677417755,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.4441447257995605,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05813189223408699,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4399144649505615,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.06313478201627731,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.453447103500366,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.05281798169016838,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.4428963661193848,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.03991910442709923,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4359219074249268,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.047981832176446915,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4506607055664062,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.03646523132920265,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8225739002227783,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6196661591529846,
+ "eval_runtime": 43.8106,
+ "eval_samples_per_second": 55.74,
+ "eval_steps_per_second": 0.457,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.03673036769032478,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.4227209091186523,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.044547371566295624,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.4073169231414795,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.04089285805821419,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4279446601867676,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.03538106009364128,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.4234392642974854,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.03152978792786598,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4214606285095215,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.03324202075600624,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.4085447788238525,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.03352442383766174,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.417776107788086,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.03582802042365074,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.411144256591797,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.0351114459335804,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.405872344970703,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.038898173719644547,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4029338359832764,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.03926656022667885,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4100089073181152,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.04285634309053421,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.410360336303711,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.04179581627249718,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4093637466430664,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.03356701880693436,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4118056297302246,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.038140635937452316,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.4154677391052246,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.04468780383467674,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416846513748169,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.062421441078186035,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4157638549804688,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06317313015460968,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.418593406677246,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.049498945474624634,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4204845428466797,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.046322405338287354,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.403489589691162,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.053081002086400986,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.4148740768432617,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05222945287823677,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.416757583618164,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04893574118614197,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4241795539855957,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04951956123113632,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.401386260986328,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04019852355122566,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.4168622493743896,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.03625848889350891,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4026846885681152,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.03464345261454582,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4021215438842773,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.03750818222761154,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4113571643829346,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.0360100194811821,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.3977622985839844,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.03464623540639877,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.413273334503174,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.03193575143814087,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.423659324645996,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.02710454724729061,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.396085262298584,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.02830432914197445,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.420255422592163,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.029717663303017616,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.3958120346069336,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.030005933716893196,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3908543586730957,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.029054716229438782,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.3873138427734375,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.031556472182273865,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3892760276794434,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03518872708082199,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4113478660583496,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.03471249341964722,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.384857654571533,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03223155066370964,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3840765953063965,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.030012644827365875,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.388247489929199,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.03162285313010216,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.3987512588500977,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.03700442612171173,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.3993430137634277,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.03710822016000748,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.387359380722046,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.028411181643605232,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3925249576568604,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031246855854988098,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.38845157623291,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.034664880484342575,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.392436981201172,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.03354701027274132,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.390890121459961,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.03527764603495598,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.3954997062683105,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03325049206614494,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3847177028656006,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.028021875768899918,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.385910987854004,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.02964954823255539,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3787484169006348,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.03068520873785019,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.400784969329834,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03416873514652252,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.391033411026001,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.03534875437617302,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.3796474933624268,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.03732474148273468,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3831934928894043,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.03648678958415985,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3785600662231445,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03801294043660164,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.387587070465088,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.04186025261878967,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.3893089294433594,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.04416101053357124,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3852195739746094,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.04137178510427475,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.391308307647705,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.05010323226451874,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.3906192779541016,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.049587685614824295,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.3846192359924316,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.04125971347093582,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.384507179260254,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.039676666259765625,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.38167142868042,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.038775283843278885,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.3789167404174805,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.03634221479296684,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.383060932159424,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03464275971055031,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.388552665710449,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.031778786331415176,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.3903298377990723,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.027904309332370758,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.377493381500244,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.028964119032025337,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.381959915161133,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.0233962032943964,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7810559272766113,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6070130467414856,
+ "eval_runtime": 50.871,
+ "eval_samples_per_second": 48.004,
+ "eval_steps_per_second": 0.393,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.030209315940737724,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.3569841384887695,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.04112662002444267,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.376680374145508,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.055858466774225235,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.3621644973754883,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.06746986508369446,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3607563972473145,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.06736797094345093,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3548707962036133,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.0562627948820591,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.364654541015625,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.051833927631378174,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.3748884201049805,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.04824110120534897,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.3502087593078613,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.05518793687224388,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3687515258789062,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.06564056873321533,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.367131233215332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.0605909489095211,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.3589868545532227,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.05744732916355133,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.370117664337158,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.06430432945489883,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.3703484535217285,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.06186915934085846,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3720293045043945,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.05977998673915863,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.3512699604034424,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.06413097679615021,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.377257823944092,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.053935080766677856,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.350717544555664,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.049110863357782364,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.3688902854919434,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.04313219338655472,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3374459743499756,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.039556123316287994,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3482556343078613,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.03796547278761864,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3547937870025635,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.03646104037761688,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.344686269760132,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.03355357423424721,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.3588008880615234,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.03004707582294941,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.352043628692627,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.030186176300048828,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3500709533691406,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.029748762026429176,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.3523058891296387,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.03238696977496147,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3635129928588867,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.032599564641714096,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.346269130706787,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.028626957908272743,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.357151746749878,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.02909751422703266,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3557677268981934,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.029951099306344986,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.34753680229187,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.029768310487270355,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3408985137939453,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.03005973994731903,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.362542152404785,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03360462561249733,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.352529287338257,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.034002047032117844,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.3334460258483887,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.03292316198348999,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3446991443634033,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.0337117575109005,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.350605010986328,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03428212180733681,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.357408046722412,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.03169601410627365,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.349088191986084,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.03389411047101021,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.361576795578003,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.036497049033641815,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.35420560836792,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03574785590171814,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3678526878356934,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.030757993459701538,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3520517349243164,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.0289121363312006,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.337864637374878,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.028012607246637344,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.353999614715576,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.027558492496609688,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.342132568359375,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.02871510572731495,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3425655364990234,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03070886805653572,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3401665687561035,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.03056657873094082,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.345277786254883,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.030907969921827316,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.346494674682617,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0318334735929966,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.3390543460845947,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.029700733721256256,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.3398728370666504,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.029687359929084778,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.3474364280700684,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.027234600856900215,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3457484245300293,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.02823043428361416,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.337700366973877,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03002646192908287,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.335954189300537,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.027246717363595963,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.330207347869873,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.029703574255108833,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.328280448913574,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.028796236962080002,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3448386192321777,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.027944380417466164,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.320234537124634,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.030626650899648666,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.334698438644409,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.029164662584662437,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3450326919555664,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03034045360982418,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.3420588970184326,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.02856968529522419,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3396072387695312,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.02892562374472618,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.340027332305908,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.02938242256641388,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.33550763130188,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.029261089861392975,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3246421813964844,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.030340535566210747,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.3465514183044434,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.033335570245981216,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.3467206954956055,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.03452218323945999,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.343702793121338,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.03326255828142166,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.337631940841675,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025433633476495743,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7402067184448242,
+ "step": 720
+ },
+ {
+ "epoch": 10.013961605584642,
+ "grad_norm": 0.03175730258226395,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.62623405456543,
+ "step": 721
+ },
+ {
+ "epoch": 10.027923211169284,
+ "grad_norm": 0.04780422896146774,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.623239517211914,
+ "step": 722
+ },
+ {
+ "epoch": 10.041884816753926,
+ "grad_norm": 0.0504758283495903,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.6369194984436035,
+ "step": 723
+ },
+ {
+ "epoch": 10.055846422338568,
+ "grad_norm": 0.04481920227408409,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.57640266418457,
+ "step": 724
+ },
+ {
+ "epoch": 10.06980802792321,
+ "grad_norm": 0.0440552793443203,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.630992889404297,
+ "step": 725
+ },
+ {
+ "epoch": 10.083769633507853,
+ "grad_norm": 0.0503336526453495,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.627910137176514,
+ "step": 726
+ },
+ {
+ "epoch": 10.097731239092496,
+ "grad_norm": 0.053885314613580704,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.637897491455078,
+ "step": 727
+ },
+ {
+ "epoch": 10.111692844677139,
+ "grad_norm": 0.05481956899166107,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.610715866088867,
+ "step": 728
+ },
+ {
+ "epoch": 10.12565445026178,
+ "grad_norm": 0.050510745495557785,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.639884948730469,
+ "step": 729
+ },
+ {
+ "epoch": 10.139616055846423,
+ "grad_norm": 0.04579266160726547,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.590188980102539,
+ "step": 730
+ },
+ {
+ "epoch": 10.153577661431065,
+ "grad_norm": 0.04817410930991173,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.636295318603516,
+ "step": 731
+ },
+ {
+ "epoch": 10.167539267015707,
+ "grad_norm": 0.058862220495939255,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.614317417144775,
+ "step": 732
+ },
+ {
+ "epoch": 10.181500872600349,
+ "grad_norm": 0.06328638643026352,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.655928611755371,
+ "step": 733
+ },
+ {
+ "epoch": 10.195462478184991,
+ "grad_norm": 0.058752477169036865,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.652004241943359,
+ "step": 734
+ },
+ {
+ "epoch": 10.209424083769633,
+ "grad_norm": 0.06049015372991562,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.6513800621032715,
+ "step": 735
+ },
+ {
+ "epoch": 10.223385689354275,
+ "grad_norm": 0.058941151946783066,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.657688140869141,
+ "step": 736
+ },
+ {
+ "epoch": 10.237347294938917,
+ "grad_norm": 0.05542639642953873,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.637376308441162,
+ "step": 737
+ },
+ {
+ "epoch": 10.25130890052356,
+ "grad_norm": 0.06024057790637016,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.63592529296875,
+ "step": 738
+ },
+ {
+ "epoch": 10.265270506108202,
+ "grad_norm": 0.06040216609835625,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.6275787353515625,
+ "step": 739
+ },
+ {
+ "epoch": 10.279232111692846,
+ "grad_norm": 0.059600673615932465,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.652995586395264,
+ "step": 740
+ },
+ {
+ "epoch": 10.293193717277488,
+ "grad_norm": 0.05379730463027954,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.597358226776123,
+ "step": 741
+ },
+ {
+ "epoch": 10.30715532286213,
+ "grad_norm": 0.05355260521173477,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.6106367111206055,
+ "step": 742
+ },
+ {
+ "epoch": 10.321116928446772,
+ "grad_norm": 0.06649646162986755,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.654443740844727,
+ "step": 743
+ },
+ {
+ "epoch": 10.335078534031414,
+ "grad_norm": 0.058309994637966156,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.6869659423828125,
+ "step": 744
+ },
+ {
+ "epoch": 10.349040139616056,
+ "grad_norm": 0.05403570458292961,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.654750823974609,
+ "step": 745
+ },
+ {
+ "epoch": 10.363001745200698,
+ "grad_norm": 0.04678157716989517,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.653955459594727,
+ "step": 746
+ },
+ {
+ "epoch": 10.37696335078534,
+ "grad_norm": 0.04297948628664017,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.644915580749512,
+ "step": 747
+ },
+ {
+ "epoch": 10.390924956369982,
+ "grad_norm": 0.04235142841935158,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.662282943725586,
+ "step": 748
+ },
+ {
+ "epoch": 10.404886561954624,
+ "grad_norm": 0.0405130498111248,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.64569091796875,
+ "step": 749
+ },
+ {
+ "epoch": 10.418848167539267,
+ "grad_norm": 0.03351007401943207,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.653165817260742,
+ "step": 750
+ },
+ {
+ "epoch": 10.432809773123909,
+ "grad_norm": 0.03182631731033325,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.606291770935059,
+ "step": 751
+ },
+ {
+ "epoch": 10.44677137870855,
+ "grad_norm": 0.031552448868751526,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.642643928527832,
+ "step": 752
+ },
+ {
+ "epoch": 10.460732984293193,
+ "grad_norm": 0.027941036969423294,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.623664379119873,
+ "step": 753
+ },
+ {
+ "epoch": 10.474694589877837,
+ "grad_norm": 0.027744563296437263,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.605886459350586,
+ "step": 754
+ },
+ {
+ "epoch": 10.488656195462479,
+ "grad_norm": 0.02962736040353775,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.606762886047363,
+ "step": 755
+ },
+ {
+ "epoch": 10.502617801047121,
+ "grad_norm": 0.027251560240983963,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.620492935180664,
+ "step": 756
+ },
+ {
+ "epoch": 10.516579406631763,
+ "grad_norm": 0.027894044294953346,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.613903999328613,
+ "step": 757
+ },
+ {
+ "epoch": 10.530541012216405,
+ "grad_norm": 0.02778882533311844,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.634457588195801,
+ "step": 758
+ },
+ {
+ "epoch": 10.544502617801047,
+ "grad_norm": 0.028082096949219704,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.610980033874512,
+ "step": 759
+ },
+ {
+ "epoch": 10.55846422338569,
+ "grad_norm": 0.026210976764559746,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.612090110778809,
+ "step": 760
+ },
+ {
+ "epoch": 10.572425828970331,
+ "grad_norm": 0.026537755504250526,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.581171035766602,
+ "step": 761
+ },
+ {
+ "epoch": 10.586387434554974,
+ "grad_norm": 0.028709523379802704,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.628799915313721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600349040139616,
+ "grad_norm": 0.028233813121914864,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.586182117462158,
+ "step": 763
+ },
+ {
+ "epoch": 10.614310645724258,
+ "grad_norm": 0.02966046892106533,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.62320613861084,
+ "step": 764
+ },
+ {
+ "epoch": 10.6282722513089,
+ "grad_norm": 0.03198636695742607,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.6018524169921875,
+ "step": 765
+ },
+ {
+ "epoch": 10.642233856893542,
+ "grad_norm": 0.033404890447854996,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.610121726989746,
+ "step": 766
+ },
+ {
+ "epoch": 10.656195462478184,
+ "grad_norm": 0.029217196628451347,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.601312160491943,
+ "step": 767
+ },
+ {
+ "epoch": 10.670157068062828,
+ "grad_norm": 0.029335757717490196,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.620820999145508,
+ "step": 768
+ },
+ {
+ "epoch": 10.68411867364747,
+ "grad_norm": 0.031007781624794006,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.626503944396973,
+ "step": 769
+ },
+ {
+ "epoch": 10.698080279232112,
+ "grad_norm": 0.028453458100557327,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.589323997497559,
+ "step": 770
+ },
+ {
+ "epoch": 10.712041884816754,
+ "grad_norm": 0.02526312880218029,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.598350524902344,
+ "step": 771
+ },
+ {
+ "epoch": 10.726003490401396,
+ "grad_norm": 0.027317512780427933,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.62388277053833,
+ "step": 772
+ },
+ {
+ "epoch": 10.739965095986038,
+ "grad_norm": 0.02710605598986149,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.606523513793945,
+ "step": 773
+ },
+ {
+ "epoch": 10.75392670157068,
+ "grad_norm": 0.027491388842463493,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.586644172668457,
+ "step": 774
+ },
+ {
+ "epoch": 10.767888307155323,
+ "grad_norm": 0.027158567681908607,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.606039047241211,
+ "step": 775
+ },
+ {
+ "epoch": 10.781849912739965,
+ "grad_norm": 0.0298276599496603,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.593776702880859,
+ "step": 776
+ },
+ {
+ "epoch": 10.795811518324607,
+ "grad_norm": 0.026622479781508446,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.59510612487793,
+ "step": 777
+ },
+ {
+ "epoch": 10.809773123909249,
+ "grad_norm": 0.028238043189048767,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.5899505615234375,
+ "step": 778
+ },
+ {
+ "epoch": 10.823734729493891,
+ "grad_norm": 0.029481008648872375,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.6092071533203125,
+ "step": 779
+ },
+ {
+ "epoch": 10.837696335078533,
+ "grad_norm": 0.030138323083519936,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.600069046020508,
+ "step": 780
+ },
+ {
+ "epoch": 10.851657940663177,
+ "grad_norm": 0.02951825223863125,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.600125789642334,
+ "step": 781
+ },
+ {
+ "epoch": 10.86561954624782,
+ "grad_norm": 0.029242202639579773,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.590723037719727,
+ "step": 782
+ },
+ {
+ "epoch": 10.879581151832461,
+ "grad_norm": 0.02941613830626011,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.623342514038086,
+ "step": 783
+ },
+ {
+ "epoch": 10.893542757417103,
+ "grad_norm": 0.029764151200652122,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.631185531616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.907504363001745,
+ "grad_norm": 0.031503792852163315,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.605094909667969,
+ "step": 785
+ },
+ {
+ "epoch": 10.921465968586388,
+ "grad_norm": 0.03184918314218521,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.592767715454102,
+ "step": 786
+ },
+ {
+ "epoch": 10.93542757417103,
+ "grad_norm": 0.03223618119955063,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.5930585861206055,
+ "step": 787
+ },
+ {
+ "epoch": 10.949389179755672,
+ "grad_norm": 0.0340263694524765,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.609347343444824,
+ "step": 788
+ },
+ {
+ "epoch": 10.963350785340314,
+ "grad_norm": 0.03245621174573898,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.598440170288086,
+ "step": 789
+ },
+ {
+ "epoch": 10.977312390924956,
+ "grad_norm": 0.03204867243766785,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.619702339172363,
+ "step": 790
+ },
+ {
+ "epoch": 10.991273996509598,
+ "grad_norm": 0.030866602435708046,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.609165191650391,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.01960863545536995,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 2.8655829429626465,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.593852162361145,
+ "eval_runtime": 60.0971,
+ "eval_samples_per_second": 40.634,
+ "eval_steps_per_second": 0.649,
+ "step": 792
+ },
+ {
+ "epoch": 11.013961605584642,
+ "grad_norm": 0.03552858531475067,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.539240837097168,
+ "step": 793
+ },
+ {
+ "epoch": 11.027923211169284,
+ "grad_norm": 0.0536109134554863,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.52609395980835,
+ "step": 794
+ },
+ {
+ "epoch": 11.041884816753926,
+ "grad_norm": 0.061411239206790924,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.536109447479248,
+ "step": 795
+ },
+ {
+ "epoch": 11.055846422338568,
+ "grad_norm": 0.06352879852056503,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.546850204467773,
+ "step": 796
+ },
+ {
+ "epoch": 11.06980802792321,
+ "grad_norm": 0.05683520436286926,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.563828468322754,
+ "step": 797
+ },
+ {
+ "epoch": 11.083769633507853,
+ "grad_norm": 0.05522637069225311,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.550929069519043,
+ "step": 798
+ },
+ {
+ "epoch": 11.097731239092496,
+ "grad_norm": 0.058990392833948135,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.554411888122559,
+ "step": 799
+ },
+ {
+ "epoch": 11.111692844677139,
+ "grad_norm": 0.057443976402282715,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.545921325683594,
+ "step": 800
+ },
+ {
+ "epoch": 11.12565445026178,
+ "grad_norm": 0.05383576825261116,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.5481414794921875,
+ "step": 801
+ },
+ {
+ "epoch": 11.139616055846423,
+ "grad_norm": 0.04797440022230148,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.55556583404541,
+ "step": 802
+ },
+ {
+ "epoch": 11.153577661431065,
+ "grad_norm": 0.04923664405941963,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.545513153076172,
+ "step": 803
+ },
+ {
+ "epoch": 11.167539267015707,
+ "grad_norm": 0.056191347539424896,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.536705017089844,
+ "step": 804
+ },
+ {
+ "epoch": 11.181500872600349,
+ "grad_norm": 0.06088989973068237,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.570624351501465,
+ "step": 805
+ },
+ {
+ "epoch": 11.195462478184991,
+ "grad_norm": 0.04802761599421501,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.526930809020996,
+ "step": 806
+ },
+ {
+ "epoch": 11.209424083769633,
+ "grad_norm": 0.04018320515751839,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.562432289123535,
+ "step": 807
+ },
+ {
+ "epoch": 11.223385689354275,
+ "grad_norm": 0.04280981421470642,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.559737205505371,
+ "step": 808
+ },
+ {
+ "epoch": 11.237347294938917,
+ "grad_norm": 0.04258878529071808,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.561749458312988,
+ "step": 809
+ },
+ {
+ "epoch": 11.25130890052356,
+ "grad_norm": 0.04411563277244568,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.562010765075684,
+ "step": 810
+ },
+ {
+ "epoch": 11.265270506108202,
+ "grad_norm": 0.041077401489019394,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.573984622955322,
+ "step": 811
+ },
+ {
+ "epoch": 11.279232111692846,
+ "grad_norm": 0.04125386103987694,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.550564289093018,
+ "step": 812
+ },
+ {
+ "epoch": 11.293193717277488,
+ "grad_norm": 0.041974026709795,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.552739143371582,
+ "step": 813
+ },
+ {
+ "epoch": 11.30715532286213,
+ "grad_norm": 0.03689907118678093,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.540107727050781,
+ "step": 814
+ },
+ {
+ "epoch": 11.321116928446772,
+ "grad_norm": 0.03586242347955704,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.554999351501465,
+ "step": 815
+ },
+ {
+ "epoch": 11.335078534031414,
+ "grad_norm": 0.036602675914764404,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.546337127685547,
+ "step": 816
+ },
+ {
+ "epoch": 11.349040139616056,
+ "grad_norm": 0.03707926720380783,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.543377876281738,
+ "step": 817
+ },
+ {
+ "epoch": 11.363001745200698,
+ "grad_norm": 0.03322125971317291,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.574325084686279,
+ "step": 818
+ },
+ {
+ "epoch": 11.37696335078534,
+ "grad_norm": 0.03359243646264076,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.5613603591918945,
+ "step": 819
+ },
+ {
+ "epoch": 11.390924956369982,
+ "grad_norm": 0.0314965583384037,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.530994415283203,
+ "step": 820
+ },
+ {
+ "epoch": 11.404886561954624,
+ "grad_norm": 0.031806789338588715,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.56138801574707,
+ "step": 821
+ },
+ {
+ "epoch": 11.418848167539267,
+ "grad_norm": 0.034988485276699066,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.553149700164795,
+ "step": 822
+ },
+ {
+ "epoch": 11.432809773123909,
+ "grad_norm": 0.037627559155225754,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.555401802062988,
+ "step": 823
+ },
+ {
+ "epoch": 11.44677137870855,
+ "grad_norm": 0.03723530471324921,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.52452278137207,
+ "step": 824
+ },
+ {
+ "epoch": 11.460732984293193,
+ "grad_norm": 0.036220453679561615,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.54799747467041,
+ "step": 825
+ },
+ {
+ "epoch": 11.474694589877837,
+ "grad_norm": 0.03228041157126427,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.574814319610596,
+ "step": 826
+ },
+ {
+ "epoch": 11.488656195462479,
+ "grad_norm": 0.030872676521539688,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.539666652679443,
+ "step": 827
+ },
+ {
+ "epoch": 11.502617801047121,
+ "grad_norm": 0.033935241401195526,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.579252243041992,
+ "step": 828
+ },
+ {
+ "epoch": 11.516579406631763,
+ "grad_norm": 0.034338828176259995,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.538324356079102,
+ "step": 829
+ },
+ {
+ "epoch": 11.530541012216405,
+ "grad_norm": 0.03441951423883438,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.540358543395996,
+ "step": 830
+ },
+ {
+ "epoch": 11.544502617801047,
+ "grad_norm": 0.03328141197562218,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.520340919494629,
+ "step": 831
+ },
+ {
+ "epoch": 11.55846422338569,
+ "grad_norm": 0.03224223107099533,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.5646209716796875,
+ "step": 832
+ },
+ {
+ "epoch": 11.572425828970331,
+ "grad_norm": 0.02999890223145485,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.504915714263916,
+ "step": 833
+ },
+ {
+ "epoch": 11.586387434554974,
+ "grad_norm": 0.03014362044632435,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.516615867614746,
+ "step": 834
+ },
+ {
+ "epoch": 11.600349040139616,
+ "grad_norm": 0.029469173401594162,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.560475826263428,
+ "step": 835
+ },
+ {
+ "epoch": 11.614310645724258,
+ "grad_norm": 0.027887821197509766,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.586333751678467,
+ "step": 836
+ },
+ {
+ "epoch": 11.6282722513089,
+ "grad_norm": 0.029970254749059677,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.530396461486816,
+ "step": 837
+ },
+ {
+ "epoch": 11.642233856893542,
+ "grad_norm": 0.031648170202970505,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.554404258728027,
+ "step": 838
+ },
+ {
+ "epoch": 11.656195462478184,
+ "grad_norm": 0.0325581431388855,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.544102668762207,
+ "step": 839
+ },
+ {
+ "epoch": 11.670157068062828,
+ "grad_norm": 0.03555625304579735,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.543701171875,
+ "step": 840
+ },
+ {
+ "epoch": 11.68411867364747,
+ "grad_norm": 0.033053912222385406,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.508375644683838,
+ "step": 841
+ },
+ {
+ "epoch": 11.698080279232112,
+ "grad_norm": 0.03059076890349388,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.545750617980957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712041884816754,
+ "grad_norm": 0.029640182852745056,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.54715633392334,
+ "step": 843
+ },
+ {
+ "epoch": 11.726003490401396,
+ "grad_norm": 0.030139951035380363,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.533708572387695,
+ "step": 844
+ },
+ {
+ "epoch": 11.739965095986038,
+ "grad_norm": 0.029753828421235085,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.543603897094727,
+ "step": 845
+ },
+ {
+ "epoch": 11.75392670157068,
+ "grad_norm": 0.029481034725904465,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.534200668334961,
+ "step": 846
+ },
+ {
+ "epoch": 11.767888307155323,
+ "grad_norm": 0.028991524130105972,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.5683674812316895,
+ "step": 847
+ },
+ {
+ "epoch": 11.781849912739965,
+ "grad_norm": 0.028145765885710716,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.543053150177002,
+ "step": 848
+ },
+ {
+ "epoch": 11.795811518324607,
+ "grad_norm": 0.02999844029545784,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.5482354164123535,
+ "step": 849
+ },
+ {
+ "epoch": 11.809773123909249,
+ "grad_norm": 0.03134117275476456,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.5360107421875,
+ "step": 850
+ },
+ {
+ "epoch": 11.823734729493891,
+ "grad_norm": 0.029277140274643898,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.5729827880859375,
+ "step": 851
+ },
+ {
+ "epoch": 11.837696335078533,
+ "grad_norm": 0.028516866266727448,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.543548583984375,
+ "step": 852
+ },
+ {
+ "epoch": 11.851657940663177,
+ "grad_norm": 0.028387080878019333,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.552616596221924,
+ "step": 853
+ },
+ {
+ "epoch": 11.86561954624782,
+ "grad_norm": 0.028977863490581512,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.5456695556640625,
+ "step": 854
+ },
+ {
+ "epoch": 11.879581151832461,
+ "grad_norm": 0.028010815382003784,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.534921646118164,
+ "step": 855
+ },
+ {
+ "epoch": 11.893542757417103,
+ "grad_norm": 0.030222605913877487,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.561526298522949,
+ "step": 856
+ },
+ {
+ "epoch": 11.907504363001745,
+ "grad_norm": 0.03140007704496384,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.500124931335449,
+ "step": 857
+ },
+ {
+ "epoch": 11.921465968586388,
+ "grad_norm": 0.029269184917211533,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.536653995513916,
+ "step": 858
+ },
+ {
+ "epoch": 11.93542757417103,
+ "grad_norm": 0.02935507707297802,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.521398544311523,
+ "step": 859
+ },
+ {
+ "epoch": 11.949389179755672,
+ "grad_norm": 0.02801917865872383,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.527442455291748,
+ "step": 860
+ },
+ {
+ "epoch": 11.963350785340314,
+ "grad_norm": 0.027995768934488297,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.556139945983887,
+ "step": 861
+ },
+ {
+ "epoch": 11.977312390924956,
+ "grad_norm": 0.028441084548830986,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.555442810058594,
+ "step": 862
+ },
+ {
+ "epoch": 11.991273996509598,
+ "grad_norm": 0.029100807383656502,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.532083511352539,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.020385215058922768,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 2.8488006591796875,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.5902590751647949,
+ "eval_runtime": 60.1424,
+ "eval_samples_per_second": 40.604,
+ "eval_steps_per_second": 0.648,
+ "step": 864
+ },
+ {
+ "epoch": 12.013961605584642,
+ "grad_norm": 0.03513244539499283,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.486500263214111,
+ "step": 865
+ },
+ {
+ "epoch": 12.027923211169284,
+ "grad_norm": 0.052679311484098434,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.489691734313965,
+ "step": 866
+ },
+ {
+ "epoch": 12.041884816753926,
+ "grad_norm": 0.06171489134430885,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.471555709838867,
+ "step": 867
+ },
+ {
+ "epoch": 12.055846422338568,
+ "grad_norm": 0.06164781004190445,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.486891746520996,
+ "step": 868
+ },
+ {
+ "epoch": 12.06980802792321,
+ "grad_norm": 0.06410388648509979,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.494929313659668,
+ "step": 869
+ },
+ {
+ "epoch": 12.083769633507853,
+ "grad_norm": 0.06399523466825485,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.5334320068359375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097731239092496,
+ "grad_norm": 0.0636502280831337,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.5039381980896,
+ "step": 871
+ },
+ {
+ "epoch": 12.111692844677139,
+ "grad_norm": 0.06139931082725525,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.512765884399414,
+ "step": 872
+ },
+ {
+ "epoch": 12.12565445026178,
+ "grad_norm": 0.07460282742977142,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.477229118347168,
+ "step": 873
+ },
+ {
+ "epoch": 12.139616055846423,
+ "grad_norm": 0.075187548995018,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.484212398529053,
+ "step": 874
+ },
+ {
+ "epoch": 12.153577661431065,
+ "grad_norm": 0.07584068924188614,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.495598793029785,
+ "step": 875
+ },
+ {
+ "epoch": 12.167539267015707,
+ "grad_norm": 0.07192158699035645,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.507598400115967,
+ "step": 876
+ },
+ {
+ "epoch": 12.181500872600349,
+ "grad_norm": 0.05366869643330574,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.514532089233398,
+ "step": 877
+ },
+ {
+ "epoch": 12.195462478184991,
+ "grad_norm": 0.06397061794996262,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.512920379638672,
+ "step": 878
+ },
+ {
+ "epoch": 12.209424083769633,
+ "grad_norm": 0.06391509622335434,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.465702056884766,
+ "step": 879
+ },
+ {
+ "epoch": 12.223385689354275,
+ "grad_norm": 0.059490568935871124,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.5351881980896,
+ "step": 880
+ },
+ {
+ "epoch": 12.237347294938917,
+ "grad_norm": 0.057501111179590225,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.510270595550537,
+ "step": 881
+ },
+ {
+ "epoch": 12.25130890052356,
+ "grad_norm": 0.056947533041238785,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.509241104125977,
+ "step": 882
+ },
+ {
+ "epoch": 12.265270506108202,
+ "grad_norm": 0.050393324345350266,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.50428581237793,
+ "step": 883
+ },
+ {
+ "epoch": 12.279232111692846,
+ "grad_norm": 0.057608287781476974,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.474919319152832,
+ "step": 884
+ },
+ {
+ "epoch": 12.293193717277488,
+ "grad_norm": 0.054470717906951904,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.521528244018555,
+ "step": 885
+ },
+ {
+ "epoch": 12.30715532286213,
+ "grad_norm": 0.054405540227890015,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.510211944580078,
+ "step": 886
+ },
+ {
+ "epoch": 12.321116928446772,
+ "grad_norm": 0.051328495144844055,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.480720520019531,
+ "step": 887
+ },
+ {
+ "epoch": 12.335078534031414,
+ "grad_norm": 0.045930180698633194,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.503024101257324,
+ "step": 888
+ },
+ {
+ "epoch": 12.349040139616056,
+ "grad_norm": 0.04298274591565132,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.49046516418457,
+ "step": 889
+ },
+ {
+ "epoch": 12.363001745200698,
+ "grad_norm": 0.0406254380941391,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.544064521789551,
+ "step": 890
+ },
+ {
+ "epoch": 12.37696335078534,
+ "grad_norm": 0.03557496517896652,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.496966361999512,
+ "step": 891
+ },
+ {
+ "epoch": 12.390924956369982,
+ "grad_norm": 0.0364493802189827,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.5047502517700195,
+ "step": 892
+ },
+ {
+ "epoch": 12.404886561954624,
+ "grad_norm": 0.034521300345659256,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.485365867614746,
+ "step": 893
+ },
+ {
+ "epoch": 12.418848167539267,
+ "grad_norm": 0.03305799141526222,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.484884738922119,
+ "step": 894
+ },
+ {
+ "epoch": 12.432809773123909,
+ "grad_norm": 0.033053625375032425,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 4.478784561157227,
+ "step": 895
+ },
+ {
+ "epoch": 12.44677137870855,
+ "grad_norm": 0.03426017984747887,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.482428073883057,
+ "step": 896
+ },
+ {
+ "epoch": 12.460732984293193,
+ "grad_norm": 0.03638783469796181,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 4.433219909667969,
+ "step": 897
+ },
+ {
+ "epoch": 12.474694589877837,
+ "grad_norm": 0.03597250208258629,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.522584915161133,
+ "step": 898
+ },
+ {
+ "epoch": 12.488656195462479,
+ "grad_norm": 0.03086097538471222,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.500803470611572,
+ "step": 899
+ },
+ {
+ "epoch": 12.502617801047121,
+ "grad_norm": 0.029642771929502487,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 4.444099426269531,
+ "step": 900
+ },
+ {
+ "epoch": 12.516579406631763,
+ "grad_norm": 0.03190716728568077,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.493034362792969,
+ "step": 901
+ },
+ {
+ "epoch": 12.530541012216405,
+ "grad_norm": 0.0312873050570488,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.487887382507324,
+ "step": 902
+ },
+ {
+ "epoch": 12.544502617801047,
+ "grad_norm": 0.03201943263411522,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 4.479177474975586,
+ "step": 903
+ },
+ {
+ "epoch": 12.55846422338569,
+ "grad_norm": 0.033088769763708115,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.5020246505737305,
+ "step": 904
+ },
+ {
+ "epoch": 12.572425828970331,
+ "grad_norm": 0.02975946106016636,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 4.455426216125488,
+ "step": 905
+ },
+ {
+ "epoch": 12.586387434554974,
+ "grad_norm": 0.03213953226804733,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.501615047454834,
+ "step": 906
+ },
+ {
+ "epoch": 12.600349040139616,
+ "grad_norm": 0.03262908384203911,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.485121250152588,
+ "step": 907
+ },
+ {
+ "epoch": 12.614310645724258,
+ "grad_norm": 0.031102091073989868,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.488893508911133,
+ "step": 908
+ },
+ {
+ "epoch": 12.6282722513089,
+ "grad_norm": 0.02945449948310852,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.490121841430664,
+ "step": 909
+ },
+ {
+ "epoch": 12.642233856893542,
+ "grad_norm": 0.033680472522974014,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.51833438873291,
+ "step": 910
+ },
+ {
+ "epoch": 12.656195462478184,
+ "grad_norm": 0.03277525305747986,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 4.480165481567383,
+ "step": 911
+ },
+ {
+ "epoch": 12.670157068062828,
+ "grad_norm": 0.03303041681647301,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.501155376434326,
+ "step": 912
+ },
+ {
+ "epoch": 12.68411867364747,
+ "grad_norm": 0.034927863627672195,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.513175010681152,
+ "step": 913
+ },
+ {
+ "epoch": 12.698080279232112,
+ "grad_norm": 0.03098093718290329,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.503798484802246,
+ "step": 914
+ },
+ {
+ "epoch": 12.712041884816754,
+ "grad_norm": 0.029513906687498093,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 4.49006986618042,
+ "step": 915
+ },
+ {
+ "epoch": 12.726003490401396,
+ "grad_norm": 0.029630785807967186,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.522514343261719,
+ "step": 916
+ },
+ {
+ "epoch": 12.739965095986038,
+ "grad_norm": 0.030888330191373825,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.4895734786987305,
+ "step": 917
+ },
+ {
+ "epoch": 12.75392670157068,
+ "grad_norm": 0.032392196357250214,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 4.471953868865967,
+ "step": 918
+ },
+ {
+ "epoch": 12.767888307155323,
+ "grad_norm": 0.03215661272406578,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.48297643661499,
+ "step": 919
+ },
+ {
+ "epoch": 12.781849912739965,
+ "grad_norm": 0.03346327692270279,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.518160820007324,
+ "step": 920
+ },
+ {
+ "epoch": 12.795811518324607,
+ "grad_norm": 0.031311362981796265,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 4.484991550445557,
+ "step": 921
+ },
+ {
+ "epoch": 12.809773123909249,
+ "grad_norm": 0.03056488186120987,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.525470733642578,
+ "step": 922
+ },
+ {
+ "epoch": 12.823734729493891,
+ "grad_norm": 0.03043230250477791,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.510650634765625,
+ "step": 923
+ },
+ {
+ "epoch": 12.837696335078533,
+ "grad_norm": 0.02916652522981167,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.496840953826904,
+ "step": 924
+ },
+ {
+ "epoch": 12.851657940663177,
+ "grad_norm": 0.030673664063215256,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.494868278503418,
+ "step": 925
+ },
+ {
+ "epoch": 12.86561954624782,
+ "grad_norm": 0.030033906921744347,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 4.479182720184326,
+ "step": 926
+ },
+ {
+ "epoch": 12.879581151832461,
+ "grad_norm": 0.028828797861933708,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 4.471983909606934,
+ "step": 927
+ },
+ {
+ "epoch": 12.893542757417103,
+ "grad_norm": 0.028316056355834007,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.5049147605896,
+ "step": 928
+ },
+ {
+ "epoch": 12.907504363001745,
+ "grad_norm": 0.03158092871308327,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 4.449270248413086,
+ "step": 929
+ },
+ {
+ "epoch": 12.921465968586388,
+ "grad_norm": 0.03041706420481205,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.508289813995361,
+ "step": 930
+ },
+ {
+ "epoch": 12.93542757417103,
+ "grad_norm": 0.02967553399503231,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 4.465579032897949,
+ "step": 931
+ },
+ {
+ "epoch": 12.949389179755672,
+ "grad_norm": 0.030721619725227356,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 4.471003532409668,
+ "step": 932
+ },
+ {
+ "epoch": 12.963350785340314,
+ "grad_norm": 0.03066832385957241,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 4.488987445831299,
+ "step": 933
+ },
+ {
+ "epoch": 12.977312390924956,
+ "grad_norm": 0.032297998666763306,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 4.491018295288086,
+ "step": 934
+ },
+ {
+ "epoch": 12.991273996509598,
+ "grad_norm": 0.03221398591995239,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 4.471237659454346,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.021120507270097733,
+ "learning_rate": 0.000520413954218197,
+ "loss": 2.8080053329467773,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.5888046026229858,
+ "eval_runtime": 59.1638,
+ "eval_samples_per_second": 41.275,
+ "eval_steps_per_second": 0.659,
+ "step": 936
+ },
+ {
+ "epoch": 13.013961605584642,
+ "grad_norm": 0.03799329698085785,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 4.401787757873535,
+ "step": 937
+ },
+ {
+ "epoch": 13.027923211169284,
+ "grad_norm": 0.05785121023654938,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 4.408820629119873,
+ "step": 938
+ },
+ {
+ "epoch": 13.041884816753926,
+ "grad_norm": 0.05964238941669464,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 4.427263259887695,
+ "step": 939
+ },
+ {
+ "epoch": 13.055846422338568,
+ "grad_norm": 0.05466582626104355,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 4.403415679931641,
+ "step": 940
+ },
+ {
+ "epoch": 13.06980802792321,
+ "grad_norm": 0.05613672360777855,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 4.413750648498535,
+ "step": 941
+ },
+ {
+ "epoch": 13.083769633507853,
+ "grad_norm": 0.05396881327033043,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 4.41916561126709,
+ "step": 942
+ },
+ {
+ "epoch": 13.097731239092496,
+ "grad_norm": 0.05689029023051262,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 4.395590782165527,
+ "step": 943
+ },
+ {
+ "epoch": 13.111692844677139,
+ "grad_norm": 0.056090328842401505,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 4.421253204345703,
+ "step": 944
+ },
+ {
+ "epoch": 13.12565445026178,
+ "grad_norm": 0.0473211295902729,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 4.385158061981201,
+ "step": 945
+ },
+ {
+ "epoch": 13.139616055846423,
+ "grad_norm": 0.04464765638113022,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 4.431746959686279,
+ "step": 946
+ },
+ {
+ "epoch": 13.153577661431065,
+ "grad_norm": 0.046421997249126434,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 4.428816795349121,
+ "step": 947
+ },
+ {
+ "epoch": 13.167539267015707,
+ "grad_norm": 0.05039866268634796,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 4.452417850494385,
+ "step": 948
+ },
+ {
+ "epoch": 13.181500872600349,
+ "grad_norm": 0.05491664633154869,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 4.447483062744141,
+ "step": 949
+ },
+ {
+ "epoch": 13.195462478184991,
+ "grad_norm": 0.05866196006536484,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 4.434992790222168,
+ "step": 950
+ },
+ {
+ "epoch": 13.209424083769633,
+ "grad_norm": 0.05532094091176987,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 4.450716018676758,
+ "step": 951
+ },
+ {
+ "epoch": 13.223385689354275,
+ "grad_norm": 0.051083534955978394,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 4.402632713317871,
+ "step": 952
+ },
+ {
+ "epoch": 13.237347294938917,
+ "grad_norm": 0.03815016150474548,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 4.433640956878662,
+ "step": 953
+ },
+ {
+ "epoch": 13.25130890052356,
+ "grad_norm": 0.04203395918011665,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 4.415888786315918,
+ "step": 954
+ },
+ {
+ "epoch": 13.265270506108202,
+ "grad_norm": 0.043729111552238464,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 4.423260688781738,
+ "step": 955
+ },
+ {
+ "epoch": 13.279232111692846,
+ "grad_norm": 0.039426498115062714,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 4.428949356079102,
+ "step": 956
+ },
+ {
+ "epoch": 13.293193717277488,
+ "grad_norm": 0.03837858885526657,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 4.434070587158203,
+ "step": 957
+ },
+ {
+ "epoch": 13.30715532286213,
+ "grad_norm": 0.038315922021865845,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.471246719360352,
+ "step": 958
+ },
+ {
+ "epoch": 13.321116928446772,
+ "grad_norm": 0.03963236138224602,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 4.4434990882873535,
+ "step": 959
+ },
+ {
+ "epoch": 13.335078534031414,
+ "grad_norm": 0.04027022793889046,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 4.436992645263672,
+ "step": 960
+ },
+ {
+ "epoch": 13.349040139616056,
+ "grad_norm": 0.0409243181347847,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 4.426936149597168,
+ "step": 961
+ },
+ {
+ "epoch": 13.363001745200698,
+ "grad_norm": 0.03404631093144417,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 4.43462610244751,
+ "step": 962
+ },
+ {
+ "epoch": 13.37696335078534,
+ "grad_norm": 0.034393735229969025,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 4.437409400939941,
+ "step": 963
+ },
+ {
+ "epoch": 13.390924956369982,
+ "grad_norm": 0.03635180741548538,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 4.451286315917969,
+ "step": 964
+ },
+ {
+ "epoch": 13.404886561954624,
+ "grad_norm": 0.03349733352661133,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 4.419209003448486,
+ "step": 965
+ },
+ {
+ "epoch": 13.418848167539267,
+ "grad_norm": 0.03238677605986595,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 4.456454277038574,
+ "step": 966
+ },
+ {
+ "epoch": 13.432809773123909,
+ "grad_norm": 0.034473925828933716,
+ "learning_rate": 0.000514700389506765,
+ "loss": 4.3988728523254395,
+ "step": 967
+ },
+ {
+ "epoch": 13.44677137870855,
+ "grad_norm": 0.03455805033445358,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 4.434884071350098,
+ "step": 968
+ },
+ {
+ "epoch": 13.460732984293193,
+ "grad_norm": 0.037486057728528976,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 4.466741561889648,
+ "step": 969
+ },
+ {
+ "epoch": 13.474694589877837,
+ "grad_norm": 0.03617125004529953,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 4.429305076599121,
+ "step": 970
+ },
+ {
+ "epoch": 13.488656195462479,
+ "grad_norm": 0.033646389842033386,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 4.407780170440674,
+ "step": 971
+ },
+ {
+ "epoch": 13.502617801047121,
+ "grad_norm": 0.03401118144392967,
+ "learning_rate": 0.000513763379483416,
+ "loss": 4.42977237701416,
+ "step": 972
+ },
+ {
+ "epoch": 13.516579406631763,
+ "grad_norm": 0.036534134298563004,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 4.451094627380371,
+ "step": 973
+ },
+ {
+ "epoch": 13.530541012216405,
+ "grad_norm": 0.0384049266576767,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 4.414777755737305,
+ "step": 974
+ },
+ {
+ "epoch": 13.544502617801047,
+ "grad_norm": 0.03460627421736717,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 4.420644760131836,
+ "step": 975
+ },
+ {
+ "epoch": 13.55846422338569,
+ "grad_norm": 0.033746346831321716,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 4.461175918579102,
+ "step": 976
+ },
+ {
+ "epoch": 13.572425828970331,
+ "grad_norm": 0.037153810262680054,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 4.444699287414551,
+ "step": 977
+ },
+ {
+ "epoch": 13.586387434554974,
+ "grad_norm": 0.03283539041876793,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 4.431022644042969,
+ "step": 978
+ },
+ {
+ "epoch": 13.600349040139616,
+ "grad_norm": 0.03248969465494156,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 4.447571277618408,
+ "step": 979
+ },
+ {
+ "epoch": 13.614310645724258,
+ "grad_norm": 0.03327852487564087,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 4.42236328125,
+ "step": 980
+ },
+ {
+ "epoch": 13.6282722513089,
+ "grad_norm": 0.031114818528294563,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 4.418244361877441,
+ "step": 981
+ },
+ {
+ "epoch": 13.642233856893542,
+ "grad_norm": 0.03231207653880119,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 4.455004692077637,
+ "step": 982
+ },
+ {
+ "epoch": 13.656195462478184,
+ "grad_norm": 0.030702251940965652,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 4.483610153198242,
+ "step": 983
+ },
+ {
+ "epoch": 13.670157068062828,
+ "grad_norm": 0.031027289107441902,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 4.4470109939575195,
+ "step": 984
+ },
+ {
+ "epoch": 13.68411867364747,
+ "grad_norm": 0.03129073604941368,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 4.452799320220947,
+ "step": 985
+ },
+ {
+ "epoch": 13.698080279232112,
+ "grad_norm": 0.03634301945567131,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 4.4661664962768555,
+ "step": 986
+ },
+ {
+ "epoch": 13.712041884816754,
+ "grad_norm": 0.03718913346529007,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 4.412006855010986,
+ "step": 987
+ },
+ {
+ "epoch": 13.726003490401396,
+ "grad_norm": 0.034018341451883316,
+ "learning_rate": 0.000510736513223685,
+ "loss": 4.436755180358887,
+ "step": 988
+ },
+ {
+ "epoch": 13.739965095986038,
+ "grad_norm": 0.032502319663763046,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 4.452866554260254,
+ "step": 989
+ },
+ {
+ "epoch": 13.75392670157068,
+ "grad_norm": 0.032902952283620834,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 4.449224472045898,
+ "step": 990
+ },
+ {
+ "epoch": 13.767888307155323,
+ "grad_norm": 0.0340626984834671,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 4.445425510406494,
+ "step": 991
+ },
+ {
+ "epoch": 13.781849912739965,
+ "grad_norm": 0.033442508429288864,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 4.432218551635742,
+ "step": 992
+ },
+ {
+ "epoch": 13.795811518324607,
+ "grad_norm": 0.03489428386092186,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 4.4205827713012695,
+ "step": 993
+ },
+ {
+ "epoch": 13.809773123909249,
+ "grad_norm": 0.031868621706962585,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 4.414196968078613,
+ "step": 994
+ },
+ {
+ "epoch": 13.823734729493891,
+ "grad_norm": 0.032986633479595184,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 4.46116304397583,
+ "step": 995
+ },
+ {
+ "epoch": 13.837696335078533,
+ "grad_norm": 0.03348355367779732,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 4.424936771392822,
+ "step": 996
+ },
+ {
+ "epoch": 13.851657940663177,
+ "grad_norm": 0.03174544870853424,
+ "learning_rate": 0.000509015031826403,
+ "loss": 4.434438705444336,
+ "step": 997
+ },
+ {
+ "epoch": 13.86561954624782,
+ "grad_norm": 0.03293656185269356,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 4.440502166748047,
+ "step": 998
+ },
+ {
+ "epoch": 13.879581151832461,
+ "grad_norm": 0.035124607384204865,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 4.440943717956543,
+ "step": 999
+ },
+ {
+ "epoch": 13.893542757417103,
+ "grad_norm": 0.03438971936702728,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 4.431408882141113,
+ "step": 1000
+ },
+ {
+ "epoch": 13.907504363001745,
+ "grad_norm": 0.03537093102931976,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 4.43867301940918,
+ "step": 1001
+ },
+ {
+ "epoch": 13.921465968586388,
+ "grad_norm": 0.03267325833439827,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 4.462226867675781,
+ "step": 1002
+ },
+ {
+ "epoch": 13.93542757417103,
+ "grad_norm": 0.030995232984423637,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 4.433740615844727,
+ "step": 1003
+ },
+ {
+ "epoch": 13.949389179755672,
+ "grad_norm": 0.03384929150342941,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 4.508179664611816,
+ "step": 1004
+ },
+ {
+ "epoch": 13.963350785340314,
+ "grad_norm": 0.032025814056396484,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 4.407378196716309,
+ "step": 1005
+ },
+ {
+ "epoch": 13.977312390924956,
+ "grad_norm": 0.03288578242063522,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 4.455531120300293,
+ "step": 1006
+ },
+ {
+ "epoch": 13.991273996509598,
+ "grad_norm": 0.032061848789453506,
+ "learning_rate": 0.000507086511366679,
+ "loss": 4.452212333679199,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.022320276126265526,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 2.7567386627197266,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.5878658294677734,
+ "eval_runtime": 59.5054,
+ "eval_samples_per_second": 41.038,
+ "eval_steps_per_second": 0.655,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013961605584642,
+ "grad_norm": 0.039114587008953094,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 4.347412586212158,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027923211169284,
+ "grad_norm": 0.05164389684796333,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 4.391279220581055,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041884816753926,
+ "grad_norm": 0.055882278829813004,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 4.355602264404297,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055846422338568,
+ "grad_norm": 0.05625041201710701,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 4.3934855461120605,
+ "step": 1012
+ },
+ {
+ "epoch": 14.06980802792321,
+ "grad_norm": 0.05861108750104904,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 4.364226341247559,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083769633507853,
+ "grad_norm": 0.052431970834732056,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 4.363334655761719,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097731239092496,
+ "grad_norm": 0.06147193908691406,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 4.384820938110352,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111692844677139,
+ "grad_norm": 0.06701470166444778,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 4.373311996459961,
+ "step": 1016
+ },
+ {
+ "epoch": 14.12565445026178,
+ "grad_norm": 0.06714873015880585,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 4.32745361328125,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139616055846423,
+ "grad_norm": 0.06684907525777817,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 4.373660087585449,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153577661431065,
+ "grad_norm": 0.05738082900643349,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 4.362325668334961,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167539267015707,
+ "grad_norm": 0.05306725949048996,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 4.351685047149658,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181500872600349,
+ "grad_norm": 0.05831029266119003,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 4.374813079833984,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195462478184991,
+ "grad_norm": 0.05547972396016121,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 4.361842632293701,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209424083769633,
+ "grad_norm": 0.048057056963443756,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 4.383205413818359,
+ "step": 1023
+ },
+ {
+ "epoch": 14.223385689354275,
+ "grad_norm": 0.04682513326406479,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 4.365846157073975,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237347294938917,
+ "grad_norm": 0.05001377314329147,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 4.362151145935059,
+ "step": 1025
+ },
+ {
+ "epoch": 14.25130890052356,
+ "grad_norm": 0.05140263959765434,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 4.376064300537109,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265270506108202,
+ "grad_norm": 0.04640384018421173,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 4.363801002502441,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279232111692846,
+ "grad_norm": 0.04175020009279251,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 4.3477983474731445,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293193717277488,
+ "grad_norm": 0.04134776070713997,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 4.337469100952148,
+ "step": 1029
+ },
+ {
+ "epoch": 14.30715532286213,
+ "grad_norm": 0.042178574949502945,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 4.351490020751953,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321116928446772,
+ "grad_norm": 0.03989115729928017,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 4.40725040435791,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335078534031414,
+ "grad_norm": 0.04086652770638466,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 4.341119766235352,
+ "step": 1032
+ },
+ {
+ "epoch": 14.349040139616056,
+ "grad_norm": 0.03931201249361038,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 4.3786468505859375,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363001745200698,
+ "grad_norm": 0.03855905309319496,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 4.3701605796813965,
+ "step": 1034
+ },
+ {
+ "epoch": 14.37696335078534,
+ "grad_norm": 0.03803952783346176,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 4.401841163635254,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390924956369982,
+ "grad_norm": 0.036908265203237534,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 4.410647392272949,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404886561954624,
+ "grad_norm": 0.0398186556994915,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 4.380431175231934,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418848167539267,
+ "grad_norm": 0.03899035602807999,
+ "learning_rate": 0.000501004302813408,
+ "loss": 4.374293804168701,
+ "step": 1038
+ },
+ {
+ "epoch": 14.432809773123909,
+ "grad_norm": 0.03946344554424286,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 4.396918773651123,
+ "step": 1039
+ },
+ {
+ "epoch": 14.44677137870855,
+ "grad_norm": 0.03738604858517647,
+ "learning_rate": 0.000500606587074199,
+ "loss": 4.404016017913818,
+ "step": 1040
+ },
+ {
+ "epoch": 14.460732984293193,
+ "grad_norm": 0.03579043596982956,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 4.396914958953857,
+ "step": 1041
+ },
+ {
+ "epoch": 14.474694589877837,
+ "grad_norm": 0.037139154970645905,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 4.407739639282227,
+ "step": 1042
+ },
+ {
+ "epoch": 14.488656195462479,
+ "grad_norm": 0.03581099584698677,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 4.411419868469238,
+ "step": 1043
+ },
+ {
+ "epoch": 14.502617801047121,
+ "grad_norm": 0.03548453748226166,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 4.332690238952637,
+ "step": 1044
+ },
+ {
+ "epoch": 14.516579406631763,
+ "grad_norm": 0.037561237812042236,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 4.381314277648926,
+ "step": 1045
+ },
+ {
+ "epoch": 14.530541012216405,
+ "grad_norm": 0.035827480256557465,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 4.371536731719971,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544502617801047,
+ "grad_norm": 0.03538314625620842,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 4.395467758178711,
+ "step": 1047
+ },
+ {
+ "epoch": 14.55846422338569,
+ "grad_norm": 0.036672890186309814,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 4.420590400695801,
+ "step": 1048
+ },
+ {
+ "epoch": 14.572425828970331,
+ "grad_norm": 0.038981080055236816,
+ "learning_rate": 0.000498809004003543,
+ "loss": 4.395296096801758,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586387434554974,
+ "grad_norm": 0.038234200328588486,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 4.415799140930176,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600349040139616,
+ "grad_norm": 0.038830097764730453,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 4.413394927978516,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614310645724258,
+ "grad_norm": 0.03990919142961502,
+ "learning_rate": 0.000498206964168724,
+ "loss": 4.39515495300293,
+ "step": 1052
+ },
+ {
+ "epoch": 14.6282722513089,
+ "grad_norm": 0.036424681544303894,
+ "learning_rate": 0.000498005969458628,
+ "loss": 4.39730978012085,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642233856893542,
+ "grad_norm": 0.03488154709339142,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 4.403696060180664,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656195462478184,
+ "grad_norm": 0.0363851860165596,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 4.368661403656006,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670157068062828,
+ "grad_norm": 0.035461440682411194,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 4.396518707275391,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68411867364747,
+ "grad_norm": 0.03647598251700401,
+ "learning_rate": 0.000497200421111647,
+ "loss": 4.440667152404785,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698080279232112,
+ "grad_norm": 0.03667188808321953,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 4.3781352043151855,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712041884816754,
+ "grad_norm": 0.034516457468271255,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 4.414007186889648,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726003490401396,
+ "grad_norm": 0.03536032512784004,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 4.42198371887207,
+ "step": 1060
+ },
+ {
+ "epoch": 14.739965095986038,
+ "grad_norm": 0.037454646080732346,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 4.409174919128418,
+ "step": 1061
+ },
+ {
+ "epoch": 14.75392670157068,
+ "grad_norm": 0.03632275015115738,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 4.3547682762146,
+ "step": 1062
+ },
+ {
+ "epoch": 14.767888307155323,
+ "grad_norm": 0.038922134786844254,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 4.39993143081665,
+ "step": 1063
+ },
+ {
+ "epoch": 14.781849912739965,
+ "grad_norm": 0.03562692925333977,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 4.409564971923828,
+ "step": 1064
+ },
+ {
+ "epoch": 14.795811518324607,
+ "grad_norm": 0.033203717321157455,
+ "learning_rate": 0.000495581824580724,
+ "loss": 4.421782493591309,
+ "step": 1065
+ },
+ {
+ "epoch": 14.809773123909249,
+ "grad_norm": 0.03830202296376228,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 4.430092811584473,
+ "step": 1066
+ },
+ {
+ "epoch": 14.823734729493891,
+ "grad_norm": 0.03835884481668472,
+ "learning_rate": 0.000495175620586125,
+ "loss": 4.379505634307861,
+ "step": 1067
+ },
+ {
+ "epoch": 14.837696335078533,
+ "grad_norm": 0.037440430372953415,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 4.403937339782715,
+ "step": 1068
+ },
+ {
+ "epoch": 14.851657940663177,
+ "grad_norm": 0.03732272610068321,
+ "learning_rate": 0.000494768797261945,
+ "loss": 4.440683364868164,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86561954624782,
+ "grad_norm": 0.03325957432389259,
+ "learning_rate": 0.000494565153760897,
+ "loss": 4.3949294090271,
+ "step": 1070
+ },
+ {
+ "epoch": 14.879581151832461,
+ "grad_norm": 0.035377148538827896,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 4.400777816772461,
+ "step": 1071
+ },
+ {
+ "epoch": 14.893542757417103,
+ "grad_norm": 0.03725374862551689,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 4.422494411468506,
+ "step": 1072
+ },
+ {
+ "epoch": 14.907504363001745,
+ "grad_norm": 0.03601110726594925,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 4.4008941650390625,
+ "step": 1073
+ },
+ {
+ "epoch": 14.921465968586388,
+ "grad_norm": 0.03193027526140213,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 4.400361061096191,
+ "step": 1074
+ },
+ {
+ "epoch": 14.93542757417103,
+ "grad_norm": 0.034382354468107224,
+ "learning_rate": 0.00049354462443268,
+ "loss": 4.408535003662109,
+ "step": 1075
+ },
+ {
+ "epoch": 14.949389179755672,
+ "grad_norm": 0.03747732564806938,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 4.412189483642578,
+ "step": 1076
+ },
+ {
+ "epoch": 14.963350785340314,
+ "grad_norm": 0.03641800582408905,
+ "learning_rate": 0.000493135336920878,
+ "loss": 4.381453037261963,
+ "step": 1077
+ },
+ {
+ "epoch": 14.977312390924956,
+ "grad_norm": 0.0315779373049736,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 4.4169769287109375,
+ "step": 1078
+ },
+ {
+ "epoch": 14.991273996509598,
+ "grad_norm": 0.03583509847521782,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 4.385746955871582,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.02574268914759159,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 2.7456769943237305,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.588273823261261,
+ "eval_runtime": 59.9356,
+ "eval_samples_per_second": 40.744,
+ "eval_steps_per_second": 0.651,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013961605584642,
+ "grad_norm": 0.037516895681619644,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 4.33476448059082,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027923211169284,
+ "grad_norm": 0.048450272530317307,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 4.296454429626465,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041884816753926,
+ "grad_norm": 0.05368480458855629,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 4.314815998077393,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055846422338568,
+ "grad_norm": 0.05520009621977806,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 4.330798625946045,
+ "step": 1084
+ },
+ {
+ "epoch": 15.06980802792321,
+ "grad_norm": 0.05928800627589226,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 4.317663192749023,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083769633507853,
+ "grad_norm": 0.06315027177333832,
+ "learning_rate": 0.000491285979372924,
+ "loss": 4.329221725463867,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097731239092496,
+ "grad_norm": 0.06342213600873947,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 4.32546329498291,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111692844677139,
+ "grad_norm": 0.0667002946138382,
+ "learning_rate": 0.000490873338088198,
+ "loss": 4.309828758239746,
+ "step": 1088
+ },
+ {
+ "epoch": 15.12565445026178,
+ "grad_norm": 0.05264135077595711,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 4.307073593139648,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139616055846423,
+ "grad_norm": 0.051895879209041595,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 4.333526611328125,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153577661431065,
+ "grad_norm": 0.05508396774530411,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 4.330171585083008,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167539267015707,
+ "grad_norm": 0.057272691279649734,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 4.329268455505371,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181500872600349,
+ "grad_norm": 0.058956146240234375,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 4.341983795166016,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195462478184991,
+ "grad_norm": 0.0557059645652771,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 4.348274230957031,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209424083769633,
+ "grad_norm": 0.05661974847316742,
+ "learning_rate": 0.000489424334303338,
+ "loss": 4.347631454467773,
+ "step": 1095
+ },
+ {
+ "epoch": 15.223385689354275,
+ "grad_norm": 0.05338657647371292,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 4.334639072418213,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237347294938917,
+ "grad_norm": 0.05091148987412453,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 4.355465412139893,
+ "step": 1097
+ },
+ {
+ "epoch": 15.25130890052356,
+ "grad_norm": 0.04996965080499649,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 4.346457004547119,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265270506108202,
+ "grad_norm": 0.05573750659823418,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 4.346520900726318,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279232111692846,
+ "grad_norm": 0.05776917189359665,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 4.359007835388184,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293193717277488,
+ "grad_norm": 0.050820041447877884,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 4.331710338592529,
+ "step": 1101
+ },
+ {
+ "epoch": 15.30715532286213,
+ "grad_norm": 0.04441820830106735,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 4.3126420974731445,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321116928446772,
+ "grad_norm": 0.04582618921995163,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 4.312925338745117,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335078534031414,
+ "grad_norm": 0.043592359870672226,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 4.345674514770508,
+ "step": 1104
+ },
+ {
+ "epoch": 15.349040139616056,
+ "grad_norm": 0.04069037735462189,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 4.334713935852051,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363001745200698,
+ "grad_norm": 0.04083041846752167,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 4.358171463012695,
+ "step": 1106
+ },
+ {
+ "epoch": 15.37696335078534,
+ "grad_norm": 0.04119594022631645,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 4.317891597747803,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390924956369982,
+ "grad_norm": 0.042060643434524536,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 4.355810165405273,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404886561954624,
+ "grad_norm": 0.04721749201416969,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 4.353546619415283,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418848167539267,
+ "grad_norm": 0.04657730832695961,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 4.341683387756348,
+ "step": 1110
+ },
+ {
+ "epoch": 15.432809773123909,
+ "grad_norm": 0.04096715524792671,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 4.357192039489746,
+ "step": 1111
+ },
+ {
+ "epoch": 15.44677137870855,
+ "grad_norm": 0.03904155269265175,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 4.354775428771973,
+ "step": 1112
+ },
+ {
+ "epoch": 15.460732984293193,
+ "grad_norm": 0.03828521445393562,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 4.356499671936035,
+ "step": 1113
+ },
+ {
+ "epoch": 15.474694589877837,
+ "grad_norm": 0.03688685595989227,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 4.339485168457031,
+ "step": 1114
+ },
+ {
+ "epoch": 15.488656195462479,
+ "grad_norm": 0.03945618495345116,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 4.363500118255615,
+ "step": 1115
+ },
+ {
+ "epoch": 15.502617801047121,
+ "grad_norm": 0.03821462020277977,
+ "learning_rate": 0.000485033394934934,
+ "loss": 4.353529930114746,
+ "step": 1116
+ },
+ {
+ "epoch": 15.516579406631763,
+ "grad_norm": 0.036265499889850616,
+ "learning_rate": 0.000484822676912974,
+ "loss": 4.337663173675537,
+ "step": 1117
+ },
+ {
+ "epoch": 15.530541012216405,
+ "grad_norm": 0.03803930804133415,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 4.34660530090332,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544502617801047,
+ "grad_norm": 0.038425102829933167,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 4.373246192932129,
+ "step": 1119
+ },
+ {
+ "epoch": 15.55846422338569,
+ "grad_norm": 0.036852262914180756,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 4.331226825714111,
+ "step": 1120
+ },
+ {
+ "epoch": 15.572425828970331,
+ "grad_norm": 0.03700809180736542,
+ "learning_rate": 0.000483978341612154,
+ "loss": 4.375818729400635,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586387434554974,
+ "grad_norm": 0.03628664091229439,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 4.32705020904541,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600349040139616,
+ "grad_norm": 0.036831192672252655,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 4.337127208709717,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614310645724258,
+ "grad_norm": 0.03736541047692299,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 4.36502742767334,
+ "step": 1124
+ },
+ {
+ "epoch": 15.6282722513089,
+ "grad_norm": 0.03658579662442207,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 4.359844207763672,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642233856893542,
+ "grad_norm": 0.03688633441925049,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 4.3355793952941895,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656195462478184,
+ "grad_norm": 0.03477008268237114,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 4.333405494689941,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670157068062828,
+ "grad_norm": 0.03655015677213669,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 4.338570594787598,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68411867364747,
+ "grad_norm": 0.03664765506982803,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 4.352496147155762,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698080279232112,
+ "grad_norm": 0.037019114941358566,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 4.335735321044922,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712041884816754,
+ "grad_norm": 0.03949085995554924,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 4.342585563659668,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726003490401396,
+ "grad_norm": 0.04084350913763046,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 4.329622268676758,
+ "step": 1132
+ },
+ {
+ "epoch": 15.739965095986038,
+ "grad_norm": 0.038439832627773285,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 4.360318660736084,
+ "step": 1133
+ },
+ {
+ "epoch": 15.75392670157068,
+ "grad_norm": 0.03932987526059151,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 4.343446254730225,
+ "step": 1134
+ },
+ {
+ "epoch": 15.767888307155323,
+ "grad_norm": 0.041595373302698135,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 4.351495742797852,
+ "step": 1135
+ },
+ {
+ "epoch": 15.781849912739965,
+ "grad_norm": 0.03917490318417549,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 4.380537033081055,
+ "step": 1136
+ },
+ {
+ "epoch": 15.795811518324607,
+ "grad_norm": 0.038597021251916885,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 4.3527116775512695,
+ "step": 1137
+ },
+ {
+ "epoch": 15.809773123909249,
+ "grad_norm": 0.03651711344718933,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 4.369256496429443,
+ "step": 1138
+ },
+ {
+ "epoch": 15.823734729493891,
+ "grad_norm": 0.035118963569402695,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 4.350616455078125,
+ "step": 1139
+ },
+ {
+ "epoch": 15.837696335078533,
+ "grad_norm": 0.034945692867040634,
+ "learning_rate": 0.000479936075927473,
+ "loss": 4.365446090698242,
+ "step": 1140
+ },
+ {
+ "epoch": 15.851657940663177,
+ "grad_norm": 0.03519820421934128,
+ "learning_rate": 0.000479721889242305,
+ "loss": 4.357390403747559,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86561954624782,
+ "grad_norm": 0.03407098725438118,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 4.350573539733887,
+ "step": 1142
+ },
+ {
+ "epoch": 15.879581151832461,
+ "grad_norm": 0.036511123180389404,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 4.328500270843506,
+ "step": 1143
+ },
+ {
+ "epoch": 15.893542757417103,
+ "grad_norm": 0.039194293320178986,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 4.346471786499023,
+ "step": 1144
+ },
+ {
+ "epoch": 15.907504363001745,
+ "grad_norm": 0.03777911514043808,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 4.358243942260742,
+ "step": 1145
+ },
+ {
+ "epoch": 15.921465968586388,
+ "grad_norm": 0.0354655385017395,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 4.368768692016602,
+ "step": 1146
+ },
+ {
+ "epoch": 15.93542757417103,
+ "grad_norm": 0.03710418567061424,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 4.3482184410095215,
+ "step": 1147
+ },
+ {
+ "epoch": 15.949389179755672,
+ "grad_norm": 0.03545419126749039,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 4.360692501068115,
+ "step": 1148
+ },
+ {
+ "epoch": 15.963350785340314,
+ "grad_norm": 0.035963475704193115,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 4.3592119216918945,
+ "step": 1149
+ },
+ {
+ "epoch": 15.977312390924956,
+ "grad_norm": 0.03425297513604164,
+ "learning_rate": 0.000477787822648007,
+ "loss": 4.329504013061523,
+ "step": 1150
+ },
+ {
+ "epoch": 15.991273996509598,
+ "grad_norm": 0.03453255817294121,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 4.298142433166504,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.025766054168343544,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 2.7026405334472656,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.5888971090316772,
+ "eval_runtime": 58.9721,
+ "eval_samples_per_second": 41.409,
+ "eval_steps_per_second": 0.661,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013961605584644,
+ "grad_norm": 0.04088004305958748,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 4.268601417541504,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027923211169284,
+ "grad_norm": 0.05194641277194023,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 4.267483711242676,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041884816753928,
+ "grad_norm": 0.05488700792193413,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 4.2527570724487305,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05584642233857,
+ "grad_norm": 0.059292279183864594,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 4.283864974975586,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069808027923212,
+ "grad_norm": 0.0567227303981781,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 4.291148662567139,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083769633507853,
+ "grad_norm": 0.05582835525274277,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 4.250524520874023,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097731239092496,
+ "grad_norm": 0.055709999054670334,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 4.259716033935547,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111692844677137,
+ "grad_norm": 0.053324732929468155,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 4.247231483459473,
+ "step": 1160
+ },
+ {
+ "epoch": 16.12565445026178,
+ "grad_norm": 0.05367910861968994,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 4.263905048370361,
+ "step": 1161
+ },
+ {
+ "epoch": 16.13961605584642,
+ "grad_norm": 0.057094886898994446,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 4.266510009765625,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153577661431065,
+ "grad_norm": 0.05921962857246399,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 4.278281211853027,
+ "step": 1163
+ },
+ {
+ "epoch": 16.167539267015705,
+ "grad_norm": 0.060246821492910385,
+ "learning_rate": 0.000474756648870413,
+ "loss": 4.2848052978515625,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18150087260035,
+ "grad_norm": 0.05686050280928612,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 4.286434173583984,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195462478184993,
+ "grad_norm": 0.05207265168428421,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 4.246121406555176,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209424083769633,
+ "grad_norm": 0.0477963425219059,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 4.287411689758301,
+ "step": 1167
+ },
+ {
+ "epoch": 16.223385689354277,
+ "grad_norm": 0.05586806684732437,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 4.2721333503723145,
+ "step": 1168
+ },
+ {
+ "epoch": 16.237347294938917,
+ "grad_norm": 0.05427060276269913,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 4.2751641273498535,
+ "step": 1169
+ },
+ {
+ "epoch": 16.25130890052356,
+ "grad_norm": 0.0518609918653965,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 4.287444114685059,
+ "step": 1170
+ },
+ {
+ "epoch": 16.2652705061082,
+ "grad_norm": 0.053606316447257996,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 4.287125587463379,
+ "step": 1171
+ },
+ {
+ "epoch": 16.279232111692846,
+ "grad_norm": 0.05118813365697861,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 4.291893005371094,
+ "step": 1172
+ },
+ {
+ "epoch": 16.293193717277486,
+ "grad_norm": 0.04774491861462593,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 4.325250148773193,
+ "step": 1173
+ },
+ {
+ "epoch": 16.30715532286213,
+ "grad_norm": 0.04634870961308479,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 4.311764717102051,
+ "step": 1174
+ },
+ {
+ "epoch": 16.32111692844677,
+ "grad_norm": 0.04411927983164787,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 4.277068138122559,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335078534031414,
+ "grad_norm": 0.043819356709718704,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 4.309638977050781,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349040139616054,
+ "grad_norm": 0.0418904684484005,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 4.306234836578369,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363001745200698,
+ "grad_norm": 0.04327298328280449,
+ "learning_rate": 0.000471698352726896,
+ "loss": 4.296874046325684,
+ "step": 1178
+ },
+ {
+ "epoch": 16.376963350785342,
+ "grad_norm": 0.045296184718608856,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 4.288957595825195,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390924956369982,
+ "grad_norm": 0.04447923228144646,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 4.313834190368652,
+ "step": 1180
+ },
+ {
+ "epoch": 16.404886561954626,
+ "grad_norm": 0.04548390209674835,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 4.284562587738037,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418848167539267,
+ "grad_norm": 0.04567817226052284,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 4.295994758605957,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43280977312391,
+ "grad_norm": 0.04277439042925835,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 4.262106895446777,
+ "step": 1183
+ },
+ {
+ "epoch": 16.44677137870855,
+ "grad_norm": 0.043597545474767685,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 4.294536590576172,
+ "step": 1184
+ },
+ {
+ "epoch": 16.460732984293195,
+ "grad_norm": 0.04424033313989639,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 4.297497749328613,
+ "step": 1185
+ },
+ {
+ "epoch": 16.474694589877835,
+ "grad_norm": 0.04354134574532509,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 4.321508884429932,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48865619546248,
+ "grad_norm": 0.04241178184747696,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 4.322726249694824,
+ "step": 1187
+ },
+ {
+ "epoch": 16.50261780104712,
+ "grad_norm": 0.039317622780799866,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 4.311155319213867,
+ "step": 1188
+ },
+ {
+ "epoch": 16.516579406631763,
+ "grad_norm": 0.04126536473631859,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 4.306636810302734,
+ "step": 1189
+ },
+ {
+ "epoch": 16.530541012216403,
+ "grad_norm": 0.040116794407367706,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 4.314671516418457,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544502617801047,
+ "grad_norm": 0.04045999422669411,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 4.301012992858887,
+ "step": 1191
+ },
+ {
+ "epoch": 16.55846422338569,
+ "grad_norm": 0.04013422504067421,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 4.29913330078125,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57242582897033,
+ "grad_norm": 0.03878502920269966,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 4.351600646972656,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586387434554975,
+ "grad_norm": 0.038935888558626175,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 4.27667236328125,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600349040139616,
+ "grad_norm": 0.04055892676115036,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 4.29019021987915,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61431064572426,
+ "grad_norm": 0.03959144279360771,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 4.327725887298584,
+ "step": 1196
+ },
+ {
+ "epoch": 16.6282722513089,
+ "grad_norm": 0.039878539741039276,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 4.297952651977539,
+ "step": 1197
+ },
+ {
+ "epoch": 16.642233856893544,
+ "grad_norm": 0.041209928691387177,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 4.313030242919922,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656195462478184,
+ "grad_norm": 0.04346393793821335,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 4.274901390075684,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670157068062828,
+ "grad_norm": 0.04090269282460213,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 4.330657958984375,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68411867364747,
+ "grad_norm": 0.04004506766796112,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 4.323106288909912,
+ "step": 1201
+ },
+ {
+ "epoch": 16.698080279232112,
+ "grad_norm": 0.041541460901498795,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 4.280947685241699,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712041884816752,
+ "grad_norm": 0.04033653065562248,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 4.341434955596924,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726003490401396,
+ "grad_norm": 0.04104280099272728,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 4.2931671142578125,
+ "step": 1204
+ },
+ {
+ "epoch": 16.739965095986037,
+ "grad_norm": 0.039398156106472015,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 4.316840648651123,
+ "step": 1205
+ },
+ {
+ "epoch": 16.75392670157068,
+ "grad_norm": 0.03980560600757599,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 4.325345039367676,
+ "step": 1206
+ },
+ {
+ "epoch": 16.767888307155324,
+ "grad_norm": 0.040656186640262604,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 4.32675838470459,
+ "step": 1207
+ },
+ {
+ "epoch": 16.781849912739965,
+ "grad_norm": 0.03775986284017563,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 4.314423561096191,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79581151832461,
+ "grad_norm": 0.04043284431099892,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 4.344202041625977,
+ "step": 1209
+ },
+ {
+ "epoch": 16.80977312390925,
+ "grad_norm": 0.04074503108859062,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 4.350123405456543,
+ "step": 1210
+ },
+ {
+ "epoch": 16.823734729493893,
+ "grad_norm": 0.04117420315742493,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 4.350917816162109,
+ "step": 1211
+ },
+ {
+ "epoch": 16.837696335078533,
+ "grad_norm": 0.03799610957503319,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 4.289987564086914,
+ "step": 1212
+ },
+ {
+ "epoch": 16.851657940663177,
+ "grad_norm": 0.04071265831589699,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 4.306744575500488,
+ "step": 1213
+ },
+ {
+ "epoch": 16.865619546247817,
+ "grad_norm": 0.04109715297818184,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 4.318753242492676,
+ "step": 1214
+ },
+ {
+ "epoch": 16.87958115183246,
+ "grad_norm": 0.03910145163536072,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 4.281325340270996,
+ "step": 1215
+ },
+ {
+ "epoch": 16.8935427574171,
+ "grad_norm": 0.03752733767032623,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 4.3231048583984375,
+ "step": 1216
+ },
+ {
+ "epoch": 16.907504363001745,
+ "grad_norm": 0.03752103075385094,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 4.351336479187012,
+ "step": 1217
+ },
+ {
+ "epoch": 16.921465968586386,
+ "grad_norm": 0.03675422817468643,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 4.282256126403809,
+ "step": 1218
+ },
+ {
+ "epoch": 16.93542757417103,
+ "grad_norm": 0.03850258141756058,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 4.322597503662109,
+ "step": 1219
+ },
+ {
+ "epoch": 16.949389179755673,
+ "grad_norm": 0.03692392632365227,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 4.320333957672119,
+ "step": 1220
+ },
+ {
+ "epoch": 16.963350785340314,
+ "grad_norm": 0.035366240888834,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 4.31341552734375,
+ "step": 1221
+ },
+ {
+ "epoch": 16.977312390924958,
+ "grad_norm": 0.037120968103408813,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 4.321141242980957,
+ "step": 1222
+ },
+ {
+ "epoch": 16.991273996509598,
+ "grad_norm": 0.03721754252910614,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 4.305757522583008,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.026698043569922447,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 2.716372489929199,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.5894472599029541,
+ "eval_runtime": 58.927,
+ "eval_samples_per_second": 41.441,
+ "eval_steps_per_second": 0.662,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013961605584644,
+ "grad_norm": 0.044532276690006256,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 4.243244171142578,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027923211169284,
+ "grad_norm": 0.050529688596725464,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 4.221382141113281,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041884816753928,
+ "grad_norm": 0.05690870061516762,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 4.240285873413086,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05584642233857,
+ "grad_norm": 0.054803814738988876,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 4.213047027587891,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069808027923212,
+ "grad_norm": 0.051825862377882004,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 4.226880073547363,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083769633507853,
+ "grad_norm": 0.049802035093307495,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 4.25303840637207,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097731239092496,
+ "grad_norm": 0.05290328338742256,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 4.199422836303711,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111692844677137,
+ "grad_norm": 0.0531814880669117,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 4.201648712158203,
+ "step": 1232
+ },
+ {
+ "epoch": 17.12565445026178,
+ "grad_norm": 0.048972588032484055,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 4.207578182220459,
+ "step": 1233
+ },
+ {
+ "epoch": 17.13961605584642,
+ "grad_norm": 0.051345713436603546,
+ "learning_rate": 0.000459203665939781,
+ "loss": 4.206205368041992,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153577661431065,
+ "grad_norm": 0.059012677520513535,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 4.228360176086426,
+ "step": 1235
+ },
+ {
+ "epoch": 17.167539267015705,
+ "grad_norm": 0.05984492972493172,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 4.234362602233887,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18150087260035,
+ "grad_norm": 0.060886185616254807,
+ "learning_rate": 0.000458522886595554,
+ "loss": 4.253525257110596,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195462478184993,
+ "grad_norm": 0.06476911902427673,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 4.249889373779297,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209424083769633,
+ "grad_norm": 0.059773124754428864,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 4.225249290466309,
+ "step": 1239
+ },
+ {
+ "epoch": 17.223385689354277,
+ "grad_norm": 0.06063077598810196,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 4.25465202331543,
+ "step": 1240
+ },
+ {
+ "epoch": 17.237347294938917,
+ "grad_norm": 0.059922896325588226,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 4.252101898193359,
+ "step": 1241
+ },
+ {
+ "epoch": 17.25130890052356,
+ "grad_norm": 0.05849695950746536,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 4.268218040466309,
+ "step": 1242
+ },
+ {
+ "epoch": 17.2652705061082,
+ "grad_norm": 0.05769657716155052,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 4.2395782470703125,
+ "step": 1243
+ },
+ {
+ "epoch": 17.279232111692846,
+ "grad_norm": 0.058189719915390015,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 4.257606506347656,
+ "step": 1244
+ },
+ {
+ "epoch": 17.293193717277486,
+ "grad_norm": 0.059679049998521805,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 4.278588771820068,
+ "step": 1245
+ },
+ {
+ "epoch": 17.30715532286213,
+ "grad_norm": 0.0579286590218544,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 4.25511360168457,
+ "step": 1246
+ },
+ {
+ "epoch": 17.32111692844677,
+ "grad_norm": 0.05269965901970863,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 4.241326332092285,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335078534031414,
+ "grad_norm": 0.04828358441591263,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 4.243182182312012,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349040139616054,
+ "grad_norm": 0.050172045826911926,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 4.295144081115723,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363001745200698,
+ "grad_norm": 0.047297943383455276,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 4.250682830810547,
+ "step": 1250
+ },
+ {
+ "epoch": 17.376963350785342,
+ "grad_norm": 0.04574039950966835,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 4.269444465637207,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390924956369982,
+ "grad_norm": 0.042310427874326706,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 4.255361557006836,
+ "step": 1252
+ },
+ {
+ "epoch": 17.404886561954626,
+ "grad_norm": 0.04268861189484596,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 4.252292156219482,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418848167539267,
+ "grad_norm": 0.0424286350607872,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 4.24439811706543,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43280977312391,
+ "grad_norm": 0.042049869894981384,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 4.236266136169434,
+ "step": 1255
+ },
+ {
+ "epoch": 17.44677137870855,
+ "grad_norm": 0.04185241088271141,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 4.268826484680176,
+ "step": 1256
+ },
+ {
+ "epoch": 17.460732984293195,
+ "grad_norm": 0.04423609375953674,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 4.240240573883057,
+ "step": 1257
+ },
+ {
+ "epoch": 17.474694589877835,
+ "grad_norm": 0.041972566395998,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 4.269453048706055,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48865619546248,
+ "grad_norm": 0.04493343085050583,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 4.270544052124023,
+ "step": 1259
+ },
+ {
+ "epoch": 17.50261780104712,
+ "grad_norm": 0.04502921551465988,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 4.271880149841309,
+ "step": 1260
+ },
+ {
+ "epoch": 17.516579406631763,
+ "grad_norm": 0.044518254697322845,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 4.249786376953125,
+ "step": 1261
+ },
+ {
+ "epoch": 17.530541012216403,
+ "grad_norm": 0.044482357800006866,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 4.300779819488525,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544502617801047,
+ "grad_norm": 0.0424671471118927,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 4.248597621917725,
+ "step": 1263
+ },
+ {
+ "epoch": 17.55846422338569,
+ "grad_norm": 0.04136480763554573,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 4.276763916015625,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57242582897033,
+ "grad_norm": 0.04094843938946724,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 4.263628005981445,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586387434554975,
+ "grad_norm": 0.04415388032793999,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 4.257946014404297,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600349040139616,
+ "grad_norm": 0.04177384823560715,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 4.289669036865234,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61431064572426,
+ "grad_norm": 0.040863484144210815,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 4.320903778076172,
+ "step": 1268
+ },
+ {
+ "epoch": 17.6282722513089,
+ "grad_norm": 0.04025879129767418,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 4.268717288970947,
+ "step": 1269
+ },
+ {
+ "epoch": 17.642233856893544,
+ "grad_norm": 0.04161325842142105,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 4.275268077850342,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656195462478184,
+ "grad_norm": 0.04260355979204178,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 4.263361930847168,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670157068062828,
+ "grad_norm": 0.043366871774196625,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 4.254581928253174,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68411867364747,
+ "grad_norm": 0.04299113526940346,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 4.268922328948975,
+ "step": 1273
+ },
+ {
+ "epoch": 17.698080279232112,
+ "grad_norm": 0.041865330189466476,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 4.290097236633301,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712041884816752,
+ "grad_norm": 0.04515853524208069,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 4.261376857757568,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726003490401396,
+ "grad_norm": 0.046030085533857346,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 4.290409564971924,
+ "step": 1276
+ },
+ {
+ "epoch": 17.739965095986037,
+ "grad_norm": 0.045034267008304596,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 4.28201961517334,
+ "step": 1277
+ },
+ {
+ "epoch": 17.75392670157068,
+ "grad_norm": 0.04293077066540718,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 4.282614707946777,
+ "step": 1278
+ },
+ {
+ "epoch": 17.767888307155324,
+ "grad_norm": 0.043866172432899475,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 4.2777557373046875,
+ "step": 1279
+ },
+ {
+ "epoch": 17.781849912739965,
+ "grad_norm": 0.04693951830267906,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 4.272597312927246,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79581151832461,
+ "grad_norm": 0.04072774201631546,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 4.284282684326172,
+ "step": 1281
+ },
+ {
+ "epoch": 17.80977312390925,
+ "grad_norm": 0.04241872951388359,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 4.254406452178955,
+ "step": 1282
+ },
+ {
+ "epoch": 17.823734729493893,
+ "grad_norm": 0.04614942893385887,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 4.297187805175781,
+ "step": 1283
+ },
+ {
+ "epoch": 17.837696335078533,
+ "grad_norm": 0.04326155036687851,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 4.292263984680176,
+ "step": 1284
+ },
+ {
+ "epoch": 17.851657940663177,
+ "grad_norm": 0.04053817689418793,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 4.2925920486450195,
+ "step": 1285
+ },
+ {
+ "epoch": 17.865619546247817,
+ "grad_norm": 0.040867388248443604,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 4.276965141296387,
+ "step": 1286
+ },
+ {
+ "epoch": 17.87958115183246,
+ "grad_norm": 0.041000183671712875,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 4.257638454437256,
+ "step": 1287
+ },
+ {
+ "epoch": 17.8935427574171,
+ "grad_norm": 0.04014866054058075,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 4.2734479904174805,
+ "step": 1288
+ },
+ {
+ "epoch": 17.907504363001745,
+ "grad_norm": 0.041237372905015945,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 4.263461112976074,
+ "step": 1289
+ },
+ {
+ "epoch": 17.921465968586386,
+ "grad_norm": 0.039109110832214355,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 4.295286655426025,
+ "step": 1290
+ },
+ {
+ "epoch": 17.93542757417103,
+ "grad_norm": 0.03911379352211952,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 4.254448413848877,
+ "step": 1291
+ },
+ {
+ "epoch": 17.949389179755673,
+ "grad_norm": 0.03997721150517464,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 4.284714698791504,
+ "step": 1292
+ },
+ {
+ "epoch": 17.963350785340314,
+ "grad_norm": 0.04011402651667595,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 4.278304100036621,
+ "step": 1293
+ },
+ {
+ "epoch": 17.977312390924958,
+ "grad_norm": 0.04040595516562462,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 4.288183212280273,
+ "step": 1294
+ },
+ {
+ "epoch": 17.991273996509598,
+ "grad_norm": 0.038857635110616684,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 4.274664402008057,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.028110472485423088,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 2.6625723838806152,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.5912551283836365,
+ "eval_runtime": 60.273,
+ "eval_samples_per_second": 40.516,
+ "eval_steps_per_second": 0.647,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013961605584644,
+ "grad_norm": 0.041634928435087204,
+ "learning_rate": 0.000444677103403854,
+ "loss": 4.181324005126953,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027923211169284,
+ "grad_norm": 0.051596369594335556,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 4.204722881317139,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041884816753928,
+ "grad_norm": 0.058580417186021805,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 4.207996845245361,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05584642233857,
+ "grad_norm": 0.05986522138118744,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 4.184011459350586,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069808027923212,
+ "grad_norm": 0.051045969128608704,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 4.148118019104004,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083769633507853,
+ "grad_norm": 0.045477766543626785,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 4.207469940185547,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097731239092496,
+ "grad_norm": 0.0488731674849987,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 4.201406478881836,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111692844677137,
+ "grad_norm": 0.05013517662882805,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 4.192239761352539,
+ "step": 1304
+ },
+ {
+ "epoch": 18.12565445026178,
+ "grad_norm": 0.05091651529073715,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 4.188252925872803,
+ "step": 1305
+ },
+ {
+ "epoch": 18.13961605584642,
+ "grad_norm": 0.051825884729623795,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 4.157929420471191,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153577661431065,
+ "grad_norm": 0.049625713378190994,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 4.208776473999023,
+ "step": 1307
+ },
+ {
+ "epoch": 18.167539267015705,
+ "grad_norm": 0.04816461354494095,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 4.180496692657471,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18150087260035,
+ "grad_norm": 0.04720531404018402,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 4.178536415100098,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195462478184993,
+ "grad_norm": 0.0475568026304245,
+ "learning_rate": 0.000441621628895411,
+ "loss": 4.203497886657715,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209424083769633,
+ "grad_norm": 0.04613947868347168,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 4.226914405822754,
+ "step": 1311
+ },
+ {
+ "epoch": 18.223385689354277,
+ "grad_norm": 0.04585929960012436,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 4.175119400024414,
+ "step": 1312
+ },
+ {
+ "epoch": 18.237347294938917,
+ "grad_norm": 0.045777060091495514,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 4.181760311126709,
+ "step": 1313
+ },
+ {
+ "epoch": 18.25130890052356,
+ "grad_norm": 0.04583602771162987,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 4.179868698120117,
+ "step": 1314
+ },
+ {
+ "epoch": 18.2652705061082,
+ "grad_norm": 0.04706503078341484,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 4.203325271606445,
+ "step": 1315
+ },
+ {
+ "epoch": 18.279232111692846,
+ "grad_norm": 0.045122336596250534,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 4.187152862548828,
+ "step": 1316
+ },
+ {
+ "epoch": 18.293193717277486,
+ "grad_norm": 0.04495825991034508,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 4.2080535888671875,
+ "step": 1317
+ },
+ {
+ "epoch": 18.30715532286213,
+ "grad_norm": 0.04767543077468872,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 4.207225322723389,
+ "step": 1318
+ },
+ {
+ "epoch": 18.32111692844677,
+ "grad_norm": 0.053396135568618774,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 4.227478981018066,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335078534031414,
+ "grad_norm": 0.0512089841067791,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 4.1904754638671875,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349040139616054,
+ "grad_norm": 0.04967977851629257,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 4.2079925537109375,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363001745200698,
+ "grad_norm": 0.04800970479846001,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 4.206037998199463,
+ "step": 1322
+ },
+ {
+ "epoch": 18.376963350785342,
+ "grad_norm": 0.04993556812405586,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 4.210054397583008,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390924956369982,
+ "grad_norm": 0.0480492040514946,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 4.229897499084473,
+ "step": 1324
+ },
+ {
+ "epoch": 18.404886561954626,
+ "grad_norm": 0.048699282109737396,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 4.17500114440918,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418848167539267,
+ "grad_norm": 0.04681490734219551,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 4.196526527404785,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43280977312391,
+ "grad_norm": 0.04674449935555458,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 4.246790409088135,
+ "step": 1327
+ },
+ {
+ "epoch": 18.44677137870855,
+ "grad_norm": 0.04650607705116272,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 4.225878715515137,
+ "step": 1328
+ },
+ {
+ "epoch": 18.460732984293195,
+ "grad_norm": 0.047200269997119904,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 4.20685338973999,
+ "step": 1329
+ },
+ {
+ "epoch": 18.474694589877835,
+ "grad_norm": 0.04322655498981476,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 4.212920665740967,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48865619546248,
+ "grad_norm": 0.04489128664135933,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 4.266653060913086,
+ "step": 1331
+ },
+ {
+ "epoch": 18.50261780104712,
+ "grad_norm": 0.048838935792446136,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 4.2364501953125,
+ "step": 1332
+ },
+ {
+ "epoch": 18.516579406631763,
+ "grad_norm": 0.04893286153674126,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 4.202394485473633,
+ "step": 1333
+ },
+ {
+ "epoch": 18.530541012216403,
+ "grad_norm": 0.046900101006031036,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 4.209832191467285,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544502617801047,
+ "grad_norm": 0.044056762009859085,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 4.229903221130371,
+ "step": 1335
+ },
+ {
+ "epoch": 18.55846422338569,
+ "grad_norm": 0.04681328684091568,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 4.22783899307251,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57242582897033,
+ "grad_norm": 0.046087440103292465,
+ "learning_rate": 0.000435215814386134,
+ "loss": 4.2078986167907715,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586387434554975,
+ "grad_norm": 0.04363090172410011,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 4.218686103820801,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600349040139616,
+ "grad_norm": 0.046255532652139664,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 4.2475128173828125,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61431064572426,
+ "grad_norm": 0.045545510947704315,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 4.233272552490234,
+ "step": 1340
+ },
+ {
+ "epoch": 18.6282722513089,
+ "grad_norm": 0.045259326696395874,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 4.252043724060059,
+ "step": 1341
+ },
+ {
+ "epoch": 18.642233856893544,
+ "grad_norm": 0.042144861072301865,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 4.247762203216553,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656195462478184,
+ "grad_norm": 0.04298558086156845,
+ "learning_rate": 0.000433781621332752,
+ "loss": 4.222329616546631,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670157068062828,
+ "grad_norm": 0.044070884585380554,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 4.222488880157471,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68411867364747,
+ "grad_norm": 0.044758375734090805,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 4.238160133361816,
+ "step": 1345
+ },
+ {
+ "epoch": 18.698080279232112,
+ "grad_norm": 0.04324857518076897,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 4.220003128051758,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712041884816752,
+ "grad_norm": 0.04349793121218681,
+ "learning_rate": 0.000432823382695327,
+ "loss": 4.251608848571777,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726003490401396,
+ "grad_norm": 0.04483242705464363,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 4.243319511413574,
+ "step": 1348
+ },
+ {
+ "epoch": 18.739965095986037,
+ "grad_norm": 0.04391994699835777,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 4.237479209899902,
+ "step": 1349
+ },
+ {
+ "epoch": 18.75392670157068,
+ "grad_norm": 0.042584922164678574,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 4.2309980392456055,
+ "step": 1350
+ },
+ {
+ "epoch": 18.767888307155324,
+ "grad_norm": 0.04363991320133209,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 4.216980934143066,
+ "step": 1351
+ },
+ {
+ "epoch": 18.781849912739965,
+ "grad_norm": 0.04662707448005676,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 4.257450103759766,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79581151832461,
+ "grad_norm": 0.0462074875831604,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 4.222329139709473,
+ "step": 1353
+ },
+ {
+ "epoch": 18.80977312390925,
+ "grad_norm": 0.043543554842472076,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 4.231960773468018,
+ "step": 1354
+ },
+ {
+ "epoch": 18.823734729493893,
+ "grad_norm": 0.04444567486643791,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 4.225714683532715,
+ "step": 1355
+ },
+ {
+ "epoch": 18.837696335078533,
+ "grad_norm": 0.04507977142930031,
+ "learning_rate": 0.000430661245733797,
+ "loss": 4.234722137451172,
+ "step": 1356
+ },
+ {
+ "epoch": 18.851657940663177,
+ "grad_norm": 0.043786536902189255,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 4.274998664855957,
+ "step": 1357
+ },
+ {
+ "epoch": 18.865619546247817,
+ "grad_norm": 0.04188325256109238,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 4.260570526123047,
+ "step": 1358
+ },
+ {
+ "epoch": 18.87958115183246,
+ "grad_norm": 0.0439927875995636,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 4.250492572784424,
+ "step": 1359
+ },
+ {
+ "epoch": 18.8935427574171,
+ "grad_norm": 0.04520147293806076,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 4.250911712646484,
+ "step": 1360
+ },
+ {
+ "epoch": 18.907504363001745,
+ "grad_norm": 0.04442249983549118,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 4.255832672119141,
+ "step": 1361
+ },
+ {
+ "epoch": 18.921465968586386,
+ "grad_norm": 0.044756144285202026,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 4.226641654968262,
+ "step": 1362
+ },
+ {
+ "epoch": 18.93542757417103,
+ "grad_norm": 0.04647127911448479,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 4.2110915184021,
+ "step": 1363
+ },
+ {
+ "epoch": 18.949389179755673,
+ "grad_norm": 0.045986004173755646,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 4.243263244628906,
+ "step": 1364
+ },
+ {
+ "epoch": 18.963350785340314,
+ "grad_norm": 0.04533308371901512,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 4.253750801086426,
+ "step": 1365
+ },
+ {
+ "epoch": 18.977312390924958,
+ "grad_norm": 0.04307471215724945,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 4.233024597167969,
+ "step": 1366
+ },
+ {
+ "epoch": 18.991273996509598,
+ "grad_norm": 0.0416078194975853,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 4.264965534210205,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.030926328152418137,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 2.651331901550293,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.592528223991394,
+ "eval_runtime": 59.67,
+ "eval_samples_per_second": 40.925,
+ "eval_steps_per_second": 0.654,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013961605584644,
+ "grad_norm": 0.04524356126785278,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 4.153796672821045,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027923211169284,
+ "grad_norm": 0.05105067789554596,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 4.154603004455566,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041884816753928,
+ "grad_norm": 0.052823506295681,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 4.1599273681640625,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05584642233857,
+ "grad_norm": 0.05688423290848732,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 4.124622344970703,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069808027923212,
+ "grad_norm": 0.053178343921899796,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 4.138644218444824,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083769633507853,
+ "grad_norm": 0.05473732203245163,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 4.138207912445068,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097731239092496,
+ "grad_norm": 0.055207233875989914,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 4.155735015869141,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111692844677137,
+ "grad_norm": 0.05484927073121071,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 4.140205383300781,
+ "step": 1376
+ },
+ {
+ "epoch": 19.12565445026178,
+ "grad_norm": 0.05589596927165985,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 4.160840034484863,
+ "step": 1377
+ },
+ {
+ "epoch": 19.13961605584642,
+ "grad_norm": 0.05222902446985245,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 4.154781341552734,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153577661431065,
+ "grad_norm": 0.051824700087308884,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 4.134243488311768,
+ "step": 1379
+ },
+ {
+ "epoch": 19.167539267015705,
+ "grad_norm": 0.0517590269446373,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 4.125945568084717,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18150087260035,
+ "grad_norm": 0.05399726703763008,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 4.146993160247803,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195462478184993,
+ "grad_norm": 0.054430052638053894,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 4.1310906410217285,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209424083769633,
+ "grad_norm": 0.0551968477666378,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 4.170640468597412,
+ "step": 1383
+ },
+ {
+ "epoch": 19.223385689354277,
+ "grad_norm": 0.06050962209701538,
+ "learning_rate": 0.000423881957237287,
+ "loss": 4.166876316070557,
+ "step": 1384
+ },
+ {
+ "epoch": 19.237347294938917,
+ "grad_norm": 0.055587392300367355,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 4.150339603424072,
+ "step": 1385
+ },
+ {
+ "epoch": 19.25130890052356,
+ "grad_norm": 0.05193328112363815,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 4.171754360198975,
+ "step": 1386
+ },
+ {
+ "epoch": 19.2652705061082,
+ "grad_norm": 0.05272402986884117,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 4.165149688720703,
+ "step": 1387
+ },
+ {
+ "epoch": 19.279232111692846,
+ "grad_norm": 0.056283943355083466,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 4.193273544311523,
+ "step": 1388
+ },
+ {
+ "epoch": 19.293193717277486,
+ "grad_norm": 0.05825977027416229,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 4.192652225494385,
+ "step": 1389
+ },
+ {
+ "epoch": 19.30715532286213,
+ "grad_norm": 0.05539379641413689,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 4.187044143676758,
+ "step": 1390
+ },
+ {
+ "epoch": 19.32111692844677,
+ "grad_norm": 0.05300498753786087,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 4.168220043182373,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335078534031414,
+ "grad_norm": 0.05425756797194481,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 4.185193061828613,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349040139616054,
+ "grad_norm": 0.05527182295918465,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 4.1618332862854,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363001745200698,
+ "grad_norm": 0.054422613233327866,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 4.195295810699463,
+ "step": 1394
+ },
+ {
+ "epoch": 19.376963350785342,
+ "grad_norm": 0.052717432379722595,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 4.143290996551514,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390924956369982,
+ "grad_norm": 0.05154350399971008,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 4.115989685058594,
+ "step": 1396
+ },
+ {
+ "epoch": 19.404886561954626,
+ "grad_norm": 0.0489451140165329,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 4.169506072998047,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418848167539267,
+ "grad_norm": 0.049093786627054214,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 4.164122104644775,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43280977312391,
+ "grad_norm": 0.04690805822610855,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 4.186679840087891,
+ "step": 1399
+ },
+ {
+ "epoch": 19.44677137870855,
+ "grad_norm": 0.046708665788173676,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 4.184325695037842,
+ "step": 1400
+ },
+ {
+ "epoch": 19.460732984293195,
+ "grad_norm": 0.04617927595973015,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 4.201678276062012,
+ "step": 1401
+ },
+ {
+ "epoch": 19.474694589877835,
+ "grad_norm": 0.05017530545592308,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 4.18928337097168,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48865619546248,
+ "grad_norm": 0.05197977274656296,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 4.188801288604736,
+ "step": 1403
+ },
+ {
+ "epoch": 19.50261780104712,
+ "grad_norm": 0.049691297113895416,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 4.209467887878418,
+ "step": 1404
+ },
+ {
+ "epoch": 19.516579406631763,
+ "grad_norm": 0.049709953367710114,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 4.19733190536499,
+ "step": 1405
+ },
+ {
+ "epoch": 19.530541012216403,
+ "grad_norm": 0.05274137482047081,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 4.221798896789551,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544502617801047,
+ "grad_norm": 0.05015392601490021,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 4.210565567016602,
+ "step": 1407
+ },
+ {
+ "epoch": 19.55846422338569,
+ "grad_norm": 0.05023058131337166,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 4.184835910797119,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57242582897033,
+ "grad_norm": 0.0479239858686924,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 4.199846267700195,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586387434554975,
+ "grad_norm": 0.0481228232383728,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 4.185157775878906,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600349040139616,
+ "grad_norm": 0.04909738525748253,
+ "learning_rate": 0.000417272427439646,
+ "loss": 4.207550525665283,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61431064572426,
+ "grad_norm": 0.0475567989051342,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 4.213225841522217,
+ "step": 1412
+ },
+ {
+ "epoch": 19.6282722513089,
+ "grad_norm": 0.048073310405015945,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 4.212669372558594,
+ "step": 1413
+ },
+ {
+ "epoch": 19.642233856893544,
+ "grad_norm": 0.04523796588182449,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 4.1950860023498535,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656195462478184,
+ "grad_norm": 0.04359745234251022,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 4.2136640548706055,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670157068062828,
+ "grad_norm": 0.047282423824071884,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 4.209680557250977,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68411867364747,
+ "grad_norm": 0.0474061481654644,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 4.188494682312012,
+ "step": 1417
+ },
+ {
+ "epoch": 19.698080279232112,
+ "grad_norm": 0.04402102902531624,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 4.207236289978027,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712041884816752,
+ "grad_norm": 0.04398655146360397,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 4.210899353027344,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726003490401396,
+ "grad_norm": 0.045767925679683685,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 4.171738624572754,
+ "step": 1420
+ },
+ {
+ "epoch": 19.739965095986037,
+ "grad_norm": 0.04498986527323723,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 4.222414970397949,
+ "step": 1421
+ },
+ {
+ "epoch": 19.75392670157068,
+ "grad_norm": 0.04282551258802414,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 4.202138900756836,
+ "step": 1422
+ },
+ {
+ "epoch": 19.767888307155324,
+ "grad_norm": 0.04392262175679207,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 4.184439659118652,
+ "step": 1423
+ },
+ {
+ "epoch": 19.781849912739965,
+ "grad_norm": 0.04343154653906822,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 4.204674243927002,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79581151832461,
+ "grad_norm": 0.04496419429779053,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 4.23316764831543,
+ "step": 1425
+ },
+ {
+ "epoch": 19.80977312390925,
+ "grad_norm": 0.04488370940089226,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 4.201659202575684,
+ "step": 1426
+ },
+ {
+ "epoch": 19.823734729493893,
+ "grad_norm": 0.045598775148391724,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 4.187412261962891,
+ "step": 1427
+ },
+ {
+ "epoch": 19.837696335078533,
+ "grad_norm": 0.043846048414707184,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 4.215444564819336,
+ "step": 1428
+ },
+ {
+ "epoch": 19.851657940663177,
+ "grad_norm": 0.043062254786491394,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 4.2074055671691895,
+ "step": 1429
+ },
+ {
+ "epoch": 19.865619546247817,
+ "grad_norm": 0.04544302448630333,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 4.184919357299805,
+ "step": 1430
+ },
+ {
+ "epoch": 19.87958115183246,
+ "grad_norm": 0.046144865453243256,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 4.2175750732421875,
+ "step": 1431
+ },
+ {
+ "epoch": 19.8935427574171,
+ "grad_norm": 0.04465445503592491,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 4.1851115226745605,
+ "step": 1432
+ },
+ {
+ "epoch": 19.907504363001745,
+ "grad_norm": 0.045913565903902054,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 4.2226152420043945,
+ "step": 1433
+ },
+ {
+ "epoch": 19.921465968586386,
+ "grad_norm": 0.04401300847530365,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 4.216772556304932,
+ "step": 1434
+ },
+ {
+ "epoch": 19.93542757417103,
+ "grad_norm": 0.042624711990356445,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 4.2133684158325195,
+ "step": 1435
+ },
+ {
+ "epoch": 19.949389179755673,
+ "grad_norm": 0.04533256217837334,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 4.218449592590332,
+ "step": 1436
+ },
+ {
+ "epoch": 19.963350785340314,
+ "grad_norm": 0.04764309898018837,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 4.191402912139893,
+ "step": 1437
+ },
+ {
+ "epoch": 19.977312390924958,
+ "grad_norm": 0.04689318686723709,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 4.2113037109375,
+ "step": 1438
+ },
+ {
+ "epoch": 19.991273996509598,
+ "grad_norm": 0.04479913413524628,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 4.194989204406738,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.03151386231184006,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 2.6285438537597656,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 0.5938133597373962,
+ "eval_runtime": 58.9656,
+ "eval_samples_per_second": 41.414,
+ "eval_steps_per_second": 0.661,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013961605584644,
+ "grad_norm": 0.04884875938296318,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 4.1257734298706055,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027923211169284,
+ "grad_norm": 0.05968257039785385,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 4.091641426086426,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041884816753928,
+ "grad_norm": 0.06437838077545166,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 4.122465133666992,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05584642233857,
+ "grad_norm": 0.057831212878227234,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 4.129654884338379,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069808027923212,
+ "grad_norm": 0.049407217651605606,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 4.116935729980469,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083769633507853,
+ "grad_norm": 0.051957421004772186,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 4.125039577484131,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097731239092496,
+ "grad_norm": 0.05425718426704407,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 4.113485336303711,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111692844677137,
+ "grad_norm": 0.055318959057331085,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 4.119422912597656,
+ "step": 1448
+ },
+ {
+ "epoch": 20.12565445026178,
+ "grad_norm": 0.0548059344291687,
+ "learning_rate": 0.000407857329622967,
+ "loss": 4.097405910491943,
+ "step": 1449
+ },
+ {
+ "epoch": 20.13961605584642,
+ "grad_norm": 0.051046356558799744,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 4.115051746368408,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153577661431065,
+ "grad_norm": 0.05079492926597595,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 4.126070499420166,
+ "step": 1451
+ },
+ {
+ "epoch": 20.167539267015705,
+ "grad_norm": 0.04777265712618828,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 4.118952751159668,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18150087260035,
+ "grad_norm": 0.04895703122019768,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 4.095477104187012,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195462478184993,
+ "grad_norm": 0.051196496933698654,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 4.116866111755371,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209424083769633,
+ "grad_norm": 0.05625293031334877,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 4.141980171203613,
+ "step": 1455
+ },
+ {
+ "epoch": 20.223385689354277,
+ "grad_norm": 0.05852660909295082,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 4.126821517944336,
+ "step": 1456
+ },
+ {
+ "epoch": 20.237347294938917,
+ "grad_norm": 0.05419151484966278,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 4.10935640335083,
+ "step": 1457
+ },
+ {
+ "epoch": 20.25130890052356,
+ "grad_norm": 0.05251913517713547,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 4.151289463043213,
+ "step": 1458
+ },
+ {
+ "epoch": 20.2652705061082,
+ "grad_norm": 0.05223092809319496,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 4.149935722351074,
+ "step": 1459
+ },
+ {
+ "epoch": 20.279232111692846,
+ "grad_norm": 0.05314765125513077,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 4.106821060180664,
+ "step": 1460
+ },
+ {
+ "epoch": 20.293193717277486,
+ "grad_norm": 0.05472502112388611,
+ "learning_rate": 0.000404858275823277,
+ "loss": 4.120498180389404,
+ "step": 1461
+ },
+ {
+ "epoch": 20.30715532286213,
+ "grad_norm": 0.05233139917254448,
+ "learning_rate": 0.000404607816444578,
+ "loss": 4.1237335205078125,
+ "step": 1462
+ },
+ {
+ "epoch": 20.32111692844677,
+ "grad_norm": 0.05210232734680176,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 4.1271209716796875,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335078534031414,
+ "grad_norm": 0.049779925495386124,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 4.121427536010742,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349040139616054,
+ "grad_norm": 0.05232279747724533,
+ "learning_rate": 0.000403855947934478,
+ "loss": 4.136429786682129,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363001745200698,
+ "grad_norm": 0.05235981196165085,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 4.1808600425720215,
+ "step": 1466
+ },
+ {
+ "epoch": 20.376963350785342,
+ "grad_norm": 0.04902610555291176,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 4.138566017150879,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390924956369982,
+ "grad_norm": 0.04923798143863678,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 4.135126113891602,
+ "step": 1468
+ },
+ {
+ "epoch": 20.404886561954626,
+ "grad_norm": 0.05059202015399933,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 4.1548285484313965,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418848167539267,
+ "grad_norm": 0.05058307573199272,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 4.149041652679443,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43280977312391,
+ "grad_norm": 0.050172705203294754,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 4.148054122924805,
+ "step": 1471
+ },
+ {
+ "epoch": 20.44677137870855,
+ "grad_norm": 0.05158299580216408,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 4.1425018310546875,
+ "step": 1472
+ },
+ {
+ "epoch": 20.460732984293195,
+ "grad_norm": 0.052176836878061295,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 4.129712104797363,
+ "step": 1473
+ },
+ {
+ "epoch": 20.474694589877835,
+ "grad_norm": 0.052095964550971985,
+ "learning_rate": 0.000401595976318565,
+ "loss": 4.134920597076416,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48865619546248,
+ "grad_norm": 0.04944197088479996,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 4.165334701538086,
+ "step": 1475
+ },
+ {
+ "epoch": 20.50261780104712,
+ "grad_norm": 0.049351636320352554,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 4.144362926483154,
+ "step": 1476
+ },
+ {
+ "epoch": 20.516579406631763,
+ "grad_norm": 0.053657080978155136,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 4.164234161376953,
+ "step": 1477
+ },
+ {
+ "epoch": 20.530541012216403,
+ "grad_norm": 0.052578140050172806,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 4.131514549255371,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544502617801047,
+ "grad_norm": 0.049683794379234314,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 4.129961967468262,
+ "step": 1479
+ },
+ {
+ "epoch": 20.55846422338569,
+ "grad_norm": 0.05115316063165665,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 4.164663314819336,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57242582897033,
+ "grad_norm": 0.051620446145534515,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 4.161067485809326,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586387434554975,
+ "grad_norm": 0.04873501881957054,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 4.161318302154541,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600349040139616,
+ "grad_norm": 0.04968130216002464,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 4.151819229125977,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61431064572426,
+ "grad_norm": 0.05476681888103485,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 4.14247989654541,
+ "step": 1484
+ },
+ {
+ "epoch": 20.6282722513089,
+ "grad_norm": 0.05098642408847809,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 4.178083419799805,
+ "step": 1485
+ },
+ {
+ "epoch": 20.642233856893544,
+ "grad_norm": 0.047264356166124344,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 4.15170955657959,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656195462478184,
+ "grad_norm": 0.050382886081933975,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 4.166364669799805,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670157068062828,
+ "grad_norm": 0.04956239461898804,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 4.164124488830566,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68411867364747,
+ "grad_norm": 0.049608584493398666,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 4.130239486694336,
+ "step": 1489
+ },
+ {
+ "epoch": 20.698080279232112,
+ "grad_norm": 0.0489494763314724,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 4.163503646850586,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712041884816752,
+ "grad_norm": 0.04847714304924011,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 4.1744585037231445,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726003490401396,
+ "grad_norm": 0.050256017595529556,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 4.1502227783203125,
+ "step": 1492
+ },
+ {
+ "epoch": 20.739965095986037,
+ "grad_norm": 0.048107463866472244,
+ "learning_rate": 0.000396803974909638,
+ "loss": 4.1790971755981445,
+ "step": 1493
+ },
+ {
+ "epoch": 20.75392670157068,
+ "grad_norm": 0.04828944429755211,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 4.128883361816406,
+ "step": 1494
+ },
+ {
+ "epoch": 20.767888307155324,
+ "grad_norm": 0.04720750078558922,
+ "learning_rate": 0.000396297942393265,
+ "loss": 4.1661577224731445,
+ "step": 1495
+ },
+ {
+ "epoch": 20.781849912739965,
+ "grad_norm": 0.047148339450359344,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 4.172522068023682,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79581151832461,
+ "grad_norm": 0.04612366110086441,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 4.149024963378906,
+ "step": 1497
+ },
+ {
+ "epoch": 20.80977312390925,
+ "grad_norm": 0.048896901309490204,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 4.179235458374023,
+ "step": 1498
+ },
+ {
+ "epoch": 20.823734729493893,
+ "grad_norm": 0.04779870808124542,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 4.1649932861328125,
+ "step": 1499
+ },
+ {
+ "epoch": 20.837696335078533,
+ "grad_norm": 0.04949707165360451,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 4.17645263671875,
+ "step": 1500
+ },
+ {
+ "epoch": 20.851657940663177,
+ "grad_norm": 0.04645023122429848,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 4.216496467590332,
+ "step": 1501
+ },
+ {
+ "epoch": 20.865619546247817,
+ "grad_norm": 0.046520136296749115,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 4.152667999267578,
+ "step": 1502
+ },
+ {
+ "epoch": 20.87958115183246,
+ "grad_norm": 0.04882653057575226,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 4.187082767486572,
+ "step": 1503
+ },
+ {
+ "epoch": 20.8935427574171,
+ "grad_norm": 0.04669157415628433,
+ "learning_rate": 0.000394017098798633,
+ "loss": 4.172348499298096,
+ "step": 1504
+ },
+ {
+ "epoch": 20.907504363001745,
+ "grad_norm": 0.04595336690545082,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 4.180559158325195,
+ "step": 1505
+ },
+ {
+ "epoch": 20.921465968586386,
+ "grad_norm": 0.048478737473487854,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 4.142880439758301,
+ "step": 1506
+ },
+ {
+ "epoch": 20.93542757417103,
+ "grad_norm": 0.04609905555844307,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 4.1459641456604,
+ "step": 1507
+ },
+ {
+ "epoch": 20.949389179755673,
+ "grad_norm": 0.04524408280849457,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 4.185099124908447,
+ "step": 1508
+ },
+ {
+ "epoch": 20.963350785340314,
+ "grad_norm": 0.0475911870598793,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 4.2085161209106445,
+ "step": 1509
+ },
+ {
+ "epoch": 20.977312390924958,
+ "grad_norm": 0.047796521335840225,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 4.165589332580566,
+ "step": 1510
+ },
+ {
+ "epoch": 20.991273996509598,
+ "grad_norm": 0.047358278185129166,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 4.1634368896484375,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.03328625112771988,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 2.6023788452148438,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.5952173471450806,
+ "eval_runtime": 58.6816,
+ "eval_samples_per_second": 41.614,
+ "eval_steps_per_second": 0.665,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013961605584644,
+ "grad_norm": 0.04970651492476463,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 4.077130317687988,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027923211169284,
+ "grad_norm": 0.060882844030857086,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 4.074376106262207,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041884816753928,
+ "grad_norm": 0.06609825044870377,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 4.107635021209717,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05584642233857,
+ "grad_norm": 0.056066352874040604,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 4.068084716796875,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069808027923212,
+ "grad_norm": 0.053013063967227936,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 4.069967269897461,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083769633507853,
+ "grad_norm": 0.059536613523960114,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 4.0639142990112305,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097731239092496,
+ "grad_norm": 0.06084490567445755,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 4.093890190124512,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111692844677137,
+ "grad_norm": 0.05299340933561325,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 4.100567817687988,
+ "step": 1520
+ },
+ {
+ "epoch": 21.12565445026178,
+ "grad_norm": 0.05136161297559738,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 4.08171272277832,
+ "step": 1521
+ },
+ {
+ "epoch": 21.13961605584642,
+ "grad_norm": 0.05201329290866852,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 4.0869975090026855,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153577661431065,
+ "grad_norm": 0.053958263248205185,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 4.0928850173950195,
+ "step": 1523
+ },
+ {
+ "epoch": 21.167539267015705,
+ "grad_norm": 0.053465038537979126,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 4.097828388214111,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18150087260035,
+ "grad_norm": 0.050980277359485626,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 4.101456165313721,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195462478184993,
+ "grad_norm": 0.05415358766913414,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 4.09096097946167,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209424083769633,
+ "grad_norm": 0.05567048117518425,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 4.061287879943848,
+ "step": 1527
+ },
+ {
+ "epoch": 21.223385689354277,
+ "grad_norm": 0.052690375596284866,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 4.080180644989014,
+ "step": 1528
+ },
+ {
+ "epoch": 21.237347294938917,
+ "grad_norm": 0.05448971316218376,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 4.079794883728027,
+ "step": 1529
+ },
+ {
+ "epoch": 21.25130890052356,
+ "grad_norm": 0.055472876876592636,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 4.072025299072266,
+ "step": 1530
+ },
+ {
+ "epoch": 21.2652705061082,
+ "grad_norm": 0.05418326333165169,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 4.103858470916748,
+ "step": 1531
+ },
+ {
+ "epoch": 21.279232111692846,
+ "grad_norm": 0.0524221733212471,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 4.0808916091918945,
+ "step": 1532
+ },
+ {
+ "epoch": 21.293193717277486,
+ "grad_norm": 0.05224563553929329,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 4.090270042419434,
+ "step": 1533
+ },
+ {
+ "epoch": 21.30715532286213,
+ "grad_norm": 0.052917953580617905,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 4.081887245178223,
+ "step": 1534
+ },
+ {
+ "epoch": 21.32111692844677,
+ "grad_norm": 0.05619358271360397,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 4.089571475982666,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335078534031414,
+ "grad_norm": 0.05455326288938522,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 4.096992015838623,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349040139616054,
+ "grad_norm": 0.049311455339193344,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 4.114621639251709,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363001745200698,
+ "grad_norm": 0.05175276845693588,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 4.107684135437012,
+ "step": 1538
+ },
+ {
+ "epoch": 21.376963350785342,
+ "grad_norm": 0.05236738547682762,
+ "learning_rate": 0.00038509205478744,
+ "loss": 4.075709342956543,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390924956369982,
+ "grad_norm": 0.04899081960320473,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 4.108396530151367,
+ "step": 1540
+ },
+ {
+ "epoch": 21.404886561954626,
+ "grad_norm": 0.05129663646221161,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 4.1135711669921875,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418848167539267,
+ "grad_norm": 0.05478592589497566,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 4.11839485168457,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43280977312391,
+ "grad_norm": 0.055830392986536026,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 4.11065149307251,
+ "step": 1543
+ },
+ {
+ "epoch": 21.44677137870855,
+ "grad_norm": 0.05370553582906723,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 4.113103866577148,
+ "step": 1544
+ },
+ {
+ "epoch": 21.460732984293195,
+ "grad_norm": 0.04968908056616783,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 4.119443893432617,
+ "step": 1545
+ },
+ {
+ "epoch": 21.474694589877835,
+ "grad_norm": 0.056149110198020935,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 4.126479625701904,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48865619546248,
+ "grad_norm": 0.053045690059661865,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 4.138535499572754,
+ "step": 1547
+ },
+ {
+ "epoch": 21.50261780104712,
+ "grad_norm": 0.04864223673939705,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 4.099347114562988,
+ "step": 1548
+ },
+ {
+ "epoch": 21.516579406631763,
+ "grad_norm": 0.05234768986701965,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 4.112222194671631,
+ "step": 1549
+ },
+ {
+ "epoch": 21.530541012216403,
+ "grad_norm": 0.050486158579587936,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 4.107913970947266,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544502617801047,
+ "grad_norm": 0.048641227185726166,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 4.114014148712158,
+ "step": 1551
+ },
+ {
+ "epoch": 21.55846422338569,
+ "grad_norm": 0.047759316861629486,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 4.128541946411133,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57242582897033,
+ "grad_norm": 0.046098388731479645,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 4.12142276763916,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586387434554975,
+ "grad_norm": 0.04934769123792648,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 4.13035249710083,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600349040139616,
+ "grad_norm": 0.048911672085523605,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 4.111344337463379,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61431064572426,
+ "grad_norm": 0.049368683248758316,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 4.1345672607421875,
+ "step": 1556
+ },
+ {
+ "epoch": 21.6282722513089,
+ "grad_norm": 0.04890736937522888,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 4.1286821365356445,
+ "step": 1557
+ },
+ {
+ "epoch": 21.642233856893544,
+ "grad_norm": 0.04812583327293396,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 4.09993314743042,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656195462478184,
+ "grad_norm": 0.05025152117013931,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 4.150769233703613,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670157068062828,
+ "grad_norm": 0.05007397383451462,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 4.074525833129883,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68411867364747,
+ "grad_norm": 0.050168465822935104,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 4.0933732986450195,
+ "step": 1561
+ },
+ {
+ "epoch": 21.698080279232112,
+ "grad_norm": 0.04990607872605324,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 4.126741886138916,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712041884816752,
+ "grad_norm": 0.05023440346121788,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 4.120314598083496,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726003490401396,
+ "grad_norm": 0.048100546002388,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 4.1242356300354,
+ "step": 1564
+ },
+ {
+ "epoch": 21.739965095986037,
+ "grad_norm": 0.047999415546655655,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 4.127745628356934,
+ "step": 1565
+ },
+ {
+ "epoch": 21.75392670157068,
+ "grad_norm": 0.05180094763636589,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 4.112565994262695,
+ "step": 1566
+ },
+ {
+ "epoch": 21.767888307155324,
+ "grad_norm": 0.053367964923381805,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 4.109419822692871,
+ "step": 1567
+ },
+ {
+ "epoch": 21.781849912739965,
+ "grad_norm": 0.05205400660634041,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 4.111886024475098,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79581151832461,
+ "grad_norm": 0.0464242659509182,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 4.104648590087891,
+ "step": 1569
+ },
+ {
+ "epoch": 21.80977312390925,
+ "grad_norm": 0.047767721116542816,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 4.156681060791016,
+ "step": 1570
+ },
+ {
+ "epoch": 21.823734729493893,
+ "grad_norm": 0.05028790235519409,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 4.132200241088867,
+ "step": 1571
+ },
+ {
+ "epoch": 21.837696335078533,
+ "grad_norm": 0.04714380204677582,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 4.132688999176025,
+ "step": 1572
+ },
+ {
+ "epoch": 21.851657940663177,
+ "grad_norm": 0.04726320132613182,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 4.173277378082275,
+ "step": 1573
+ },
+ {
+ "epoch": 21.865619546247817,
+ "grad_norm": 0.049558594822883606,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 4.15473747253418,
+ "step": 1574
+ },
+ {
+ "epoch": 21.87958115183246,
+ "grad_norm": 0.04742267727851868,
+ "learning_rate": 0.000375827577564042,
+ "loss": 4.10634708404541,
+ "step": 1575
+ },
+ {
+ "epoch": 21.8935427574171,
+ "grad_norm": 0.048223357647657394,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 4.133949279785156,
+ "step": 1576
+ },
+ {
+ "epoch": 21.907504363001745,
+ "grad_norm": 0.046766068786382675,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 4.153807640075684,
+ "step": 1577
+ },
+ {
+ "epoch": 21.921465968586386,
+ "grad_norm": 0.04586656019091606,
+ "learning_rate": 0.000375051971546236,
+ "loss": 4.169903755187988,
+ "step": 1578
+ },
+ {
+ "epoch": 21.93542757417103,
+ "grad_norm": 0.047912538051605225,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 4.148102283477783,
+ "step": 1579
+ },
+ {
+ "epoch": 21.949389179755673,
+ "grad_norm": 0.048949215561151505,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 4.149707317352295,
+ "step": 1580
+ },
+ {
+ "epoch": 21.963350785340314,
+ "grad_norm": 0.04765952378511429,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 4.128008842468262,
+ "step": 1581
+ },
+ {
+ "epoch": 21.977312390924958,
+ "grad_norm": 0.04726482927799225,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 4.1268720626831055,
+ "step": 1582
+ },
+ {
+ "epoch": 21.991273996509598,
+ "grad_norm": 0.0468299426138401,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 4.1332926750183105,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.03415689244866371,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 2.5875940322875977,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.5966277122497559,
+ "eval_runtime": 59.7704,
+ "eval_samples_per_second": 40.856,
+ "eval_steps_per_second": 0.652,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013961605584644,
+ "grad_norm": 0.04954535514116287,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 4.054527759552002,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027923211169284,
+ "grad_norm": 0.0498427115380764,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 4.048842430114746,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041884816753928,
+ "grad_norm": 0.051013074815273285,
+ "learning_rate": 0.000372722041257003,
+ "loss": 4.036187171936035,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05584642233857,
+ "grad_norm": 0.04973271116614342,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 4.050318717956543,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069808027923212,
+ "grad_norm": 0.051144856959581375,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 4.036319732666016,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083769633507853,
+ "grad_norm": 0.051874030381441116,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 4.017857551574707,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097731239092496,
+ "grad_norm": 0.05181736871600151,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 4.0519819259643555,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111692844677137,
+ "grad_norm": 0.05467049032449722,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 4.064291954040527,
+ "step": 1592
+ },
+ {
+ "epoch": 22.12565445026178,
+ "grad_norm": 0.05251661688089371,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 4.056951522827148,
+ "step": 1593
+ },
+ {
+ "epoch": 22.13961605584642,
+ "grad_norm": 0.05278680846095085,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 4.037125587463379,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153577661431065,
+ "grad_norm": 0.05276667699217796,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 4.053675174713135,
+ "step": 1595
+ },
+ {
+ "epoch": 22.167539267015705,
+ "grad_norm": 0.05007271468639374,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 4.044122695922852,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18150087260035,
+ "grad_norm": 0.051073770970106125,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 4.0265960693359375,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195462478184993,
+ "grad_norm": 0.05045647919178009,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 4.060938835144043,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209424083769633,
+ "grad_norm": 0.05075158178806305,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 4.035902976989746,
+ "step": 1599
+ },
+ {
+ "epoch": 22.223385689354277,
+ "grad_norm": 0.04867244139313698,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 4.055220603942871,
+ "step": 1600
+ },
+ {
+ "epoch": 22.237347294938917,
+ "grad_norm": 0.049976419657468796,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 4.026124954223633,
+ "step": 1601
+ },
+ {
+ "epoch": 22.25130890052356,
+ "grad_norm": 0.04968957230448723,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 4.081721782684326,
+ "step": 1602
+ },
+ {
+ "epoch": 22.2652705061082,
+ "grad_norm": 0.0503058098256588,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 4.0384721755981445,
+ "step": 1603
+ },
+ {
+ "epoch": 22.279232111692846,
+ "grad_norm": 0.050643209367990494,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 4.049397945404053,
+ "step": 1604
+ },
+ {
+ "epoch": 22.293193717277486,
+ "grad_norm": 0.05297046899795532,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 4.046389579772949,
+ "step": 1605
+ },
+ {
+ "epoch": 22.30715532286213,
+ "grad_norm": 0.053165238350629807,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 4.090221405029297,
+ "step": 1606
+ },
+ {
+ "epoch": 22.32111692844677,
+ "grad_norm": 0.05104832351207733,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 4.023227691650391,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335078534031414,
+ "grad_norm": 0.048736415803432465,
+ "learning_rate": 0.000367268105959126,
+ "loss": 4.07478666305542,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349040139616054,
+ "grad_norm": 0.05190873518586159,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 4.068147659301758,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363001745200698,
+ "grad_norm": 0.05409928783774376,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 4.060364723205566,
+ "step": 1610
+ },
+ {
+ "epoch": 22.376963350785342,
+ "grad_norm": 0.05112519487738609,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 4.072710990905762,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390924956369982,
+ "grad_norm": 0.04818312078714371,
+ "learning_rate": 0.000366226612810024,
+ "loss": 4.042757034301758,
+ "step": 1612
+ },
+ {
+ "epoch": 22.404886561954626,
+ "grad_norm": 0.050733599811792374,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 4.064810752868652,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418848167539267,
+ "grad_norm": 0.054070211946964264,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 4.054926872253418,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43280977312391,
+ "grad_norm": 0.05459332838654518,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 4.0652923583984375,
+ "step": 1615
+ },
+ {
+ "epoch": 22.44677137870855,
+ "grad_norm": 0.05064154416322708,
+ "learning_rate": 0.000365184304613104,
+ "loss": 4.076216220855713,
+ "step": 1616
+ },
+ {
+ "epoch": 22.460732984293195,
+ "grad_norm": 0.05071895569562912,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 4.070466041564941,
+ "step": 1617
+ },
+ {
+ "epoch": 22.474694589877835,
+ "grad_norm": 0.05049987882375717,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 4.062691688537598,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48865619546248,
+ "grad_norm": 0.05174451321363449,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 4.07501220703125,
+ "step": 1619
+ },
+ {
+ "epoch": 22.50261780104712,
+ "grad_norm": 0.05360838770866394,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 4.054711818695068,
+ "step": 1620
+ },
+ {
+ "epoch": 22.516579406631763,
+ "grad_norm": 0.051809027791023254,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 4.075811386108398,
+ "step": 1621
+ },
+ {
+ "epoch": 22.530541012216403,
+ "grad_norm": 0.05586795508861542,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 4.0638203620910645,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544502617801047,
+ "grad_norm": 0.05571979284286499,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 4.089414596557617,
+ "step": 1623
+ },
+ {
+ "epoch": 22.55846422338569,
+ "grad_norm": 0.05098723992705345,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 4.091159820556641,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57242582897033,
+ "grad_norm": 0.05359277129173279,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 4.060613632202148,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586387434554975,
+ "grad_norm": 0.05281100794672966,
+ "learning_rate": 0.000362575056487632,
+ "loss": 4.08416748046875,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600349040139616,
+ "grad_norm": 0.053913842886686325,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 4.088703155517578,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61431064572426,
+ "grad_norm": 0.05476798862218857,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 4.064074516296387,
+ "step": 1628
+ },
+ {
+ "epoch": 22.6282722513089,
+ "grad_norm": 0.051888495683670044,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 4.116634368896484,
+ "step": 1629
+ },
+ {
+ "epoch": 22.642233856893544,
+ "grad_norm": 0.051054876297712326,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 4.0800018310546875,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656195462478184,
+ "grad_norm": 0.054063159972429276,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 4.112773418426514,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670157068062828,
+ "grad_norm": 0.05173034593462944,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 4.062599182128906,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68411867364747,
+ "grad_norm": 0.054618336260318756,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 4.110188961029053,
+ "step": 1633
+ },
+ {
+ "epoch": 22.698080279232112,
+ "grad_norm": 0.055602896958589554,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 4.140998840332031,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712041884816752,
+ "grad_norm": 0.05332222580909729,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 4.119804382324219,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726003490401396,
+ "grad_norm": 0.05162122845649719,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 4.0731282234191895,
+ "step": 1636
+ },
+ {
+ "epoch": 22.739965095986037,
+ "grad_norm": 0.0514204278588295,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 4.068007946014404,
+ "step": 1637
+ },
+ {
+ "epoch": 22.75392670157068,
+ "grad_norm": 0.05027637258172035,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 4.097658157348633,
+ "step": 1638
+ },
+ {
+ "epoch": 22.767888307155324,
+ "grad_norm": 0.051812347024679184,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 4.083312511444092,
+ "step": 1639
+ },
+ {
+ "epoch": 22.781849912739965,
+ "grad_norm": 0.05145815759897232,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 4.129471778869629,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79581151832461,
+ "grad_norm": 0.05093684792518616,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 4.091841220855713,
+ "step": 1641
+ },
+ {
+ "epoch": 22.80977312390925,
+ "grad_norm": 0.05100737139582634,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 4.076615810394287,
+ "step": 1642
+ },
+ {
+ "epoch": 22.823734729493893,
+ "grad_norm": 0.05079149082303047,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 4.067737579345703,
+ "step": 1643
+ },
+ {
+ "epoch": 22.837696335078533,
+ "grad_norm": 0.05044509842991829,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 4.118824481964111,
+ "step": 1644
+ },
+ {
+ "epoch": 22.851657940663177,
+ "grad_norm": 0.05010584369301796,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 4.113405227661133,
+ "step": 1645
+ },
+ {
+ "epoch": 22.865619546247817,
+ "grad_norm": 0.052232302725315094,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 4.101553440093994,
+ "step": 1646
+ },
+ {
+ "epoch": 22.87958115183246,
+ "grad_norm": 0.05080661550164223,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 4.116809844970703,
+ "step": 1647
+ },
+ {
+ "epoch": 22.8935427574171,
+ "grad_norm": 0.05051033943891525,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 4.128484725952148,
+ "step": 1648
+ },
+ {
+ "epoch": 22.907504363001745,
+ "grad_norm": 0.051927294582128525,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 4.092285633087158,
+ "step": 1649
+ },
+ {
+ "epoch": 22.921465968586386,
+ "grad_norm": 0.049063827842473984,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 4.088473796844482,
+ "step": 1650
+ },
+ {
+ "epoch": 22.93542757417103,
+ "grad_norm": 0.05048854276537895,
+ "learning_rate": 0.000356031397755195,
+ "loss": 4.081521987915039,
+ "step": 1651
+ },
+ {
+ "epoch": 22.949389179755673,
+ "grad_norm": 0.049307651817798615,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 4.122158050537109,
+ "step": 1652
+ },
+ {
+ "epoch": 22.963350785340314,
+ "grad_norm": 0.048311345279216766,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 4.100488662719727,
+ "step": 1653
+ },
+ {
+ "epoch": 22.977312390924958,
+ "grad_norm": 0.05016734451055527,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 4.088587760925293,
+ "step": 1654
+ },
+ {
+ "epoch": 22.991273996509598,
+ "grad_norm": 0.04944450408220291,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 4.114686012268066,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.03592754527926445,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 2.545376777648926,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.5981184840202332,
+ "eval_runtime": 59.9558,
+ "eval_samples_per_second": 40.73,
+ "eval_steps_per_second": 0.65,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013961605584644,
+ "grad_norm": 0.05112785845994949,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 4.00495719909668,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027923211169284,
+ "grad_norm": 0.05807499587535858,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 4.015654563903809,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041884816753928,
+ "grad_norm": 0.05383148789405823,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.979349136352539,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05584642233857,
+ "grad_norm": 0.05221369490027428,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 4.009493827819824,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069808027923212,
+ "grad_norm": 0.05144324526190758,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 4.045680999755859,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083769633507853,
+ "grad_norm": 0.051422782242298126,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.994011878967285,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097731239092496,
+ "grad_norm": 0.05142863467335701,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 4.03209114074707,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111692844677137,
+ "grad_norm": 0.050954073667526245,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 4.00391960144043,
+ "step": 1664
+ },
+ {
+ "epoch": 23.12565445026178,
+ "grad_norm": 0.05365024879574776,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.9839367866516113,
+ "step": 1665
+ },
+ {
+ "epoch": 23.13961605584642,
+ "grad_norm": 0.051181286573410034,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 4.017707824707031,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153577661431065,
+ "grad_norm": 0.054279860109090805,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.989722490310669,
+ "step": 1667
+ },
+ {
+ "epoch": 23.167539267015705,
+ "grad_norm": 0.052196286618709564,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 4.014932155609131,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18150087260035,
+ "grad_norm": 0.05098043382167816,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 4.026510238647461,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195462478184993,
+ "grad_norm": 0.05241888761520386,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.979708433151245,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209424083769633,
+ "grad_norm": 0.051801104098558426,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 4.029015064239502,
+ "step": 1671
+ },
+ {
+ "epoch": 23.223385689354277,
+ "grad_norm": 0.052666619420051575,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.9932432174682617,
+ "step": 1672
+ },
+ {
+ "epoch": 23.237347294938917,
+ "grad_norm": 0.054418303072452545,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 4.05676794052124,
+ "step": 1673
+ },
+ {
+ "epoch": 23.25130890052356,
+ "grad_norm": 0.055074021220207214,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.999929666519165,
+ "step": 1674
+ },
+ {
+ "epoch": 23.2652705061082,
+ "grad_norm": 0.05626136437058449,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 4.013473987579346,
+ "step": 1675
+ },
+ {
+ "epoch": 23.279232111692846,
+ "grad_norm": 0.054028257727622986,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 4.015802383422852,
+ "step": 1676
+ },
+ {
+ "epoch": 23.293193717277486,
+ "grad_norm": 0.054044198244810104,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 4.020464897155762,
+ "step": 1677
+ },
+ {
+ "epoch": 23.30715532286213,
+ "grad_norm": 0.05465259402990341,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 4.010867118835449,
+ "step": 1678
+ },
+ {
+ "epoch": 23.32111692844677,
+ "grad_norm": 0.05545585975050926,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 4.054412841796875,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335078534031414,
+ "grad_norm": 0.055265430361032486,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 4.020756244659424,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349040139616054,
+ "grad_norm": 0.053363267332315445,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 4.033337116241455,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363001745200698,
+ "grad_norm": 0.05347830802202225,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 4.024794578552246,
+ "step": 1682
+ },
+ {
+ "epoch": 23.376963350785342,
+ "grad_norm": 0.05496951565146446,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 4.013302803039551,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390924956369982,
+ "grad_norm": 0.05179804563522339,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 4.030914306640625,
+ "step": 1684
+ },
+ {
+ "epoch": 23.404886561954626,
+ "grad_norm": 0.054958634078502655,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 4.012727737426758,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418848167539267,
+ "grad_norm": 0.05533262714743614,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 4.0543365478515625,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43280977312391,
+ "grad_norm": 0.052329424768686295,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 4.031486511230469,
+ "step": 1687
+ },
+ {
+ "epoch": 23.44677137870855,
+ "grad_norm": 0.05273434892296791,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 4.040292739868164,
+ "step": 1688
+ },
+ {
+ "epoch": 23.460732984293195,
+ "grad_norm": 0.05445481091737747,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 4.025630950927734,
+ "step": 1689
+ },
+ {
+ "epoch": 23.474694589877835,
+ "grad_norm": 0.052102912217378616,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 4.049562454223633,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48865619546248,
+ "grad_norm": 0.051272716373205185,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 4.0431365966796875,
+ "step": 1691
+ },
+ {
+ "epoch": 23.50261780104712,
+ "grad_norm": 0.051311176270246506,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 4.042119026184082,
+ "step": 1692
+ },
+ {
+ "epoch": 23.516579406631763,
+ "grad_norm": 0.05042742192745209,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 4.044881820678711,
+ "step": 1693
+ },
+ {
+ "epoch": 23.530541012216403,
+ "grad_norm": 0.05103316903114319,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 4.033669471740723,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544502617801047,
+ "grad_norm": 0.05441150814294815,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 4.076536178588867,
+ "step": 1695
+ },
+ {
+ "epoch": 23.55846422338569,
+ "grad_norm": 0.05529651418328285,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 4.073678016662598,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57242582897033,
+ "grad_norm": 0.0526033453643322,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 4.067898750305176,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586387434554975,
+ "grad_norm": 0.052817948162555695,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 4.017403602600098,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600349040139616,
+ "grad_norm": 0.052461300045251846,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 4.026311874389648,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61431064572426,
+ "grad_norm": 0.05097689852118492,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 4.076074600219727,
+ "step": 1700
+ },
+ {
+ "epoch": 23.6282722513089,
+ "grad_norm": 0.05387786403298378,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 4.022763729095459,
+ "step": 1701
+ },
+ {
+ "epoch": 23.642233856893544,
+ "grad_norm": 0.05132855102419853,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 4.077712535858154,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656195462478184,
+ "grad_norm": 0.05380886048078537,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 4.038937091827393,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670157068062828,
+ "grad_norm": 0.053299430757761,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 4.063364028930664,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68411867364747,
+ "grad_norm": 0.051913514733314514,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 4.05495548248291,
+ "step": 1705
+ },
+ {
+ "epoch": 23.698080279232112,
+ "grad_norm": 0.05660073831677437,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 4.057949066162109,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712041884816752,
+ "grad_norm": 0.054757390171289444,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 4.05377721786499,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726003490401396,
+ "grad_norm": 0.0517287515103817,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 4.035928726196289,
+ "step": 1708
+ },
+ {
+ "epoch": 23.739965095986037,
+ "grad_norm": 0.05187627300620079,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 4.043362617492676,
+ "step": 1709
+ },
+ {
+ "epoch": 23.75392670157068,
+ "grad_norm": 0.053502220660448074,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 4.007593154907227,
+ "step": 1710
+ },
+ {
+ "epoch": 23.767888307155324,
+ "grad_norm": 0.05755629390478134,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 4.076575756072998,
+ "step": 1711
+ },
+ {
+ "epoch": 23.781849912739965,
+ "grad_norm": 0.05619359388947487,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 4.072240352630615,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79581151832461,
+ "grad_norm": 0.05604318529367447,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 4.0486650466918945,
+ "step": 1713
+ },
+ {
+ "epoch": 23.80977312390925,
+ "grad_norm": 0.05630200728774071,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 4.086282730102539,
+ "step": 1714
+ },
+ {
+ "epoch": 23.823734729493893,
+ "grad_norm": 0.053262822329998016,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 4.089612007141113,
+ "step": 1715
+ },
+ {
+ "epoch": 23.837696335078533,
+ "grad_norm": 0.05699621140956879,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 4.053062438964844,
+ "step": 1716
+ },
+ {
+ "epoch": 23.851657940663177,
+ "grad_norm": 0.05963658541440964,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 4.072884559631348,
+ "step": 1717
+ },
+ {
+ "epoch": 23.865619546247817,
+ "grad_norm": 0.0547492615878582,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 4.066218852996826,
+ "step": 1718
+ },
+ {
+ "epoch": 23.87958115183246,
+ "grad_norm": 0.054908450692892075,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 4.035344123840332,
+ "step": 1719
+ },
+ {
+ "epoch": 23.8935427574171,
+ "grad_norm": 0.055191367864608765,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 4.08471155166626,
+ "step": 1720
+ },
+ {
+ "epoch": 23.907504363001745,
+ "grad_norm": 0.05334226414561272,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 4.076174736022949,
+ "step": 1721
+ },
+ {
+ "epoch": 23.921465968586386,
+ "grad_norm": 0.05579005926847458,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 4.05291223526001,
+ "step": 1722
+ },
+ {
+ "epoch": 23.93542757417103,
+ "grad_norm": 0.05464969575405121,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 4.07335090637207,
+ "step": 1723
+ },
+ {
+ "epoch": 23.949389179755673,
+ "grad_norm": 0.05240754038095474,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 4.0995073318481445,
+ "step": 1724
+ },
+ {
+ "epoch": 23.963350785340314,
+ "grad_norm": 0.05416056886315346,
+ "learning_rate": 0.000336518346307424,
+ "loss": 4.072793960571289,
+ "step": 1725
+ },
+ {
+ "epoch": 23.977312390924958,
+ "grad_norm": 0.05496834218502045,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 4.059562683105469,
+ "step": 1726
+ },
+ {
+ "epoch": 23.991273996509598,
+ "grad_norm": 0.05537887290120125,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 4.050772666931152,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.036677468568086624,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 2.5407257080078125,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.5995977520942688,
+ "eval_runtime": 60.2704,
+ "eval_samples_per_second": 40.517,
+ "eval_steps_per_second": 0.647,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013961605584644,
+ "grad_norm": 0.05583682656288147,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.9873995780944824,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027923211169284,
+ "grad_norm": 0.0605769157409668,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.999206066131592,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041884816753928,
+ "grad_norm": 0.057975225150585175,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.9923095703125,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05584642233857,
+ "grad_norm": 0.056104276329278946,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.946317672729492,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069808027923212,
+ "grad_norm": 0.05481491982936859,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.9973483085632324,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083769633507853,
+ "grad_norm": 0.05460227280855179,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.984617233276367,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097731239092496,
+ "grad_norm": 0.058481909334659576,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 4.0138773918151855,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111692844677137,
+ "grad_norm": 0.05672750994563103,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.9683098793029785,
+ "step": 1736
+ },
+ {
+ "epoch": 24.12565445026178,
+ "grad_norm": 0.05315157398581505,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.9909560680389404,
+ "step": 1737
+ },
+ {
+ "epoch": 24.13961605584642,
+ "grad_norm": 0.05565650761127472,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.954531192779541,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153577661431065,
+ "grad_norm": 0.05550679191946983,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.97672700881958,
+ "step": 1739
+ },
+ {
+ "epoch": 24.167539267015705,
+ "grad_norm": 0.05608576908707619,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.987941265106201,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18150087260035,
+ "grad_norm": 0.05266846343874931,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.97526216506958,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195462478184993,
+ "grad_norm": 0.05439139902591705,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 4.012054920196533,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209424083769633,
+ "grad_norm": 0.05473192036151886,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 4.022945404052734,
+ "step": 1743
+ },
+ {
+ "epoch": 24.223385689354277,
+ "grad_norm": 0.054737839847803116,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.981072425842285,
+ "step": 1744
+ },
+ {
+ "epoch": 24.237347294938917,
+ "grad_norm": 0.05696691572666168,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.9708170890808105,
+ "step": 1745
+ },
+ {
+ "epoch": 24.25130890052356,
+ "grad_norm": 0.05677401274442673,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.985995292663574,
+ "step": 1746
+ },
+ {
+ "epoch": 24.2652705061082,
+ "grad_norm": 0.05895088613033295,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.991887092590332,
+ "step": 1747
+ },
+ {
+ "epoch": 24.279232111692846,
+ "grad_norm": 0.05758961662650108,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.9695687294006348,
+ "step": 1748
+ },
+ {
+ "epoch": 24.293193717277486,
+ "grad_norm": 0.05174405500292778,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.959138870239258,
+ "step": 1749
+ },
+ {
+ "epoch": 24.30715532286213,
+ "grad_norm": 0.055284611880779266,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 4.012698173522949,
+ "step": 1750
+ },
+ {
+ "epoch": 24.32111692844677,
+ "grad_norm": 0.06042446941137314,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.9892425537109375,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335078534031414,
+ "grad_norm": 0.05489135906100273,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.9962222576141357,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349040139616054,
+ "grad_norm": 0.0542326495051384,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.9787094593048096,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363001745200698,
+ "grad_norm": 0.05612313374876976,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.9881339073181152,
+ "step": 1754
+ },
+ {
+ "epoch": 24.376963350785342,
+ "grad_norm": 0.05661709979176521,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 4.009440898895264,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390924956369982,
+ "grad_norm": 0.055537596344947815,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 4.00221061706543,
+ "step": 1756
+ },
+ {
+ "epoch": 24.404886561954626,
+ "grad_norm": 0.054210033267736435,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.9807987213134766,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418848167539267,
+ "grad_norm": 0.05847020074725151,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.960838794708252,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43280977312391,
+ "grad_norm": 0.05386386811733246,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 4.017860412597656,
+ "step": 1759
+ },
+ {
+ "epoch": 24.44677137870855,
+ "grad_norm": 0.056131865829229355,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.980543613433838,
+ "step": 1760
+ },
+ {
+ "epoch": 24.460732984293195,
+ "grad_norm": 0.05648239329457283,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.9948508739471436,
+ "step": 1761
+ },
+ {
+ "epoch": 24.474694589877835,
+ "grad_norm": 0.054138120263814926,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 4.002998352050781,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48865619546248,
+ "grad_norm": 0.05725586414337158,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 4.01864767074585,
+ "step": 1763
+ },
+ {
+ "epoch": 24.50261780104712,
+ "grad_norm": 0.05680665746331215,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 4.023143768310547,
+ "step": 1764
+ },
+ {
+ "epoch": 24.516579406631763,
+ "grad_norm": 0.052627697587013245,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 4.006389617919922,
+ "step": 1765
+ },
+ {
+ "epoch": 24.530541012216403,
+ "grad_norm": 0.053234782069921494,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 4.007419586181641,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544502617801047,
+ "grad_norm": 0.05756475403904915,
+ "learning_rate": 0.000325372061241796,
+ "loss": 4.0226945877075195,
+ "step": 1767
+ },
+ {
+ "epoch": 24.55846422338569,
+ "grad_norm": 0.055328067392110825,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 4.022767066955566,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57242582897033,
+ "grad_norm": 0.05378621071577072,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 4.005022048950195,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586387434554975,
+ "grad_norm": 0.056049644947052,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 4.025008201599121,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600349040139616,
+ "grad_norm": 0.057860117405653,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 4.003593444824219,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61431064572426,
+ "grad_norm": 0.056685276329517365,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.997868299484253,
+ "step": 1772
+ },
+ {
+ "epoch": 24.6282722513089,
+ "grad_norm": 0.053294967859983444,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 4.0211992263793945,
+ "step": 1773
+ },
+ {
+ "epoch": 24.642233856893544,
+ "grad_norm": 0.053967639803886414,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 4.022367477416992,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656195462478184,
+ "grad_norm": 0.052357546985149384,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 4.021792411804199,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670157068062828,
+ "grad_norm": 0.0547296479344368,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 4.008403778076172,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68411867364747,
+ "grad_norm": 0.056104280054569244,
+ "learning_rate": 0.000322712903929477,
+ "loss": 4.033827781677246,
+ "step": 1777
+ },
+ {
+ "epoch": 24.698080279232112,
+ "grad_norm": 0.055415406823158264,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 4.015371322631836,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712041884816752,
+ "grad_norm": 0.054506298154592514,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 4.0000152587890625,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726003490401396,
+ "grad_norm": 0.05468682199716568,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.9905481338500977,
+ "step": 1780
+ },
+ {
+ "epoch": 24.739965095986037,
+ "grad_norm": 0.05323269963264465,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 4.040729999542236,
+ "step": 1781
+ },
+ {
+ "epoch": 24.75392670157068,
+ "grad_norm": 0.052569612860679626,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 4.028299331665039,
+ "step": 1782
+ },
+ {
+ "epoch": 24.767888307155324,
+ "grad_norm": 0.052494753152132034,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 4.028860569000244,
+ "step": 1783
+ },
+ {
+ "epoch": 24.781849912739965,
+ "grad_norm": 0.05346379801630974,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 4.000349044799805,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79581151832461,
+ "grad_norm": 0.051543332636356354,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 4.016782760620117,
+ "step": 1785
+ },
+ {
+ "epoch": 24.80977312390925,
+ "grad_norm": 0.05330873280763626,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 4.024371147155762,
+ "step": 1786
+ },
+ {
+ "epoch": 24.823734729493893,
+ "grad_norm": 0.05306831747293472,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 4.0101213455200195,
+ "step": 1787
+ },
+ {
+ "epoch": 24.837696335078533,
+ "grad_norm": 0.05148417130112648,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 4.020768642425537,
+ "step": 1788
+ },
+ {
+ "epoch": 24.851657940663177,
+ "grad_norm": 0.050848908722400665,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 4.030246734619141,
+ "step": 1789
+ },
+ {
+ "epoch": 24.865619546247817,
+ "grad_norm": 0.05279014632105827,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 4.04910135269165,
+ "step": 1790
+ },
+ {
+ "epoch": 24.87958115183246,
+ "grad_norm": 0.0514594204723835,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 4.025290489196777,
+ "step": 1791
+ },
+ {
+ "epoch": 24.8935427574171,
+ "grad_norm": 0.05200032889842987,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 4.019560813903809,
+ "step": 1792
+ },
+ {
+ "epoch": 24.907504363001745,
+ "grad_norm": 0.052474118769168854,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 4.035177230834961,
+ "step": 1793
+ },
+ {
+ "epoch": 24.921465968586386,
+ "grad_norm": 0.05411219224333763,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.9920859336853027,
+ "step": 1794
+ },
+ {
+ "epoch": 24.93542757417103,
+ "grad_norm": 0.0543992817401886,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 4.017391204833984,
+ "step": 1795
+ },
+ {
+ "epoch": 24.949389179755673,
+ "grad_norm": 0.05187678337097168,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 4.042266368865967,
+ "step": 1796
+ },
+ {
+ "epoch": 24.963350785340314,
+ "grad_norm": 0.053762417286634445,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 4.029013633728027,
+ "step": 1797
+ },
+ {
+ "epoch": 24.977312390924958,
+ "grad_norm": 0.051723796874284744,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 4.038886070251465,
+ "step": 1798
+ },
+ {
+ "epoch": 24.991273996509598,
+ "grad_norm": 0.05242295190691948,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 4.056063652038574,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.03998562693595886,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 2.534770965576172,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 0.6009227633476257,
+ "eval_runtime": 58.9421,
+ "eval_samples_per_second": 41.43,
+ "eval_steps_per_second": 0.662,
+ "step": 1800
+ },
+ {
+ "epoch": 25.013961605584644,
+ "grad_norm": 0.05450186878442764,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 3.936816692352295,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027923211169284,
+ "grad_norm": 0.06389589607715607,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 3.9368391036987305,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041884816753928,
+ "grad_norm": 0.06028348580002785,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 3.953296661376953,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05584642233857,
+ "grad_norm": 0.05359262600541115,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 3.924558639526367,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069808027923212,
+ "grad_norm": 0.059837762266397476,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 3.9608402252197266,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083769633507853,
+ "grad_norm": 0.05523517727851868,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 3.943776845932007,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097731239092496,
+ "grad_norm": 0.05617433041334152,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 3.9635634422302246,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111692844677137,
+ "grad_norm": 0.06066269427537918,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 3.9647364616394043,
+ "step": 1808
+ },
+ {
+ "epoch": 25.12565445026178,
+ "grad_norm": 0.05491397902369499,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 3.95222806930542,
+ "step": 1809
+ },
+ {
+ "epoch": 25.13961605584642,
+ "grad_norm": 0.05731203034520149,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 3.9440197944641113,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153577661431065,
+ "grad_norm": 0.05428403243422508,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 3.9428441524505615,
+ "step": 1811
+ },
+ {
+ "epoch": 25.167539267015705,
+ "grad_norm": 0.05487453565001488,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 3.9489784240722656,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18150087260035,
+ "grad_norm": 0.05725884810090065,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 3.9643478393554688,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195462478184993,
+ "grad_norm": 0.05752462521195412,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 3.963301181793213,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209424083769633,
+ "grad_norm": 0.057387594133615494,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 3.9427294731140137,
+ "step": 1815
+ },
+ {
+ "epoch": 25.223385689354277,
+ "grad_norm": 0.057492151856422424,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 3.9229259490966797,
+ "step": 1816
+ },
+ {
+ "epoch": 25.237347294938917,
+ "grad_norm": 0.05202317237854004,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 3.938037395477295,
+ "step": 1817
+ },
+ {
+ "epoch": 25.25130890052356,
+ "grad_norm": 0.05574658885598183,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 3.9532949924468994,
+ "step": 1818
+ },
+ {
+ "epoch": 25.2652705061082,
+ "grad_norm": 0.05714752897620201,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 3.962207794189453,
+ "step": 1819
+ },
+ {
+ "epoch": 25.279232111692846,
+ "grad_norm": 0.05540093779563904,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 3.96329402923584,
+ "step": 1820
+ },
+ {
+ "epoch": 25.293193717277486,
+ "grad_norm": 0.056597232818603516,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 3.948793411254883,
+ "step": 1821
+ },
+ {
+ "epoch": 25.30715532286213,
+ "grad_norm": 0.055902402848005295,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 3.9111557006835938,
+ "step": 1822
+ },
+ {
+ "epoch": 25.32111692844677,
+ "grad_norm": 0.05509798601269722,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 3.929591655731201,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335078534031414,
+ "grad_norm": 0.05763949453830719,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 3.9758448600769043,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349040139616054,
+ "grad_norm": 0.053838349878787994,
+ "learning_rate": 0.000309930621834911,
+ "loss": 3.92457914352417,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363001745200698,
+ "grad_norm": 0.05467558279633522,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 3.9501285552978516,
+ "step": 1826
+ },
+ {
+ "epoch": 25.376963350785342,
+ "grad_norm": 0.05715392529964447,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 3.9734368324279785,
+ "step": 1827
+ },
+ {
+ "epoch": 25.390924956369982,
+ "grad_norm": 0.0566536970436573,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 3.9933152198791504,
+ "step": 1828
+ },
+ {
+ "epoch": 25.404886561954626,
+ "grad_norm": 0.053221236914396286,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 3.967787265777588,
+ "step": 1829
+ },
+ {
+ "epoch": 25.418848167539267,
+ "grad_norm": 0.05779031664133072,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 3.9415433406829834,
+ "step": 1830
+ },
+ {
+ "epoch": 25.43280977312391,
+ "grad_norm": 0.062013521790504456,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 3.9801456928253174,
+ "step": 1831
+ },
+ {
+ "epoch": 25.44677137870855,
+ "grad_norm": 0.05819004401564598,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 3.966928005218506,
+ "step": 1832
+ },
+ {
+ "epoch": 25.460732984293195,
+ "grad_norm": 0.05754564702510834,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 3.9612770080566406,
+ "step": 1833
+ },
+ {
+ "epoch": 25.474694589877835,
+ "grad_norm": 0.05707791447639465,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 3.935521125793457,
+ "step": 1834
+ },
+ {
+ "epoch": 25.48865619546248,
+ "grad_norm": 0.056380245834589005,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 4.006120681762695,
+ "step": 1835
+ },
+ {
+ "epoch": 25.50261780104712,
+ "grad_norm": 0.0551077239215374,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 3.9544754028320312,
+ "step": 1836
+ },
+ {
+ "epoch": 25.516579406631763,
+ "grad_norm": 0.05720360204577446,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 3.9564664363861084,
+ "step": 1837
+ },
+ {
+ "epoch": 25.530541012216403,
+ "grad_norm": 0.05522238463163376,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 4.0157880783081055,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544502617801047,
+ "grad_norm": 0.05304659530520439,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 3.957207202911377,
+ "step": 1839
+ },
+ {
+ "epoch": 25.55846422338569,
+ "grad_norm": 0.05569292604923248,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 3.9909608364105225,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57242582897033,
+ "grad_norm": 0.05347870662808418,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 3.9537010192871094,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586387434554975,
+ "grad_norm": 0.05572560802102089,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 3.9818642139434814,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600349040139616,
+ "grad_norm": 0.05624900385737419,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 3.995117425918579,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61431064572426,
+ "grad_norm": 0.05600332096219063,
+ "learning_rate": 0.000304866093757771,
+ "loss": 3.9480154514312744,
+ "step": 1844
+ },
+ {
+ "epoch": 25.6282722513089,
+ "grad_norm": 0.05488704890012741,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 3.993014335632324,
+ "step": 1845
+ },
+ {
+ "epoch": 25.642233856893544,
+ "grad_norm": 0.05647030100226402,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 3.994615077972412,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656195462478184,
+ "grad_norm": 0.05460594967007637,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 3.9714808464050293,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670157068062828,
+ "grad_norm": 0.05629213526844978,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 3.962033271789551,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68411867364747,
+ "grad_norm": 0.05666861683130264,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 3.995985269546509,
+ "step": 1849
+ },
+ {
+ "epoch": 25.698080279232112,
+ "grad_norm": 0.05580414459109306,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 4.001766681671143,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712041884816752,
+ "grad_norm": 0.05638023093342781,
+ "learning_rate": 0.000303,
+ "loss": 4.005331993103027,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726003490401396,
+ "grad_norm": 0.05684564262628555,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 4.017581939697266,
+ "step": 1852
+ },
+ {
+ "epoch": 25.739965095986037,
+ "grad_norm": 0.057602159678936005,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 3.9668312072753906,
+ "step": 1853
+ },
+ {
+ "epoch": 25.75392670157068,
+ "grad_norm": 0.055540211498737335,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 3.9885573387145996,
+ "step": 1854
+ },
+ {
+ "epoch": 25.767888307155324,
+ "grad_norm": 0.05655810981988907,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 3.990997314453125,
+ "step": 1855
+ },
+ {
+ "epoch": 25.781849912739965,
+ "grad_norm": 0.05443689972162247,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 3.9554123878479004,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79581151832461,
+ "grad_norm": 0.05570529028773308,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 3.983283519744873,
+ "step": 1857
+ },
+ {
+ "epoch": 25.80977312390925,
+ "grad_norm": 0.05399419739842415,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 4.004507064819336,
+ "step": 1858
+ },
+ {
+ "epoch": 25.823734729493893,
+ "grad_norm": 0.05587748438119888,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 4.00900936126709,
+ "step": 1859
+ },
+ {
+ "epoch": 25.837696335078533,
+ "grad_norm": 0.05465575307607651,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 3.9826579093933105,
+ "step": 1860
+ },
+ {
+ "epoch": 25.851657940663177,
+ "grad_norm": 0.05286499112844467,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 3.973453998565674,
+ "step": 1861
+ },
+ {
+ "epoch": 25.865619546247817,
+ "grad_norm": 0.05495590344071388,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 4.003095626831055,
+ "step": 1862
+ },
+ {
+ "epoch": 25.87958115183246,
+ "grad_norm": 0.053388018161058426,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 4.00042200088501,
+ "step": 1863
+ },
+ {
+ "epoch": 25.8935427574171,
+ "grad_norm": 0.05323237553238869,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 3.992929458618164,
+ "step": 1864
+ },
+ {
+ "epoch": 25.907504363001745,
+ "grad_norm": 0.053946711122989655,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 3.996519088745117,
+ "step": 1865
+ },
+ {
+ "epoch": 25.921465968586386,
+ "grad_norm": 0.05455862358212471,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 4.007123947143555,
+ "step": 1866
+ },
+ {
+ "epoch": 25.93542757417103,
+ "grad_norm": 0.05290355160832405,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 4.001724720001221,
+ "step": 1867
+ },
+ {
+ "epoch": 25.949389179755673,
+ "grad_norm": 0.052808016538619995,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 3.9964821338653564,
+ "step": 1868
+ },
+ {
+ "epoch": 25.963350785340314,
+ "grad_norm": 0.05416771396994591,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 3.98549222946167,
+ "step": 1869
+ },
+ {
+ "epoch": 25.977312390924958,
+ "grad_norm": 0.05264521390199661,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 3.9729766845703125,
+ "step": 1870
+ },
+ {
+ "epoch": 25.991273996509598,
+ "grad_norm": 0.05483962222933769,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 4.018877029418945,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.03927824646234512,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 2.456951856613159,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 0.6026532053947449,
+ "eval_runtime": 60.4079,
+ "eval_samples_per_second": 40.425,
+ "eval_steps_per_second": 0.646,
+ "step": 1872
+ },
+ {
+ "epoch": 26.013961605584644,
+ "grad_norm": 0.05756894499063492,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 3.93898606300354,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027923211169284,
+ "grad_norm": 0.06288876384496689,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 3.9142136573791504,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041884816753928,
+ "grad_norm": 0.0600322000682354,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 3.918700695037842,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05584642233857,
+ "grad_norm": 0.05729348957538605,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 3.9040162563323975,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069808027923212,
+ "grad_norm": 0.05658194422721863,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 3.8809690475463867,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083769633507853,
+ "grad_norm": 0.06274262815713882,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 3.928222179412842,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097731239092496,
+ "grad_norm": 0.06313030421733856,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 3.9395527839660645,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111692844677137,
+ "grad_norm": 0.0554068423807621,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 3.9314563274383545,
+ "step": 1880
+ },
+ {
+ "epoch": 26.12565445026178,
+ "grad_norm": 0.05697336047887802,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 3.9071288108825684,
+ "step": 1881
+ },
+ {
+ "epoch": 26.13961605584642,
+ "grad_norm": 0.05603424087166786,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 3.932054042816162,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153577661431065,
+ "grad_norm": 0.055603109300136566,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 3.9169063568115234,
+ "step": 1883
+ },
+ {
+ "epoch": 26.167539267015705,
+ "grad_norm": 0.055022306740283966,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 3.8985838890075684,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18150087260035,
+ "grad_norm": 0.05383468046784401,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 3.896359443664551,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195462478184993,
+ "grad_norm": 0.055076900869607925,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 3.905672073364258,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209424083769633,
+ "grad_norm": 0.05629882961511612,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 3.9222521781921387,
+ "step": 1887
+ },
+ {
+ "epoch": 26.223385689354277,
+ "grad_norm": 0.056736111640930176,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 3.8918867111206055,
+ "step": 1888
+ },
+ {
+ "epoch": 26.237347294938917,
+ "grad_norm": 0.05625338479876518,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 3.9249935150146484,
+ "step": 1889
+ },
+ {
+ "epoch": 26.25130890052356,
+ "grad_norm": 0.06145375967025757,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 3.904536724090576,
+ "step": 1890
+ },
+ {
+ "epoch": 26.2652705061082,
+ "grad_norm": 0.05601578950881958,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 3.8845863342285156,
+ "step": 1891
+ },
+ {
+ "epoch": 26.279232111692846,
+ "grad_norm": 0.056899815797805786,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 3.917966842651367,
+ "step": 1892
+ },
+ {
+ "epoch": 26.293193717277486,
+ "grad_norm": 0.060314927250146866,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 3.9230940341949463,
+ "step": 1893
+ },
+ {
+ "epoch": 26.30715532286213,
+ "grad_norm": 0.05449546128511429,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 3.9082441329956055,
+ "step": 1894
+ },
+ {
+ "epoch": 26.32111692844677,
+ "grad_norm": 0.05601051449775696,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 3.9174978733062744,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335078534031414,
+ "grad_norm": 0.05770742893218994,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 3.9183058738708496,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349040139616054,
+ "grad_norm": 0.05718272924423218,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 3.9438772201538086,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363001745200698,
+ "grad_norm": 0.052873238921165466,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 3.9554827213287354,
+ "step": 1898
+ },
+ {
+ "epoch": 26.376963350785342,
+ "grad_norm": 0.056853555142879486,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 3.886812210083008,
+ "step": 1899
+ },
+ {
+ "epoch": 26.390924956369982,
+ "grad_norm": 0.05662025138735771,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 3.9384284019470215,
+ "step": 1900
+ },
+ {
+ "epoch": 26.404886561954626,
+ "grad_norm": 0.05693329498171806,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 3.9123764038085938,
+ "step": 1901
+ },
+ {
+ "epoch": 26.418848167539267,
+ "grad_norm": 0.05425408110022545,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 3.9441678524017334,
+ "step": 1902
+ },
+ {
+ "epoch": 26.43280977312391,
+ "grad_norm": 0.05533420667052269,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 3.9086289405822754,
+ "step": 1903
+ },
+ {
+ "epoch": 26.44677137870855,
+ "grad_norm": 0.05527901649475098,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 3.9469594955444336,
+ "step": 1904
+ },
+ {
+ "epoch": 26.460732984293195,
+ "grad_norm": 0.052982039749622345,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 3.9182326793670654,
+ "step": 1905
+ },
+ {
+ "epoch": 26.474694589877835,
+ "grad_norm": 0.05520346015691757,
+ "learning_rate": 0.000288343693342466,
+ "loss": 3.954651117324829,
+ "step": 1906
+ },
+ {
+ "epoch": 26.48865619546248,
+ "grad_norm": 0.05564865842461586,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 3.951840400695801,
+ "step": 1907
+ },
+ {
+ "epoch": 26.50261780104712,
+ "grad_norm": 0.05588578060269356,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 3.933762550354004,
+ "step": 1908
+ },
+ {
+ "epoch": 26.516579406631763,
+ "grad_norm": 0.05727043375372887,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 3.9299421310424805,
+ "step": 1909
+ },
+ {
+ "epoch": 26.530541012216403,
+ "grad_norm": 0.05533654987812042,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 3.941525936126709,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544502617801047,
+ "grad_norm": 0.05524713918566704,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 3.935666084289551,
+ "step": 1911
+ },
+ {
+ "epoch": 26.55846422338569,
+ "grad_norm": 0.055435653775930405,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 3.951631546020508,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57242582897033,
+ "grad_norm": 0.05678093433380127,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 3.960881233215332,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586387434554975,
+ "grad_norm": 0.054896436631679535,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 3.9262163639068604,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600349040139616,
+ "grad_norm": 0.055213335901498795,
+ "learning_rate": 0.000285947841605349,
+ "loss": 3.9284324645996094,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61431064572426,
+ "grad_norm": 0.05644496902823448,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 3.9727559089660645,
+ "step": 1916
+ },
+ {
+ "epoch": 26.6282722513089,
+ "grad_norm": 0.05483568087220192,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 3.9410910606384277,
+ "step": 1917
+ },
+ {
+ "epoch": 26.642233856893544,
+ "grad_norm": 0.054636143147945404,
+ "learning_rate": 0.000285149463934261,
+ "loss": 3.945554733276367,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656195462478184,
+ "grad_norm": 0.056062210351228714,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 3.950249671936035,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670157068062828,
+ "grad_norm": 0.05734739080071449,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 3.9527649879455566,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68411867364747,
+ "grad_norm": 0.05562867224216461,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 3.9563350677490234,
+ "step": 1921
+ },
+ {
+ "epoch": 26.698080279232112,
+ "grad_norm": 0.05565498396754265,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 3.9478225708007812,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712041884816752,
+ "grad_norm": 0.057103481143713,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 3.9302189350128174,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726003490401396,
+ "grad_norm": 0.058241210877895355,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 3.9773902893066406,
+ "step": 1924
+ },
+ {
+ "epoch": 26.739965095986037,
+ "grad_norm": 0.055981557816267014,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 3.923896074295044,
+ "step": 1925
+ },
+ {
+ "epoch": 26.75392670157068,
+ "grad_norm": 0.05711016431450844,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 3.972334861755371,
+ "step": 1926
+ },
+ {
+ "epoch": 26.767888307155324,
+ "grad_norm": 0.05707632377743721,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 3.958393096923828,
+ "step": 1927
+ },
+ {
+ "epoch": 26.781849912739965,
+ "grad_norm": 0.055073633790016174,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 3.9449703693389893,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79581151832461,
+ "grad_norm": 0.05765518546104431,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 3.9442358016967773,
+ "step": 1929
+ },
+ {
+ "epoch": 26.80977312390925,
+ "grad_norm": 0.05692309886217117,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 3.9665679931640625,
+ "step": 1930
+ },
+ {
+ "epoch": 26.823734729493893,
+ "grad_norm": 0.05776236951351166,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 3.9153542518615723,
+ "step": 1931
+ },
+ {
+ "epoch": 26.837696335078533,
+ "grad_norm": 0.05465872213244438,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 3.919874429702759,
+ "step": 1932
+ },
+ {
+ "epoch": 26.851657940663177,
+ "grad_norm": 0.057827629148960114,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 3.9578394889831543,
+ "step": 1933
+ },
+ {
+ "epoch": 26.865619546247817,
+ "grad_norm": 0.055902253836393356,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 3.995711326599121,
+ "step": 1934
+ },
+ {
+ "epoch": 26.87958115183246,
+ "grad_norm": 0.05868638679385185,
+ "learning_rate": 0.000280627938758204,
+ "loss": 3.96519136428833,
+ "step": 1935
+ },
+ {
+ "epoch": 26.8935427574171,
+ "grad_norm": 0.059722673147916794,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 3.9566941261291504,
+ "step": 1936
+ },
+ {
+ "epoch": 26.907504363001745,
+ "grad_norm": 0.055209770798683167,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 3.9929540157318115,
+ "step": 1937
+ },
+ {
+ "epoch": 26.921465968586386,
+ "grad_norm": 0.05694293603301048,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 3.9529929161071777,
+ "step": 1938
+ },
+ {
+ "epoch": 26.93542757417103,
+ "grad_norm": 0.058492012321949005,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 3.9465792179107666,
+ "step": 1939
+ },
+ {
+ "epoch": 26.949389179755673,
+ "grad_norm": 0.056036073714494705,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 3.9590697288513184,
+ "step": 1940
+ },
+ {
+ "epoch": 26.963350785340314,
+ "grad_norm": 0.0547766387462616,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 3.969907522201538,
+ "step": 1941
+ },
+ {
+ "epoch": 26.977312390924958,
+ "grad_norm": 0.05592350661754608,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 3.9509897232055664,
+ "step": 1942
+ },
+ {
+ "epoch": 26.991273996509598,
+ "grad_norm": 0.05589321628212929,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 3.9648609161376953,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.03934016078710556,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 2.4784152507781982,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 0.6035652160644531,
+ "eval_runtime": 60.0656,
+ "eval_samples_per_second": 40.656,
+ "eval_steps_per_second": 0.649,
+ "step": 1944
+ },
+ {
+ "epoch": 27.013961605584644,
+ "grad_norm": 0.05773136392235756,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 3.8625826835632324,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027923211169284,
+ "grad_norm": 0.05844702199101448,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 3.8796658515930176,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041884816753928,
+ "grad_norm": 0.05718810483813286,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 3.8808112144470215,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05584642233857,
+ "grad_norm": 0.05264506861567497,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 3.8776121139526367,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069808027923212,
+ "grad_norm": 0.055468373000621796,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 3.866316318511963,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083769633507853,
+ "grad_norm": 0.053565360605716705,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 3.8870849609375,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097731239092496,
+ "grad_norm": 0.05468900501728058,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 3.8581550121307373,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111692844677137,
+ "grad_norm": 0.05454741418361664,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 3.8752541542053223,
+ "step": 1952
+ },
+ {
+ "epoch": 27.12565445026178,
+ "grad_norm": 0.05647914111614227,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 3.8818912506103516,
+ "step": 1953
+ },
+ {
+ "epoch": 27.13961605584642,
+ "grad_norm": 0.05325528234243393,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 3.8562560081481934,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153577661431065,
+ "grad_norm": 0.05581580847501755,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 3.879669189453125,
+ "step": 1955
+ },
+ {
+ "epoch": 27.167539267015705,
+ "grad_norm": 0.058219097554683685,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 3.8530635833740234,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18150087260035,
+ "grad_norm": 0.05613216012716293,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 3.8905630111694336,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195462478184993,
+ "grad_norm": 0.05486270785331726,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 3.8945157527923584,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209424083769633,
+ "grad_norm": 0.05338508263230324,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 3.909775733947754,
+ "step": 1959
+ },
+ {
+ "epoch": 27.223385689354277,
+ "grad_norm": 0.05602312833070755,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 3.8643651008605957,
+ "step": 1960
+ },
+ {
+ "epoch": 27.237347294938917,
+ "grad_norm": 0.05373037979006767,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 3.8915491104125977,
+ "step": 1961
+ },
+ {
+ "epoch": 27.25130890052356,
+ "grad_norm": 0.05805065110325813,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 3.8836984634399414,
+ "step": 1962
+ },
+ {
+ "epoch": 27.2652705061082,
+ "grad_norm": 0.05469847097992897,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 3.8812215328216553,
+ "step": 1963
+ },
+ {
+ "epoch": 27.279232111692846,
+ "grad_norm": 0.05575987696647644,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 3.8887271881103516,
+ "step": 1964
+ },
+ {
+ "epoch": 27.293193717277486,
+ "grad_norm": 0.05577889457345009,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 3.8917694091796875,
+ "step": 1965
+ },
+ {
+ "epoch": 27.30715532286213,
+ "grad_norm": 0.05604076758027077,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 3.881073474884033,
+ "step": 1966
+ },
+ {
+ "epoch": 27.32111692844677,
+ "grad_norm": 0.055674705654382706,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 3.9303488731384277,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335078534031414,
+ "grad_norm": 0.05300940200686455,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 3.871678113937378,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349040139616054,
+ "grad_norm": 0.05647176504135132,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 3.8848423957824707,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363001745200698,
+ "grad_norm": 0.05647560954093933,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 3.931574583053589,
+ "step": 1970
+ },
+ {
+ "epoch": 27.376963350785342,
+ "grad_norm": 0.055338893085718155,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 3.908456563949585,
+ "step": 1971
+ },
+ {
+ "epoch": 27.390924956369982,
+ "grad_norm": 0.05508219450712204,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 3.8933262825012207,
+ "step": 1972
+ },
+ {
+ "epoch": 27.404886561954626,
+ "grad_norm": 0.05787097290158272,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 3.8908863067626953,
+ "step": 1973
+ },
+ {
+ "epoch": 27.418848167539267,
+ "grad_norm": 0.05587272346019745,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 3.9261322021484375,
+ "step": 1974
+ },
+ {
+ "epoch": 27.43280977312391,
+ "grad_norm": 0.05573277547955513,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 3.9129767417907715,
+ "step": 1975
+ },
+ {
+ "epoch": 27.44677137870855,
+ "grad_norm": 0.05628194287419319,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 3.906001091003418,
+ "step": 1976
+ },
+ {
+ "epoch": 27.460732984293195,
+ "grad_norm": 0.055410996079444885,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 3.9163825511932373,
+ "step": 1977
+ },
+ {
+ "epoch": 27.474694589877835,
+ "grad_norm": 0.05579163134098053,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 3.9286627769470215,
+ "step": 1978
+ },
+ {
+ "epoch": 27.48865619546248,
+ "grad_norm": 0.05793057754635811,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 3.8955917358398438,
+ "step": 1979
+ },
+ {
+ "epoch": 27.50261780104712,
+ "grad_norm": 0.05592824146151543,
+ "learning_rate": 0.00026868712586269,
+ "loss": 3.922128677368164,
+ "step": 1980
+ },
+ {
+ "epoch": 27.516579406631763,
+ "grad_norm": 0.05768568441271782,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 3.897780418395996,
+ "step": 1981
+ },
+ {
+ "epoch": 27.530541012216403,
+ "grad_norm": 0.055304404348134995,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 3.886094570159912,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544502617801047,
+ "grad_norm": 0.054745838046073914,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 3.9272961616516113,
+ "step": 1983
+ },
+ {
+ "epoch": 27.55846422338569,
+ "grad_norm": 0.05616277828812599,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 3.8794174194335938,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57242582897033,
+ "grad_norm": 0.05605581775307655,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 3.9165992736816406,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586387434554975,
+ "grad_norm": 0.057485803961753845,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 3.902085542678833,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600349040139616,
+ "grad_norm": 0.0575849711894989,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 3.92405366897583,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61431064572426,
+ "grad_norm": 0.05563105270266533,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 3.8700900077819824,
+ "step": 1988
+ },
+ {
+ "epoch": 27.6282722513089,
+ "grad_norm": 0.05615481734275818,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 3.9526686668395996,
+ "step": 1989
+ },
+ {
+ "epoch": 27.642233856893544,
+ "grad_norm": 0.056746870279312134,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 3.9000678062438965,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656195462478184,
+ "grad_norm": 0.057345934212207794,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 3.910407543182373,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670157068062828,
+ "grad_norm": 0.05434483289718628,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 3.9132578372955322,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68411867364747,
+ "grad_norm": 0.0582718588411808,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 3.9188318252563477,
+ "step": 1993
+ },
+ {
+ "epoch": 27.698080279232112,
+ "grad_norm": 0.05566178634762764,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 3.8843226432800293,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712041884816752,
+ "grad_norm": 0.05550086498260498,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 3.904834032058716,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726003490401396,
+ "grad_norm": 0.05517689883708954,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 3.9082438945770264,
+ "step": 1996
+ },
+ {
+ "epoch": 27.739965095986037,
+ "grad_norm": 0.05598316714167595,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 3.909560203552246,
+ "step": 1997
+ },
+ {
+ "epoch": 27.75392670157068,
+ "grad_norm": 0.054991044104099274,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 3.903012275695801,
+ "step": 1998
+ },
+ {
+ "epoch": 27.767888307155324,
+ "grad_norm": 0.05612454190850258,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 3.9183692932128906,
+ "step": 1999
+ },
+ {
+ "epoch": 27.781849912739965,
+ "grad_norm": 0.055587366223335266,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 3.918454647064209,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79581151832461,
+ "grad_norm": 0.05595824122428894,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 3.895993232727051,
+ "step": 2001
+ },
+ {
+ "epoch": 27.80977312390925,
+ "grad_norm": 0.054730597883462906,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 3.9305458068847656,
+ "step": 2002
+ },
+ {
+ "epoch": 27.823734729493893,
+ "grad_norm": 0.05543222650885582,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 3.9375224113464355,
+ "step": 2003
+ },
+ {
+ "epoch": 27.837696335078533,
+ "grad_norm": 0.055638376623392105,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 3.9242658615112305,
+ "step": 2004
+ },
+ {
+ "epoch": 27.851657940663177,
+ "grad_norm": 0.05635111406445503,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 3.889681339263916,
+ "step": 2005
+ },
+ {
+ "epoch": 27.865619546247817,
+ "grad_norm": 0.056509099900722504,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 3.9191479682922363,
+ "step": 2006
+ },
+ {
+ "epoch": 27.87958115183246,
+ "grad_norm": 0.055455707013607025,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 3.936049222946167,
+ "step": 2007
+ },
+ {
+ "epoch": 27.8935427574171,
+ "grad_norm": 0.05508789047598839,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 3.901210069656372,
+ "step": 2008
+ },
+ {
+ "epoch": 27.907504363001745,
+ "grad_norm": 0.05542607232928276,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 3.9348654747009277,
+ "step": 2009
+ },
+ {
+ "epoch": 27.921465968586386,
+ "grad_norm": 0.05693232640624046,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 3.940847158432007,
+ "step": 2010
+ },
+ {
+ "epoch": 27.93542757417103,
+ "grad_norm": 0.05707051232457161,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 3.9295883178710938,
+ "step": 2011
+ },
+ {
+ "epoch": 27.949389179755673,
+ "grad_norm": 0.05778150260448456,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 3.9317305088043213,
+ "step": 2012
+ },
+ {
+ "epoch": 27.963350785340314,
+ "grad_norm": 0.05821850523352623,
+ "learning_rate": 0.00025996500713765,
+ "loss": 3.9197943210601807,
+ "step": 2013
+ },
+ {
+ "epoch": 27.977312390924958,
+ "grad_norm": 0.05592802166938782,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 3.9471435546875,
+ "step": 2014
+ },
+ {
+ "epoch": 27.991273996509598,
+ "grad_norm": 0.059447627514600754,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 3.9220168590545654,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.04082372784614563,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 2.433501958847046,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 0.6050393581390381,
+ "eval_runtime": 57.4222,
+ "eval_samples_per_second": 42.527,
+ "eval_steps_per_second": 0.679,
+ "step": 2016
+ },
+ {
+ "epoch": 28.013961605584644,
+ "grad_norm": 0.059585411101579666,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 3.8270139694213867,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027923211169284,
+ "grad_norm": 0.0652594342827797,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 3.842228412628174,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041884816753928,
+ "grad_norm": 0.05607042834162712,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 3.8498799800872803,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05584642233857,
+ "grad_norm": 0.05797959491610527,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 3.8537564277648926,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069808027923212,
+ "grad_norm": 0.06011393293738365,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 3.865638256072998,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083769633507853,
+ "grad_norm": 0.05946647375822067,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 3.8406717777252197,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097731239092496,
+ "grad_norm": 0.05657580494880676,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 3.8516461849212646,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111692844677137,
+ "grad_norm": 0.06000044196844101,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 3.859069347381592,
+ "step": 2024
+ },
+ {
+ "epoch": 28.12565445026178,
+ "grad_norm": 0.05713335797190666,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 3.8558874130249023,
+ "step": 2025
+ },
+ {
+ "epoch": 28.13961605584642,
+ "grad_norm": 0.05555511265993118,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 3.882650852203369,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153577661431065,
+ "grad_norm": 0.05772312358021736,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 3.847323179244995,
+ "step": 2027
+ },
+ {
+ "epoch": 28.167539267015705,
+ "grad_norm": 0.056661207228899,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 3.8922581672668457,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18150087260035,
+ "grad_norm": 0.05714283883571625,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 3.849281072616577,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195462478184993,
+ "grad_norm": 0.05582018196582794,
+ "learning_rate": 0.000255486047794226,
+ "loss": 3.8659892082214355,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209424083769633,
+ "grad_norm": 0.05674750730395317,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 3.8591721057891846,
+ "step": 2031
+ },
+ {
+ "epoch": 28.223385689354277,
+ "grad_norm": 0.060421206057071686,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 3.8599629402160645,
+ "step": 2032
+ },
+ {
+ "epoch": 28.237347294938917,
+ "grad_norm": 0.05727452412247658,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 3.8559670448303223,
+ "step": 2033
+ },
+ {
+ "epoch": 28.25130890052356,
+ "grad_norm": 0.05708890035748482,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 3.8643712997436523,
+ "step": 2034
+ },
+ {
+ "epoch": 28.2652705061082,
+ "grad_norm": 0.05768970027565956,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 3.842012405395508,
+ "step": 2035
+ },
+ {
+ "epoch": 28.279232111692846,
+ "grad_norm": 0.05973518267273903,
+ "learning_rate": 0.000253907826333243,
+ "loss": 3.841464042663574,
+ "step": 2036
+ },
+ {
+ "epoch": 28.293193717277486,
+ "grad_norm": 0.057536788284778595,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 3.8655552864074707,
+ "step": 2037
+ },
+ {
+ "epoch": 28.30715532286213,
+ "grad_norm": 0.05864274501800537,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 3.852839469909668,
+ "step": 2038
+ },
+ {
+ "epoch": 28.32111692844677,
+ "grad_norm": 0.056395065039396286,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 3.859243392944336,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335078534031414,
+ "grad_norm": 0.05933082476258278,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 3.843926191329956,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349040139616054,
+ "grad_norm": 0.05862351134419441,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 3.8284354209899902,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363001745200698,
+ "grad_norm": 0.05564470961689949,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 3.8802285194396973,
+ "step": 2042
+ },
+ {
+ "epoch": 28.376963350785342,
+ "grad_norm": 0.06139146164059639,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 3.867100954055786,
+ "step": 2043
+ },
+ {
+ "epoch": 28.390924956369982,
+ "grad_norm": 0.058729205280542374,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 3.8322975635528564,
+ "step": 2044
+ },
+ {
+ "epoch": 28.404886561954626,
+ "grad_norm": 0.05771554261445999,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 3.855440616607666,
+ "step": 2045
+ },
+ {
+ "epoch": 28.418848167539267,
+ "grad_norm": 0.06000557541847229,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 3.8510828018188477,
+ "step": 2046
+ },
+ {
+ "epoch": 28.43280977312391,
+ "grad_norm": 0.05935550108551979,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 3.8572049140930176,
+ "step": 2047
+ },
+ {
+ "epoch": 28.44677137870855,
+ "grad_norm": 0.056745223701000214,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 3.8892972469329834,
+ "step": 2048
+ },
+ {
+ "epoch": 28.460732984293195,
+ "grad_norm": 0.0588800348341465,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 3.865957260131836,
+ "step": 2049
+ },
+ {
+ "epoch": 28.474694589877835,
+ "grad_norm": 0.05756748467683792,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 3.8530519008636475,
+ "step": 2050
+ },
+ {
+ "epoch": 28.48865619546248,
+ "grad_norm": 0.057398825883865356,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 3.9022018909454346,
+ "step": 2051
+ },
+ {
+ "epoch": 28.50261780104712,
+ "grad_norm": 0.05809267610311508,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 3.853647232055664,
+ "step": 2052
+ },
+ {
+ "epoch": 28.516579406631763,
+ "grad_norm": 0.05625335872173309,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 3.890363931655884,
+ "step": 2053
+ },
+ {
+ "epoch": 28.530541012216403,
+ "grad_norm": 0.05852413922548294,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 3.8836288452148438,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544502617801047,
+ "grad_norm": 0.05545232817530632,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 3.902006149291992,
+ "step": 2055
+ },
+ {
+ "epoch": 28.55846422338569,
+ "grad_norm": 0.05776599794626236,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 3.8910727500915527,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57242582897033,
+ "grad_norm": 0.05847984179854393,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 3.890106201171875,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586387434554975,
+ "grad_norm": 0.056724466383457184,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 3.836899995803833,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600349040139616,
+ "grad_norm": 0.05690842866897583,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 3.834292411804199,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61431064572426,
+ "grad_norm": 0.058501504361629486,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 3.891890048980713,
+ "step": 2060
+ },
+ {
+ "epoch": 28.6282722513089,
+ "grad_norm": 0.059110380709171295,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 3.883753538131714,
+ "step": 2061
+ },
+ {
+ "epoch": 28.642233856893544,
+ "grad_norm": 0.0574255995452404,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 3.8829569816589355,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656195462478184,
+ "grad_norm": 0.05842957645654678,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 3.845698118209839,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670157068062828,
+ "grad_norm": 0.05536387115716934,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 3.9008209705352783,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68411867364747,
+ "grad_norm": 0.05769932642579079,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 3.890972137451172,
+ "step": 2065
+ },
+ {
+ "epoch": 28.698080279232112,
+ "grad_norm": 0.058036014437675476,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 3.873745918273926,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712041884816752,
+ "grad_norm": 0.05829216167330742,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 3.879685401916504,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726003490401396,
+ "grad_norm": 0.06169239804148674,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 3.9156579971313477,
+ "step": 2068
+ },
+ {
+ "epoch": 28.739965095986037,
+ "grad_norm": 0.05861479043960571,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 3.876373052597046,
+ "step": 2069
+ },
+ {
+ "epoch": 28.75392670157068,
+ "grad_norm": 0.05878114700317383,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 3.918684959411621,
+ "step": 2070
+ },
+ {
+ "epoch": 28.767888307155324,
+ "grad_norm": 0.059904664754867554,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 3.8893160820007324,
+ "step": 2071
+ },
+ {
+ "epoch": 28.781849912739965,
+ "grad_norm": 0.057385317981243134,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 3.880828619003296,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79581151832461,
+ "grad_norm": 0.05815180763602257,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 3.8638038635253906,
+ "step": 2073
+ },
+ {
+ "epoch": 28.80977312390925,
+ "grad_norm": 0.05871085450053215,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 3.8427436351776123,
+ "step": 2074
+ },
+ {
+ "epoch": 28.823734729493893,
+ "grad_norm": 0.056275319308042526,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 3.858442544937134,
+ "step": 2075
+ },
+ {
+ "epoch": 28.837696335078533,
+ "grad_norm": 0.05850474536418915,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 3.871570110321045,
+ "step": 2076
+ },
+ {
+ "epoch": 28.851657940663177,
+ "grad_norm": 0.0580497644841671,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 3.89687442779541,
+ "step": 2077
+ },
+ {
+ "epoch": 28.865619546247817,
+ "grad_norm": 0.05944157764315605,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 3.8877720832824707,
+ "step": 2078
+ },
+ {
+ "epoch": 28.87958115183246,
+ "grad_norm": 0.05749465152621269,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 3.8787431716918945,
+ "step": 2079
+ },
+ {
+ "epoch": 28.8935427574171,
+ "grad_norm": 0.05645301938056946,
+ "learning_rate": 0.000242380656502223,
+ "loss": 3.9244091510772705,
+ "step": 2080
+ },
+ {
+ "epoch": 28.907504363001745,
+ "grad_norm": 0.05700884759426117,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 3.9192817211151123,
+ "step": 2081
+ },
+ {
+ "epoch": 28.921465968586386,
+ "grad_norm": 0.05823809653520584,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 3.905806064605713,
+ "step": 2082
+ },
+ {
+ "epoch": 28.93542757417103,
+ "grad_norm": 0.05736592411994934,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 3.8907763957977295,
+ "step": 2083
+ },
+ {
+ "epoch": 28.949389179755673,
+ "grad_norm": 0.05646584555506706,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 3.892714500427246,
+ "step": 2084
+ },
+ {
+ "epoch": 28.963350785340314,
+ "grad_norm": 0.06081141158938408,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 3.864286184310913,
+ "step": 2085
+ },
+ {
+ "epoch": 28.977312390924958,
+ "grad_norm": 0.05633515864610672,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 3.8791451454162598,
+ "step": 2086
+ },
+ {
+ "epoch": 28.991273996509598,
+ "grad_norm": 0.05825566127896309,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 3.8895483016967773,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.042954832315444946,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 2.408613681793213,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 0.6065697073936462,
+ "eval_runtime": 59.1276,
+ "eval_samples_per_second": 41.301,
+ "eval_steps_per_second": 0.66,
+ "step": 2088
+ },
+ {
+ "epoch": 29.013961605584644,
+ "grad_norm": 0.06000809744000435,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 3.8419973850250244,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027923211169284,
+ "grad_norm": 0.06233612447977066,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 3.83025860786438,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041884816753928,
+ "grad_norm": 0.0551699623465538,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 3.8283262252807617,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05584642233857,
+ "grad_norm": 0.06225701421499252,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 3.822310447692871,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069808027923212,
+ "grad_norm": 0.05747336149215698,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 3.8208513259887695,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083769633507853,
+ "grad_norm": 0.0582275427877903,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 3.8013901710510254,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097731239092496,
+ "grad_norm": 0.059701140969991684,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 3.829238176345825,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111692844677137,
+ "grad_norm": 0.057550523430109024,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 3.7963826656341553,
+ "step": 2096
+ },
+ {
+ "epoch": 29.12565445026178,
+ "grad_norm": 0.05855116620659828,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 3.824028491973877,
+ "step": 2097
+ },
+ {
+ "epoch": 29.13961605584642,
+ "grad_norm": 0.056533779948949814,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 3.8151779174804688,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153577661431065,
+ "grad_norm": 0.0573289655148983,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 3.7990314960479736,
+ "step": 2099
+ },
+ {
+ "epoch": 29.167539267015705,
+ "grad_norm": 0.060061171650886536,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 3.8221707344055176,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18150087260035,
+ "grad_norm": 0.05851743370294571,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 3.793701648712158,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195462478184993,
+ "grad_norm": 0.059191569685935974,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 3.8194010257720947,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209424083769633,
+ "grad_norm": 0.05996483936905861,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 3.8316292762756348,
+ "step": 2103
+ },
+ {
+ "epoch": 29.223385689354277,
+ "grad_norm": 0.0555657260119915,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 3.821362018585205,
+ "step": 2104
+ },
+ {
+ "epoch": 29.237347294938917,
+ "grad_norm": 0.058476634323596954,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 3.8098394870758057,
+ "step": 2105
+ },
+ {
+ "epoch": 29.25130890052356,
+ "grad_norm": 0.05669668689370155,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 3.843484878540039,
+ "step": 2106
+ },
+ {
+ "epoch": 29.2652705061082,
+ "grad_norm": 0.05648602545261383,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 3.811458110809326,
+ "step": 2107
+ },
+ {
+ "epoch": 29.279232111692846,
+ "grad_norm": 0.05574130266904831,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 3.837676525115967,
+ "step": 2108
+ },
+ {
+ "epoch": 29.293193717277486,
+ "grad_norm": 0.05538785457611084,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 3.810798168182373,
+ "step": 2109
+ },
+ {
+ "epoch": 29.30715532286213,
+ "grad_norm": 0.05477031692862511,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 3.8156371116638184,
+ "step": 2110
+ },
+ {
+ "epoch": 29.32111692844677,
+ "grad_norm": 0.05664678290486336,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 3.8472559452056885,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335078534031414,
+ "grad_norm": 0.05649891495704651,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 3.824376106262207,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349040139616054,
+ "grad_norm": 0.05665554851293564,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 3.7963147163391113,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363001745200698,
+ "grad_norm": 0.054111722856760025,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 3.819735527038574,
+ "step": 2114
+ },
+ {
+ "epoch": 29.376963350785342,
+ "grad_norm": 0.05805648863315582,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 3.790189743041992,
+ "step": 2115
+ },
+ {
+ "epoch": 29.390924956369982,
+ "grad_norm": 0.05634011700749397,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 3.852161169052124,
+ "step": 2116
+ },
+ {
+ "epoch": 29.404886561954626,
+ "grad_norm": 0.0555867962539196,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 3.846884250640869,
+ "step": 2117
+ },
+ {
+ "epoch": 29.418848167539267,
+ "grad_norm": 0.057229746133089066,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 3.8432483673095703,
+ "step": 2118
+ },
+ {
+ "epoch": 29.43280977312391,
+ "grad_norm": 0.05664287880063057,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 3.842709541320801,
+ "step": 2119
+ },
+ {
+ "epoch": 29.44677137870855,
+ "grad_norm": 0.0567023791372776,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 3.844529628753662,
+ "step": 2120
+ },
+ {
+ "epoch": 29.460732984293195,
+ "grad_norm": 0.057870302349328995,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 3.853799343109131,
+ "step": 2121
+ },
+ {
+ "epoch": 29.474694589877835,
+ "grad_norm": 0.05681667849421501,
+ "learning_rate": 0.000231465389734324,
+ "loss": 3.849405288696289,
+ "step": 2122
+ },
+ {
+ "epoch": 29.48865619546248,
+ "grad_norm": 0.05738653615117073,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 3.8083410263061523,
+ "step": 2123
+ },
+ {
+ "epoch": 29.50261780104712,
+ "grad_norm": 0.058900732547044754,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 3.840458869934082,
+ "step": 2124
+ },
+ {
+ "epoch": 29.516579406631763,
+ "grad_norm": 0.05823393166065216,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 3.821415662765503,
+ "step": 2125
+ },
+ {
+ "epoch": 29.530541012216403,
+ "grad_norm": 0.057413019239902496,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 3.8713812828063965,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544502617801047,
+ "grad_norm": 0.05935502052307129,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 3.8488903045654297,
+ "step": 2127
+ },
+ {
+ "epoch": 29.55846422338569,
+ "grad_norm": 0.055285241454839706,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 3.8196067810058594,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57242582897033,
+ "grad_norm": 0.05894080176949501,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 3.8444557189941406,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586387434554975,
+ "grad_norm": 0.05878668278455734,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 3.8731908798217773,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600349040139616,
+ "grad_norm": 0.05855663865804672,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 3.831028938293457,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61431064572426,
+ "grad_norm": 0.057419586926698685,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 3.849435806274414,
+ "step": 2132
+ },
+ {
+ "epoch": 29.6282722513089,
+ "grad_norm": 0.058863408863544464,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 3.8495519161224365,
+ "step": 2133
+ },
+ {
+ "epoch": 29.642233856893544,
+ "grad_norm": 0.06081194058060646,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 3.844421863555908,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656195462478184,
+ "grad_norm": 0.05597590655088425,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 3.8413243293762207,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670157068062828,
+ "grad_norm": 0.058956339955329895,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 3.861703872680664,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68411867364747,
+ "grad_norm": 0.05883771926164627,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 3.8756394386291504,
+ "step": 2137
+ },
+ {
+ "epoch": 29.698080279232112,
+ "grad_norm": 0.06036122143268585,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 3.855403423309326,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712041884816752,
+ "grad_norm": 0.061243657022714615,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 3.861691474914551,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726003490401396,
+ "grad_norm": 0.05917691811919212,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 3.870516777038574,
+ "step": 2140
+ },
+ {
+ "epoch": 29.739965095986037,
+ "grad_norm": 0.06157418712973595,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 3.8292267322540283,
+ "step": 2141
+ },
+ {
+ "epoch": 29.75392670157068,
+ "grad_norm": 0.0605802945792675,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 3.836577892303467,
+ "step": 2142
+ },
+ {
+ "epoch": 29.767888307155324,
+ "grad_norm": 0.05750525742769241,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 3.8472204208374023,
+ "step": 2143
+ },
+ {
+ "epoch": 29.781849912739965,
+ "grad_norm": 0.057982541620731354,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 3.8682217597961426,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79581151832461,
+ "grad_norm": 0.057891640812158585,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 3.8397727012634277,
+ "step": 2145
+ },
+ {
+ "epoch": 29.80977312390925,
+ "grad_norm": 0.058175452053546906,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 3.8578720092773438,
+ "step": 2146
+ },
+ {
+ "epoch": 29.823734729493893,
+ "grad_norm": 0.05686056986451149,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 3.8467812538146973,
+ "step": 2147
+ },
+ {
+ "epoch": 29.837696335078533,
+ "grad_norm": 0.05755440890789032,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 3.844052314758301,
+ "step": 2148
+ },
+ {
+ "epoch": 29.851657940663177,
+ "grad_norm": 0.057706039398908615,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 3.818631172180176,
+ "step": 2149
+ },
+ {
+ "epoch": 29.865619546247817,
+ "grad_norm": 0.058824241161346436,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 3.8732900619506836,
+ "step": 2150
+ },
+ {
+ "epoch": 29.87958115183246,
+ "grad_norm": 0.058583300560712814,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 3.8373632431030273,
+ "step": 2151
+ },
+ {
+ "epoch": 29.8935427574171,
+ "grad_norm": 0.057403020560741425,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 3.8761391639709473,
+ "step": 2152
+ },
+ {
+ "epoch": 29.907504363001745,
+ "grad_norm": 0.0586736761033535,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 3.8518989086151123,
+ "step": 2153
+ },
+ {
+ "epoch": 29.921465968586386,
+ "grad_norm": 0.05727249011397362,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 3.839515209197998,
+ "step": 2154
+ },
+ {
+ "epoch": 29.93542757417103,
+ "grad_norm": 0.0594087690114975,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 3.834868907928467,
+ "step": 2155
+ },
+ {
+ "epoch": 29.949389179755673,
+ "grad_norm": 0.057742297649383545,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 3.861978530883789,
+ "step": 2156
+ },
+ {
+ "epoch": 29.963350785340314,
+ "grad_norm": 0.05707386136054993,
+ "learning_rate": 0.00022244633283095,
+ "loss": 3.886831760406494,
+ "step": 2157
+ },
+ {
+ "epoch": 29.977312390924958,
+ "grad_norm": 0.05631854757666588,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 3.83940052986145,
+ "step": 2158
+ },
+ {
+ "epoch": 29.991273996509598,
+ "grad_norm": 0.057860348373651505,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 3.8477416038513184,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.04080710560083389,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 2.4169907569885254,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 0.6078308820724487,
+ "eval_runtime": 59.4599,
+ "eval_samples_per_second": 41.07,
+ "eval_steps_per_second": 0.656,
+ "step": 2160
+ },
+ {
+ "epoch": 30.013961605584644,
+ "grad_norm": 0.05596645176410675,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 3.798748016357422,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027923211169284,
+ "grad_norm": 0.054763782769441605,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 3.7787699699401855,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041884816753928,
+ "grad_norm": 0.053927499800920486,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 3.779787063598633,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05584642233857,
+ "grad_norm": 0.05513143539428711,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 3.782046318054199,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069808027923212,
+ "grad_norm": 0.055234167724847794,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 3.8064064979553223,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083769633507853,
+ "grad_norm": 0.05320988595485687,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 3.8084359169006348,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097731239092496,
+ "grad_norm": 0.05671925097703934,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 3.7805888652801514,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111692844677137,
+ "grad_norm": 0.05502697452902794,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 3.791616916656494,
+ "step": 2168
+ },
+ {
+ "epoch": 30.12565445026178,
+ "grad_norm": 0.055426280945539474,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 3.7947494983673096,
+ "step": 2169
+ },
+ {
+ "epoch": 30.13961605584642,
+ "grad_norm": 0.05564495921134949,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 3.7958974838256836,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153577661431065,
+ "grad_norm": 0.055089402943849564,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 3.783299446105957,
+ "step": 2171
+ },
+ {
+ "epoch": 30.167539267015705,
+ "grad_norm": 0.055240437388420105,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 3.769829750061035,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18150087260035,
+ "grad_norm": 0.055886462330818176,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 3.805971622467041,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195462478184993,
+ "grad_norm": 0.05560130998492241,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 3.801429271697998,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209424083769633,
+ "grad_norm": 0.054780587553977966,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 3.794417381286621,
+ "step": 2175
+ },
+ {
+ "epoch": 30.223385689354277,
+ "grad_norm": 0.05751654878258705,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 3.7895469665527344,
+ "step": 2176
+ },
+ {
+ "epoch": 30.237347294938917,
+ "grad_norm": 0.054392315447330475,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 3.786479949951172,
+ "step": 2177
+ },
+ {
+ "epoch": 30.25130890052356,
+ "grad_norm": 0.057609833776950836,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 3.809680700302124,
+ "step": 2178
+ },
+ {
+ "epoch": 30.2652705061082,
+ "grad_norm": 0.054882537573575974,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 3.7928991317749023,
+ "step": 2179
+ },
+ {
+ "epoch": 30.279232111692846,
+ "grad_norm": 0.056957781314849854,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 3.790977716445923,
+ "step": 2180
+ },
+ {
+ "epoch": 30.293193717277486,
+ "grad_norm": 0.0580085888504982,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 3.851022958755493,
+ "step": 2181
+ },
+ {
+ "epoch": 30.30715532286213,
+ "grad_norm": 0.055905163288116455,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 3.784822463989258,
+ "step": 2182
+ },
+ {
+ "epoch": 30.32111692844677,
+ "grad_norm": 0.05998636782169342,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 3.811324119567871,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335078534031414,
+ "grad_norm": 0.057061366736888885,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 3.8131840229034424,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349040139616054,
+ "grad_norm": 0.05680330842733383,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 3.806260824203491,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363001745200698,
+ "grad_norm": 0.05539876967668533,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 3.793149471282959,
+ "step": 2186
+ },
+ {
+ "epoch": 30.376963350785342,
+ "grad_norm": 0.05771343782544136,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 3.7772932052612305,
+ "step": 2187
+ },
+ {
+ "epoch": 30.390924956369982,
+ "grad_norm": 0.055599041283130646,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 3.7704858779907227,
+ "step": 2188
+ },
+ {
+ "epoch": 30.404886561954626,
+ "grad_norm": 0.05746382102370262,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 3.8296337127685547,
+ "step": 2189
+ },
+ {
+ "epoch": 30.418848167539267,
+ "grad_norm": 0.05619506537914276,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 3.8134541511535645,
+ "step": 2190
+ },
+ {
+ "epoch": 30.43280977312391,
+ "grad_norm": 0.05699177086353302,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 3.8070058822631836,
+ "step": 2191
+ },
+ {
+ "epoch": 30.44677137870855,
+ "grad_norm": 0.056783925741910934,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 3.7931368350982666,
+ "step": 2192
+ },
+ {
+ "epoch": 30.460732984293195,
+ "grad_norm": 0.055537473410367966,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 3.809480905532837,
+ "step": 2193
+ },
+ {
+ "epoch": 30.474694589877835,
+ "grad_norm": 0.057826802134513855,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 3.8422718048095703,
+ "step": 2194
+ },
+ {
+ "epoch": 30.48865619546248,
+ "grad_norm": 0.0562913753092289,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 3.803966760635376,
+ "step": 2195
+ },
+ {
+ "epoch": 30.50261780104712,
+ "grad_norm": 0.05720103159546852,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 3.767852306365967,
+ "step": 2196
+ },
+ {
+ "epoch": 30.516579406631763,
+ "grad_norm": 0.05813048407435417,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 3.8423361778259277,
+ "step": 2197
+ },
+ {
+ "epoch": 30.530541012216403,
+ "grad_norm": 0.057173099368810654,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 3.8059489727020264,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544502617801047,
+ "grad_norm": 0.059024229645729065,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 3.7705254554748535,
+ "step": 2199
+ },
+ {
+ "epoch": 30.55846422338569,
+ "grad_norm": 0.056807395070791245,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 3.8056063652038574,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57242582897033,
+ "grad_norm": 0.05868417024612427,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 3.829719066619873,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586387434554975,
+ "grad_norm": 0.06023262441158295,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 3.8277781009674072,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600349040139616,
+ "grad_norm": 0.057253073900938034,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 3.8088555335998535,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61431064572426,
+ "grad_norm": 0.05946561321616173,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 3.7907581329345703,
+ "step": 2204
+ },
+ {
+ "epoch": 30.6282722513089,
+ "grad_norm": 0.05909013748168945,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 3.812042236328125,
+ "step": 2205
+ },
+ {
+ "epoch": 30.642233856893544,
+ "grad_norm": 0.05843796953558922,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 3.814551830291748,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656195462478184,
+ "grad_norm": 0.05953318253159523,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 3.8058853149414062,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670157068062828,
+ "grad_norm": 0.0616983063519001,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 3.7986061573028564,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68411867364747,
+ "grad_norm": 0.0588211789727211,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 3.8080081939697266,
+ "step": 2209
+ },
+ {
+ "epoch": 30.698080279232112,
+ "grad_norm": 0.05937618389725685,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 3.7851452827453613,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712041884816752,
+ "grad_norm": 0.05869529768824577,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 3.7893905639648438,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726003490401396,
+ "grad_norm": 0.060255810618400574,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 3.7998719215393066,
+ "step": 2212
+ },
+ {
+ "epoch": 30.739965095986037,
+ "grad_norm": 0.05810222774744034,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 3.800131320953369,
+ "step": 2213
+ },
+ {
+ "epoch": 30.75392670157068,
+ "grad_norm": 0.057548508048057556,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 3.7940425872802734,
+ "step": 2214
+ },
+ {
+ "epoch": 30.767888307155324,
+ "grad_norm": 0.05792916938662529,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 3.8089685440063477,
+ "step": 2215
+ },
+ {
+ "epoch": 30.781849912739965,
+ "grad_norm": 0.06035939231514931,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 3.8355417251586914,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79581151832461,
+ "grad_norm": 0.058805160224437714,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 3.8146378993988037,
+ "step": 2217
+ },
+ {
+ "epoch": 30.80977312390925,
+ "grad_norm": 0.05828496813774109,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 3.8210601806640625,
+ "step": 2218
+ },
+ {
+ "epoch": 30.823734729493893,
+ "grad_norm": 0.059257738292217255,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 3.807644844055176,
+ "step": 2219
+ },
+ {
+ "epoch": 30.837696335078533,
+ "grad_norm": 0.05937729775905609,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 3.8312151432037354,
+ "step": 2220
+ },
+ {
+ "epoch": 30.851657940663177,
+ "grad_norm": 0.058580249547958374,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 3.7979660034179688,
+ "step": 2221
+ },
+ {
+ "epoch": 30.865619546247817,
+ "grad_norm": 0.0585496611893177,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 3.827874183654785,
+ "step": 2222
+ },
+ {
+ "epoch": 30.87958115183246,
+ "grad_norm": 0.05949759483337402,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 3.808143138885498,
+ "step": 2223
+ },
+ {
+ "epoch": 30.8935427574171,
+ "grad_norm": 0.058828581124544144,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 3.8351407051086426,
+ "step": 2224
+ },
+ {
+ "epoch": 30.907504363001745,
+ "grad_norm": 0.0600498728454113,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 3.8025779724121094,
+ "step": 2225
+ },
+ {
+ "epoch": 30.921465968586386,
+ "grad_norm": 0.06004577502608299,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 3.8157856464385986,
+ "step": 2226
+ },
+ {
+ "epoch": 30.93542757417103,
+ "grad_norm": 0.057676251977682114,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 3.8136918544769287,
+ "step": 2227
+ },
+ {
+ "epoch": 30.949389179755673,
+ "grad_norm": 0.0627770945429802,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 3.8470370769500732,
+ "step": 2228
+ },
+ {
+ "epoch": 30.963350785340314,
+ "grad_norm": 0.058836981654167175,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 3.8422586917877197,
+ "step": 2229
+ },
+ {
+ "epoch": 30.977312390924958,
+ "grad_norm": 0.060720525681972504,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 3.7992711067199707,
+ "step": 2230
+ },
+ {
+ "epoch": 30.991273996509598,
+ "grad_norm": 0.05844535678625107,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 3.823277473449707,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.0424773246049881,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 2.3872625827789307,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 0.6090936064720154,
+ "eval_runtime": 58.7367,
+ "eval_samples_per_second": 41.575,
+ "eval_steps_per_second": 0.664,
+ "step": 2232
+ },
+ {
+ "epoch": 31.013961605584644,
+ "grad_norm": 0.057239554822444916,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 3.764132022857666,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027923211169284,
+ "grad_norm": 0.05666050314903259,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 3.767040729522705,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041884816753928,
+ "grad_norm": 0.054038409143686295,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 3.7878732681274414,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05584642233857,
+ "grad_norm": 0.056939832866191864,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 3.7827179431915283,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069808027923212,
+ "grad_norm": 0.05488067492842674,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 3.7544898986816406,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083769633507853,
+ "grad_norm": 0.05604229122400284,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 3.7706053256988525,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097731239092496,
+ "grad_norm": 0.0550883486866951,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 3.755401611328125,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111692844677137,
+ "grad_norm": 0.055959347635507584,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 3.741966724395752,
+ "step": 2240
+ },
+ {
+ "epoch": 31.12565445026178,
+ "grad_norm": 0.05733809620141983,
+ "learning_rate": 0.000201141724176723,
+ "loss": 3.7537970542907715,
+ "step": 2241
+ },
+ {
+ "epoch": 31.13961605584642,
+ "grad_norm": 0.0551145002245903,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 3.763530731201172,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153577661431065,
+ "grad_norm": 0.055499665439128876,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 3.747159481048584,
+ "step": 2243
+ },
+ {
+ "epoch": 31.167539267015705,
+ "grad_norm": 0.0553385354578495,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 3.769044876098633,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18150087260035,
+ "grad_norm": 0.05678795278072357,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 3.7528388500213623,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195462478184993,
+ "grad_norm": 0.055931974202394485,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 3.742429733276367,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209424083769633,
+ "grad_norm": 0.05553249269723892,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 3.7760162353515625,
+ "step": 2247
+ },
+ {
+ "epoch": 31.223385689354277,
+ "grad_norm": 0.05742773041129112,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 3.7682301998138428,
+ "step": 2248
+ },
+ {
+ "epoch": 31.237347294938917,
+ "grad_norm": 0.056315284222364426,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 3.755894660949707,
+ "step": 2249
+ },
+ {
+ "epoch": 31.25130890052356,
+ "grad_norm": 0.05695716291666031,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 3.733503580093384,
+ "step": 2250
+ },
+ {
+ "epoch": 31.2652705061082,
+ "grad_norm": 0.05714774876832962,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 3.7504918575286865,
+ "step": 2251
+ },
+ {
+ "epoch": 31.279232111692846,
+ "grad_norm": 0.058987151831388474,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 3.7579331398010254,
+ "step": 2252
+ },
+ {
+ "epoch": 31.293193717277486,
+ "grad_norm": 0.05853817239403725,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 3.781409740447998,
+ "step": 2253
+ },
+ {
+ "epoch": 31.30715532286213,
+ "grad_norm": 0.057014722377061844,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 3.763247013092041,
+ "step": 2254
+ },
+ {
+ "epoch": 31.32111692844677,
+ "grad_norm": 0.05816958099603653,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 3.7801408767700195,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335078534031414,
+ "grad_norm": 0.05724027752876282,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 3.7466397285461426,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349040139616054,
+ "grad_norm": 0.055836353451013565,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 3.7770347595214844,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363001745200698,
+ "grad_norm": 0.05831955373287201,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 3.7769415378570557,
+ "step": 2258
+ },
+ {
+ "epoch": 31.376963350785342,
+ "grad_norm": 0.056602735072374344,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 3.781635284423828,
+ "step": 2259
+ },
+ {
+ "epoch": 31.390924956369982,
+ "grad_norm": 0.05784839391708374,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 3.7605199813842773,
+ "step": 2260
+ },
+ {
+ "epoch": 31.404886561954626,
+ "grad_norm": 0.057923249900341034,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 3.7671380043029785,
+ "step": 2261
+ },
+ {
+ "epoch": 31.418848167539267,
+ "grad_norm": 0.05663329362869263,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 3.7952239513397217,
+ "step": 2262
+ },
+ {
+ "epoch": 31.43280977312391,
+ "grad_norm": 0.057000353932380676,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 3.7924318313598633,
+ "step": 2263
+ },
+ {
+ "epoch": 31.44677137870855,
+ "grad_norm": 0.059976726770401,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 3.7782416343688965,
+ "step": 2264
+ },
+ {
+ "epoch": 31.460732984293195,
+ "grad_norm": 0.05681120976805687,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 3.7918577194213867,
+ "step": 2265
+ },
+ {
+ "epoch": 31.474694589877835,
+ "grad_norm": 0.05824654549360275,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 3.7725377082824707,
+ "step": 2266
+ },
+ {
+ "epoch": 31.48865619546248,
+ "grad_norm": 0.058513302356004715,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 3.769562244415283,
+ "step": 2267
+ },
+ {
+ "epoch": 31.50261780104712,
+ "grad_norm": 0.05720348283648491,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 3.773402214050293,
+ "step": 2268
+ },
+ {
+ "epoch": 31.516579406631763,
+ "grad_norm": 0.06029265746474266,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 3.7562055587768555,
+ "step": 2269
+ },
+ {
+ "epoch": 31.530541012216403,
+ "grad_norm": 0.05892042815685272,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 3.779987335205078,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544502617801047,
+ "grad_norm": 0.06146431341767311,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 3.749899387359619,
+ "step": 2271
+ },
+ {
+ "epoch": 31.55846422338569,
+ "grad_norm": 0.056403227150440216,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 3.763558864593506,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57242582897033,
+ "grad_norm": 0.058624617755413055,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 3.7811503410339355,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586387434554975,
+ "grad_norm": 0.05978401005268097,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 3.7820377349853516,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600349040139616,
+ "grad_norm": 0.05971162021160126,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 3.7685751914978027,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61431064572426,
+ "grad_norm": 0.05576494708657265,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 3.7588582038879395,
+ "step": 2276
+ },
+ {
+ "epoch": 31.6282722513089,
+ "grad_norm": 0.05953748896718025,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 3.7525811195373535,
+ "step": 2277
+ },
+ {
+ "epoch": 31.642233856893544,
+ "grad_norm": 0.05771994963288307,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 3.7891383171081543,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656195462478184,
+ "grad_norm": 0.05822686105966568,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 3.7933766841888428,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670157068062828,
+ "grad_norm": 0.06044064089655876,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 3.783848762512207,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68411867364747,
+ "grad_norm": 0.05953317880630493,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 3.775938034057617,
+ "step": 2281
+ },
+ {
+ "epoch": 31.698080279232112,
+ "grad_norm": 0.05928120017051697,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 3.7867183685302734,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712041884816752,
+ "grad_norm": 0.05855962261557579,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 3.7872838973999023,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726003490401396,
+ "grad_norm": 0.058542776852846146,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 3.8022043704986572,
+ "step": 2284
+ },
+ {
+ "epoch": 31.739965095986037,
+ "grad_norm": 0.059078946709632874,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 3.79337739944458,
+ "step": 2285
+ },
+ {
+ "epoch": 31.75392670157068,
+ "grad_norm": 0.05854807421565056,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 3.769434928894043,
+ "step": 2286
+ },
+ {
+ "epoch": 31.767888307155324,
+ "grad_norm": 0.058214519172906876,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 3.778744697570801,
+ "step": 2287
+ },
+ {
+ "epoch": 31.781849912739965,
+ "grad_norm": 0.05981272831559181,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 3.8015875816345215,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79581151832461,
+ "grad_norm": 0.05833584442734718,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 3.7718634605407715,
+ "step": 2289
+ },
+ {
+ "epoch": 31.80977312390925,
+ "grad_norm": 0.058713216334581375,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 3.759026050567627,
+ "step": 2290
+ },
+ {
+ "epoch": 31.823734729493893,
+ "grad_norm": 0.06085464730858803,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 3.762263059616089,
+ "step": 2291
+ },
+ {
+ "epoch": 31.837696335078533,
+ "grad_norm": 0.05838872492313385,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 3.7770943641662598,
+ "step": 2292
+ },
+ {
+ "epoch": 31.851657940663177,
+ "grad_norm": 0.05941537022590637,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 3.8041505813598633,
+ "step": 2293
+ },
+ {
+ "epoch": 31.865619546247817,
+ "grad_norm": 0.058818284422159195,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 3.7724623680114746,
+ "step": 2294
+ },
+ {
+ "epoch": 31.87958115183246,
+ "grad_norm": 0.05925480276346207,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 3.7881040573120117,
+ "step": 2295
+ },
+ {
+ "epoch": 31.8935427574171,
+ "grad_norm": 0.058119483292102814,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 3.7992963790893555,
+ "step": 2296
+ },
+ {
+ "epoch": 31.907504363001745,
+ "grad_norm": 0.06139586120843887,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 3.800816297531128,
+ "step": 2297
+ },
+ {
+ "epoch": 31.921465968586386,
+ "grad_norm": 0.05812571942806244,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 3.7683157920837402,
+ "step": 2298
+ },
+ {
+ "epoch": 31.93542757417103,
+ "grad_norm": 0.05929575115442276,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 3.808281660079956,
+ "step": 2299
+ },
+ {
+ "epoch": 31.949389179755673,
+ "grad_norm": 0.057833533734083176,
+ "learning_rate": 0.000186516746349841,
+ "loss": 3.821108818054199,
+ "step": 2300
+ },
+ {
+ "epoch": 31.963350785340314,
+ "grad_norm": 0.05857503414154053,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 3.8082172870635986,
+ "step": 2301
+ },
+ {
+ "epoch": 31.977312390924958,
+ "grad_norm": 0.05893079936504364,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 3.802304267883301,
+ "step": 2302
+ },
+ {
+ "epoch": 31.991273996509598,
+ "grad_norm": 0.057888105511665344,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 3.7589073181152344,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.04306843504309654,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 2.359179735183716,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 0.6101394891738892,
+ "eval_runtime": 59.2767,
+ "eval_samples_per_second": 41.197,
+ "eval_steps_per_second": 0.658,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01396160558464,
+ "grad_norm": 0.059457872062921524,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 3.746081829071045,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02792321116929,
+ "grad_norm": 0.05826259404420853,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 3.7450389862060547,
+ "step": 2306
+ },
+ {
+ "epoch": 32.04188481675393,
+ "grad_norm": 0.05683758482336998,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 3.718315601348877,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05584642233857,
+ "grad_norm": 0.05709684267640114,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 3.7391161918640137,
+ "step": 2308
+ },
+ {
+ "epoch": 32.06980802792321,
+ "grad_norm": 0.0563763789832592,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 3.714268684387207,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083769633507856,
+ "grad_norm": 0.058482736349105835,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 3.7212839126586914,
+ "step": 2310
+ },
+ {
+ "epoch": 32.0977312390925,
+ "grad_norm": 0.05536908283829689,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 3.7573728561401367,
+ "step": 2311
+ },
+ {
+ "epoch": 32.11169284467714,
+ "grad_norm": 0.05950985103845596,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 3.715609312057495,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12565445026178,
+ "grad_norm": 0.05600661039352417,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 3.7144830226898193,
+ "step": 2313
+ },
+ {
+ "epoch": 32.139616055846425,
+ "grad_norm": 0.05814438313245773,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 3.7489242553710938,
+ "step": 2314
+ },
+ {
+ "epoch": 32.153577661431065,
+ "grad_norm": 0.0568280853331089,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 3.729698657989502,
+ "step": 2315
+ },
+ {
+ "epoch": 32.167539267015705,
+ "grad_norm": 0.05647209286689758,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 3.7484517097473145,
+ "step": 2316
+ },
+ {
+ "epoch": 32.181500872600346,
+ "grad_norm": 0.05811113119125366,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 3.7534449100494385,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19546247818499,
+ "grad_norm": 0.055132538080215454,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 3.695220470428467,
+ "step": 2318
+ },
+ {
+ "epoch": 32.20942408376963,
+ "grad_norm": 0.05722372978925705,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 3.7204041481018066,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223385689354274,
+ "grad_norm": 0.05533589795231819,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 3.7357606887817383,
+ "step": 2320
+ },
+ {
+ "epoch": 32.23734729493892,
+ "grad_norm": 0.056150201708078384,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 3.742290735244751,
+ "step": 2321
+ },
+ {
+ "epoch": 32.25130890052356,
+ "grad_norm": 0.05737832188606262,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 3.7484631538391113,
+ "step": 2322
+ },
+ {
+ "epoch": 32.2652705061082,
+ "grad_norm": 0.056017421185970306,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 3.744821071624756,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27923211169284,
+ "grad_norm": 0.05658293142914772,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 3.717287063598633,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29319371727749,
+ "grad_norm": 0.05632270127534866,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 3.7172422409057617,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30715532286213,
+ "grad_norm": 0.05777734890580177,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 3.7551779747009277,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32111692844677,
+ "grad_norm": 0.05593368038535118,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 3.7148218154907227,
+ "step": 2327
+ },
+ {
+ "epoch": 32.33507853403141,
+ "grad_norm": 0.05844123661518097,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 3.7418324947357178,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34904013961606,
+ "grad_norm": 0.05616185814142227,
+ "learning_rate": 0.000179445406945268,
+ "loss": 3.7213454246520996,
+ "step": 2329
+ },
+ {
+ "epoch": 32.3630017452007,
+ "grad_norm": 0.058453164994716644,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 3.730037212371826,
+ "step": 2330
+ },
+ {
+ "epoch": 32.37696335078534,
+ "grad_norm": 0.05819960683584213,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 3.7571628093719482,
+ "step": 2331
+ },
+ {
+ "epoch": 32.390924956369986,
+ "grad_norm": 0.05889322608709335,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 3.7306394577026367,
+ "step": 2332
+ },
+ {
+ "epoch": 32.404886561954626,
+ "grad_norm": 0.05727435275912285,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 3.7304887771606445,
+ "step": 2333
+ },
+ {
+ "epoch": 32.41884816753927,
+ "grad_norm": 0.058260343968868256,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 3.714399814605713,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43280977312391,
+ "grad_norm": 0.06092117354273796,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 3.743901491165161,
+ "step": 2335
+ },
+ {
+ "epoch": 32.446771378708554,
+ "grad_norm": 0.05715249851346016,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 3.747819423675537,
+ "step": 2336
+ },
+ {
+ "epoch": 32.460732984293195,
+ "grad_norm": 0.060237716883420944,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 3.770688533782959,
+ "step": 2337
+ },
+ {
+ "epoch": 32.474694589877835,
+ "grad_norm": 0.059161923825740814,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 3.724484443664551,
+ "step": 2338
+ },
+ {
+ "epoch": 32.488656195462475,
+ "grad_norm": 0.05693944916129112,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 3.7228128910064697,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50261780104712,
+ "grad_norm": 0.05745118111371994,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 3.7756295204162598,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51657940663176,
+ "grad_norm": 0.05782153829932213,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 3.757032871246338,
+ "step": 2341
+ },
+ {
+ "epoch": 32.5305410122164,
+ "grad_norm": 0.058188699185848236,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 3.7462215423583984,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544502617801044,
+ "grad_norm": 0.05856512114405632,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 3.748447895050049,
+ "step": 2343
+ },
+ {
+ "epoch": 32.55846422338569,
+ "grad_norm": 0.05581355467438698,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 3.726031541824341,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57242582897033,
+ "grad_norm": 0.05866248160600662,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 3.7235984802246094,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58638743455497,
+ "grad_norm": 0.05732253938913345,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 3.7372875213623047,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60034904013962,
+ "grad_norm": 0.057102739810943604,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 3.768643617630005,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61431064572426,
+ "grad_norm": 0.05914342775940895,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 3.748710870742798,
+ "step": 2348
+ },
+ {
+ "epoch": 32.6282722513089,
+ "grad_norm": 0.05618954077363014,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 3.7640533447265625,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64223385689354,
+ "grad_norm": 0.05810703709721565,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 3.7230985164642334,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65619546247819,
+ "grad_norm": 0.05668202042579651,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 3.736083984375,
+ "step": 2351
+ },
+ {
+ "epoch": 32.67015706806283,
+ "grad_norm": 0.05779054015874863,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 3.7652668952941895,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68411867364747,
+ "grad_norm": 0.05945349857211113,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 3.75258207321167,
+ "step": 2353
+ },
+ {
+ "epoch": 32.69808027923211,
+ "grad_norm": 0.05697475001215935,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 3.7300591468811035,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712041884816756,
+ "grad_norm": 0.05820303037762642,
+ "learning_rate": 0.000173176617304673,
+ "loss": 3.7426252365112305,
+ "step": 2355
+ },
+ {
+ "epoch": 32.726003490401396,
+ "grad_norm": 0.05830353870987892,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 3.7716498374938965,
+ "step": 2356
+ },
+ {
+ "epoch": 32.73996509598604,
+ "grad_norm": 0.05690579116344452,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 3.7565040588378906,
+ "step": 2357
+ },
+ {
+ "epoch": 32.753926701570684,
+ "grad_norm": 0.059615254402160645,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 3.7484030723571777,
+ "step": 2358
+ },
+ {
+ "epoch": 32.767888307155324,
+ "grad_norm": 0.058043934404850006,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 3.76991868019104,
+ "step": 2359
+ },
+ {
+ "epoch": 32.781849912739965,
+ "grad_norm": 0.059192877262830734,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 3.7510385513305664,
+ "step": 2360
+ },
+ {
+ "epoch": 32.795811518324605,
+ "grad_norm": 0.05763391777873039,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 3.7479896545410156,
+ "step": 2361
+ },
+ {
+ "epoch": 32.80977312390925,
+ "grad_norm": 0.05946139991283417,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 3.7780566215515137,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82373472949389,
+ "grad_norm": 0.05837136134505272,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 3.7610299587249756,
+ "step": 2363
+ },
+ {
+ "epoch": 32.83769633507853,
+ "grad_norm": 0.05832294747233391,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 3.7216475009918213,
+ "step": 2364
+ },
+ {
+ "epoch": 32.85165794066317,
+ "grad_norm": 0.05887595936655998,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 3.767125129699707,
+ "step": 2365
+ },
+ {
+ "epoch": 32.86561954624782,
+ "grad_norm": 0.058594878762960434,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 3.751053810119629,
+ "step": 2366
+ },
+ {
+ "epoch": 32.87958115183246,
+ "grad_norm": 0.05895577743649483,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 3.7889444828033447,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8935427574171,
+ "grad_norm": 0.05831810459494591,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 3.7719945907592773,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90750436300174,
+ "grad_norm": 0.06090841069817543,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 3.7372164726257324,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92146596858639,
+ "grad_norm": 0.05728885903954506,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 3.782083511352539,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93542757417103,
+ "grad_norm": 0.061425331979990005,
+ "learning_rate": 0.00016935382741164,
+ "loss": 3.7716455459594727,
+ "step": 2371
+ },
+ {
+ "epoch": 32.94938917975567,
+ "grad_norm": 0.05963108688592911,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 3.7798497676849365,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96335078534032,
+ "grad_norm": 0.05818144232034683,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 3.746483325958252,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97731239092496,
+ "grad_norm": 0.05970250442624092,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 3.7578821182250977,
+ "step": 2374
+ },
+ {
+ "epoch": 32.9912739965096,
+ "grad_norm": 0.05898123234510422,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 3.762807846069336,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.042496208101511,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 2.353346347808838,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 0.611886203289032,
+ "eval_runtime": 58.794,
+ "eval_samples_per_second": 41.535,
+ "eval_steps_per_second": 0.663,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01396160558464,
+ "grad_norm": 0.05598964914679527,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 3.7107529640197754,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02792321116929,
+ "grad_norm": 0.057739973068237305,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 3.6825904846191406,
+ "step": 2378
+ },
+ {
+ "epoch": 33.04188481675393,
+ "grad_norm": 0.05335357040166855,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 3.684080123901367,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05584642233857,
+ "grad_norm": 0.05573783814907074,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 3.69419527053833,
+ "step": 2380
+ },
+ {
+ "epoch": 33.06980802792321,
+ "grad_norm": 0.055793728679418564,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 3.730604648590088,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083769633507856,
+ "grad_norm": 0.054736193269491196,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 3.6922354698181152,
+ "step": 2382
+ },
+ {
+ "epoch": 33.0977312390925,
+ "grad_norm": 0.05806516483426094,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 3.7343029975891113,
+ "step": 2383
+ },
+ {
+ "epoch": 33.11169284467714,
+ "grad_norm": 0.0541117899119854,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 3.7044200897216797,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12565445026178,
+ "grad_norm": 0.057081446051597595,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 3.6954641342163086,
+ "step": 2385
+ },
+ {
+ "epoch": 33.139616055846425,
+ "grad_norm": 0.05561508610844612,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 3.709800958633423,
+ "step": 2386
+ },
+ {
+ "epoch": 33.153577661431065,
+ "grad_norm": 0.05675433203577995,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 3.7057042121887207,
+ "step": 2387
+ },
+ {
+ "epoch": 33.167539267015705,
+ "grad_norm": 0.05598882585763931,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 3.7071051597595215,
+ "step": 2388
+ },
+ {
+ "epoch": 33.181500872600346,
+ "grad_norm": 0.05500704050064087,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 3.7088711261749268,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19546247818499,
+ "grad_norm": 0.057210177183151245,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 3.685708522796631,
+ "step": 2390
+ },
+ {
+ "epoch": 33.20942408376963,
+ "grad_norm": 0.05615527555346489,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 3.695154905319214,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223385689354274,
+ "grad_norm": 0.056796107441186905,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 3.7289557456970215,
+ "step": 2392
+ },
+ {
+ "epoch": 33.23734729493892,
+ "grad_norm": 0.056371185928583145,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 3.7177929878234863,
+ "step": 2393
+ },
+ {
+ "epoch": 33.25130890052356,
+ "grad_norm": 0.05646727234125137,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 3.7265491485595703,
+ "step": 2394
+ },
+ {
+ "epoch": 33.2652705061082,
+ "grad_norm": 0.05717485025525093,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 3.7198410034179688,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27923211169284,
+ "grad_norm": 0.056603338569402695,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 3.692781686782837,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29319371727749,
+ "grad_norm": 0.05612906441092491,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 3.7237164974212646,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30715532286213,
+ "grad_norm": 0.05704745277762413,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 3.698744058609009,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32111692844677,
+ "grad_norm": 0.055712319910526276,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 3.7240562438964844,
+ "step": 2399
+ },
+ {
+ "epoch": 33.33507853403141,
+ "grad_norm": 0.05763087794184685,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 3.709852695465088,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34904013961606,
+ "grad_norm": 0.05521798133850098,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 3.7102928161621094,
+ "step": 2401
+ },
+ {
+ "epoch": 33.3630017452007,
+ "grad_norm": 0.05802507698535919,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 3.697176933288574,
+ "step": 2402
+ },
+ {
+ "epoch": 33.37696335078534,
+ "grad_norm": 0.05547751113772392,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 3.7416112422943115,
+ "step": 2403
+ },
+ {
+ "epoch": 33.390924956369986,
+ "grad_norm": 0.05753115564584732,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 3.7048134803771973,
+ "step": 2404
+ },
+ {
+ "epoch": 33.404886561954626,
+ "grad_norm": 0.05642146244645119,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 3.690187931060791,
+ "step": 2405
+ },
+ {
+ "epoch": 33.41884816753927,
+ "grad_norm": 0.05780908465385437,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 3.7087855339050293,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43280977312391,
+ "grad_norm": 0.05772170051932335,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 3.7156143188476562,
+ "step": 2407
+ },
+ {
+ "epoch": 33.446771378708554,
+ "grad_norm": 0.05813417211174965,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 3.7155094146728516,
+ "step": 2408
+ },
+ {
+ "epoch": 33.460732984293195,
+ "grad_norm": 0.056307602673769,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 3.7018208503723145,
+ "step": 2409
+ },
+ {
+ "epoch": 33.474694589877835,
+ "grad_norm": 0.05811525136232376,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 3.7002341747283936,
+ "step": 2410
+ },
+ {
+ "epoch": 33.488656195462475,
+ "grad_norm": 0.056484196335077286,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 3.715003490447998,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50261780104712,
+ "grad_norm": 0.057813555002212524,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 3.744851589202881,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51657940663176,
+ "grad_norm": 0.05758954957127571,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 3.712819814682007,
+ "step": 2413
+ },
+ {
+ "epoch": 33.5305410122164,
+ "grad_norm": 0.05723990499973297,
+ "learning_rate": 0.000159218843594243,
+ "loss": 3.7235021591186523,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544502617801044,
+ "grad_norm": 0.058723341673612595,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 3.681633949279785,
+ "step": 2415
+ },
+ {
+ "epoch": 33.55846422338569,
+ "grad_norm": 0.05725013092160225,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 3.7118563652038574,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57242582897033,
+ "grad_norm": 0.05801474303007126,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 3.7071094512939453,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58638743455497,
+ "grad_norm": 0.05690481513738632,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 3.7366995811462402,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60034904013962,
+ "grad_norm": 0.05696716532111168,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 3.7125186920166016,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61431064572426,
+ "grad_norm": 0.05725441873073578,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 3.729522705078125,
+ "step": 2420
+ },
+ {
+ "epoch": 33.6282722513089,
+ "grad_norm": 0.05838564410805702,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 3.713085412979126,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64223385689354,
+ "grad_norm": 0.05807472765445709,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 3.722982883453369,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65619546247819,
+ "grad_norm": 0.05865970253944397,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 3.7254886627197266,
+ "step": 2423
+ },
+ {
+ "epoch": 33.67015706806283,
+ "grad_norm": 0.05879775062203407,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 3.7419567108154297,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68411867364747,
+ "grad_norm": 0.058425575494766235,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 3.7334160804748535,
+ "step": 2425
+ },
+ {
+ "epoch": 33.69808027923211,
+ "grad_norm": 0.05795733258128166,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 3.6824467182159424,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712041884816756,
+ "grad_norm": 0.05840831995010376,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 3.7226808071136475,
+ "step": 2427
+ },
+ {
+ "epoch": 33.726003490401396,
+ "grad_norm": 0.05915387347340584,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 3.7192320823669434,
+ "step": 2428
+ },
+ {
+ "epoch": 33.73996509598604,
+ "grad_norm": 0.05898639187216759,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 3.731309175491333,
+ "step": 2429
+ },
+ {
+ "epoch": 33.753926701570684,
+ "grad_norm": 0.058383479714393616,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 3.6850638389587402,
+ "step": 2430
+ },
+ {
+ "epoch": 33.767888307155324,
+ "grad_norm": 0.06078571453690529,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 3.7126986980438232,
+ "step": 2431
+ },
+ {
+ "epoch": 33.781849912739965,
+ "grad_norm": 0.05873129144310951,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 3.726377010345459,
+ "step": 2432
+ },
+ {
+ "epoch": 33.795811518324605,
+ "grad_norm": 0.060784872621297836,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 3.7208423614501953,
+ "step": 2433
+ },
+ {
+ "epoch": 33.80977312390925,
+ "grad_norm": 0.05992938578128815,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 3.7019124031066895,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82373472949389,
+ "grad_norm": 0.060495104640722275,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 3.7536425590515137,
+ "step": 2435
+ },
+ {
+ "epoch": 33.83769633507853,
+ "grad_norm": 0.059453774243593216,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 3.7357583045959473,
+ "step": 2436
+ },
+ {
+ "epoch": 33.85165794066317,
+ "grad_norm": 0.06152774766087532,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 3.7266006469726562,
+ "step": 2437
+ },
+ {
+ "epoch": 33.86561954624782,
+ "grad_norm": 0.05948188528418541,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 3.712332248687744,
+ "step": 2438
+ },
+ {
+ "epoch": 33.87958115183246,
+ "grad_norm": 0.05947373807430267,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 3.696099042892456,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8935427574171,
+ "grad_norm": 0.060062918812036514,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 3.7297117710113525,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90750436300174,
+ "grad_norm": 0.057695310562849045,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 3.7090835571289062,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92146596858639,
+ "grad_norm": 0.05904509872198105,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 3.763152599334717,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93542757417103,
+ "grad_norm": 0.05886756256222725,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 3.7213687896728516,
+ "step": 2443
+ },
+ {
+ "epoch": 33.94938917975567,
+ "grad_norm": 0.06159936264157295,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 3.7479395866394043,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96335078534032,
+ "grad_norm": 0.057995621114969254,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 3.7327473163604736,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97731239092496,
+ "grad_norm": 0.058990150690078735,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 3.699246883392334,
+ "step": 2446
+ },
+ {
+ "epoch": 33.9912739965096,
+ "grad_norm": 0.05924935266375542,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 3.730475425720215,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.04313633218407631,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 2.336765766143799,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 0.6134145855903625,
+ "eval_runtime": 59.4711,
+ "eval_samples_per_second": 41.062,
+ "eval_steps_per_second": 0.656,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01396160558464,
+ "grad_norm": 0.05632878839969635,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 3.667578935623169,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02792321116929,
+ "grad_norm": 0.05653175711631775,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 3.6973605155944824,
+ "step": 2450
+ },
+ {
+ "epoch": 34.04188481675393,
+ "grad_norm": 0.05482367426156998,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 3.675959587097168,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05584642233857,
+ "grad_norm": 0.05636861175298691,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 3.673849582672119,
+ "step": 2452
+ },
+ {
+ "epoch": 34.06980802792321,
+ "grad_norm": 0.05500229448080063,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 3.6991372108459473,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083769633507856,
+ "grad_norm": 0.056783828884363174,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 3.6838951110839844,
+ "step": 2454
+ },
+ {
+ "epoch": 34.0977312390925,
+ "grad_norm": 0.05543818324804306,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 3.6698226928710938,
+ "step": 2455
+ },
+ {
+ "epoch": 34.11169284467714,
+ "grad_norm": 0.05481055751442909,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 3.678802967071533,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12565445026178,
+ "grad_norm": 0.055152181535959244,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 3.706040620803833,
+ "step": 2457
+ },
+ {
+ "epoch": 34.139616055846425,
+ "grad_norm": 0.05652368441224098,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 3.669009208679199,
+ "step": 2458
+ },
+ {
+ "epoch": 34.153577661431065,
+ "grad_norm": 0.05562841519713402,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 3.6889967918395996,
+ "step": 2459
+ },
+ {
+ "epoch": 34.167539267015705,
+ "grad_norm": 0.05579658970236778,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 3.68369197845459,
+ "step": 2460
+ },
+ {
+ "epoch": 34.181500872600346,
+ "grad_norm": 0.055683478713035583,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 3.670243740081787,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19546247818499,
+ "grad_norm": 0.05624355748295784,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 3.633751392364502,
+ "step": 2462
+ },
+ {
+ "epoch": 34.20942408376963,
+ "grad_norm": 0.05518807843327522,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 3.689868927001953,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223385689354274,
+ "grad_norm": 0.05650133267045021,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 3.694890022277832,
+ "step": 2464
+ },
+ {
+ "epoch": 34.23734729493892,
+ "grad_norm": 0.0555979385972023,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 3.6806082725524902,
+ "step": 2465
+ },
+ {
+ "epoch": 34.25130890052356,
+ "grad_norm": 0.0558185949921608,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 3.6637582778930664,
+ "step": 2466
+ },
+ {
+ "epoch": 34.2652705061082,
+ "grad_norm": 0.056465502828359604,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 3.693030834197998,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27923211169284,
+ "grad_norm": 0.0570668987929821,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 3.6703829765319824,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29319371727749,
+ "grad_norm": 0.05592618137598038,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 3.6874935626983643,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30715532286213,
+ "grad_norm": 0.055560462176799774,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 3.656379222869873,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32111692844677,
+ "grad_norm": 0.056786514818668365,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 3.685967445373535,
+ "step": 2471
+ },
+ {
+ "epoch": 34.33507853403141,
+ "grad_norm": 0.055693406611680984,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 3.693864345550537,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34904013961606,
+ "grad_norm": 0.05816012993454933,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 3.7092270851135254,
+ "step": 2473
+ },
+ {
+ "epoch": 34.3630017452007,
+ "grad_norm": 0.05769353732466698,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 3.6529183387756348,
+ "step": 2474
+ },
+ {
+ "epoch": 34.37696335078534,
+ "grad_norm": 0.058298323303461075,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 3.7092885971069336,
+ "step": 2475
+ },
+ {
+ "epoch": 34.390924956369986,
+ "grad_norm": 0.055963121354579926,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 3.6982855796813965,
+ "step": 2476
+ },
+ {
+ "epoch": 34.404886561954626,
+ "grad_norm": 0.057205602526664734,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 3.690000534057617,
+ "step": 2477
+ },
+ {
+ "epoch": 34.41884816753927,
+ "grad_norm": 0.05652588978409767,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 3.6800036430358887,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43280977312391,
+ "grad_norm": 0.059187956154346466,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 3.677968978881836,
+ "step": 2479
+ },
+ {
+ "epoch": 34.446771378708554,
+ "grad_norm": 0.05786960572004318,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 3.7025253772735596,
+ "step": 2480
+ },
+ {
+ "epoch": 34.460732984293195,
+ "grad_norm": 0.06170958653092384,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 3.6797211170196533,
+ "step": 2481
+ },
+ {
+ "epoch": 34.474694589877835,
+ "grad_norm": 0.0579230934381485,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 3.6861648559570312,
+ "step": 2482
+ },
+ {
+ "epoch": 34.488656195462475,
+ "grad_norm": 0.059319235384464264,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 3.6565022468566895,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50261780104712,
+ "grad_norm": 0.05783943459391594,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 3.722566604614258,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51657940663176,
+ "grad_norm": 0.062192801386117935,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 3.6969285011291504,
+ "step": 2485
+ },
+ {
+ "epoch": 34.5305410122164,
+ "grad_norm": 0.05853543430566788,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 3.6851119995117188,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544502617801044,
+ "grad_norm": 0.05946554243564606,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 3.6720314025878906,
+ "step": 2487
+ },
+ {
+ "epoch": 34.55846422338569,
+ "grad_norm": 0.05766003578901291,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 3.665830135345459,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57242582897033,
+ "grad_norm": 0.06046074256300926,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 3.6955373287200928,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58638743455497,
+ "grad_norm": 0.057485196739435196,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 3.7058653831481934,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60034904013962,
+ "grad_norm": 0.0580572746694088,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 3.6663808822631836,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61431064572426,
+ "grad_norm": 0.05744212865829468,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 3.7024950981140137,
+ "step": 2492
+ },
+ {
+ "epoch": 34.6282722513089,
+ "grad_norm": 0.058456238359212875,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 3.6953721046447754,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64223385689354,
+ "grad_norm": 0.05971676856279373,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 3.6731884479522705,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65619546247819,
+ "grad_norm": 0.058337483555078506,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 3.7169854640960693,
+ "step": 2495
+ },
+ {
+ "epoch": 34.67015706806283,
+ "grad_norm": 0.059567082673311234,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 3.6867635250091553,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68411867364747,
+ "grad_norm": 0.05710845813155174,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 3.6897926330566406,
+ "step": 2497
+ },
+ {
+ "epoch": 34.69808027923211,
+ "grad_norm": 0.05975678935647011,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 3.699439764022827,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712041884816756,
+ "grad_norm": 0.05828933045268059,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 3.677755355834961,
+ "step": 2499
+ },
+ {
+ "epoch": 34.726003490401396,
+ "grad_norm": 0.05896460637450218,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 3.687589645385742,
+ "step": 2500
+ },
+ {
+ "epoch": 34.73996509598604,
+ "grad_norm": 0.05770917236804962,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 3.7188401222229004,
+ "step": 2501
+ },
+ {
+ "epoch": 34.753926701570684,
+ "grad_norm": 0.05971298739314079,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 3.6803629398345947,
+ "step": 2502
+ },
+ {
+ "epoch": 34.767888307155324,
+ "grad_norm": 0.05826874077320099,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 3.7006959915161133,
+ "step": 2503
+ },
+ {
+ "epoch": 34.781849912739965,
+ "grad_norm": 0.05863497033715248,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 3.680962562561035,
+ "step": 2504
+ },
+ {
+ "epoch": 34.795811518324605,
+ "grad_norm": 0.058764465153217316,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 3.7000298500061035,
+ "step": 2505
+ },
+ {
+ "epoch": 34.80977312390925,
+ "grad_norm": 0.05830385908484459,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 3.6838178634643555,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82373472949389,
+ "grad_norm": 0.05855685845017433,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 3.709963321685791,
+ "step": 2507
+ },
+ {
+ "epoch": 34.83769633507853,
+ "grad_norm": 0.05872069671750069,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 3.702852725982666,
+ "step": 2508
+ },
+ {
+ "epoch": 34.85165794066317,
+ "grad_norm": 0.05955277755856514,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 3.699869155883789,
+ "step": 2509
+ },
+ {
+ "epoch": 34.86561954624782,
+ "grad_norm": 0.061080168932676315,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 3.7028279304504395,
+ "step": 2510
+ },
+ {
+ "epoch": 34.87958115183246,
+ "grad_norm": 0.05777806416153908,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 3.6681065559387207,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8935427574171,
+ "grad_norm": 0.0598529689013958,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 3.7238996028900146,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90750436300174,
+ "grad_norm": 0.05878746509552002,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 3.696312427520752,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92146596858639,
+ "grad_norm": 0.06028176471590996,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 3.6749677658081055,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93542757417103,
+ "grad_norm": 0.05837864801287651,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 3.675783634185791,
+ "step": 2515
+ },
+ {
+ "epoch": 34.94938917975567,
+ "grad_norm": 0.060354143381118774,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 3.6795129776000977,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96335078534032,
+ "grad_norm": 0.05837022140622139,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 3.711968183517456,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97731239092496,
+ "grad_norm": 0.06035105511546135,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 3.6994686126708984,
+ "step": 2518
+ },
+ {
+ "epoch": 34.9912739965096,
+ "grad_norm": 0.05695690959692001,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 3.6764185428619385,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.04363907873630524,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 2.2967724800109863,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 0.614627480506897,
+ "eval_runtime": 59.3089,
+ "eval_samples_per_second": 41.174,
+ "eval_steps_per_second": 0.658,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01396160558464,
+ "grad_norm": 0.05742518976330757,
+ "learning_rate": 0.0001349604788260823,
+ "loss": 3.6706724166870117,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02792321116929,
+ "grad_norm": 0.05394700914621353,
+ "learning_rate": 0.00013474073268599152,
+ "loss": 3.6419858932495117,
+ "step": 2522
+ },
+ {
+ "epoch": 35.04188481675393,
+ "grad_norm": 0.05702907219529152,
+ "learning_rate": 0.00013452112210935339,
+ "loss": 3.6621429920196533,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05584642233857,
+ "grad_norm": 0.05404812470078468,
+ "learning_rate": 0.00013430164727310399,
+ "loss": 3.6589713096618652,
+ "step": 2524
+ },
+ {
+ "epoch": 35.06980802792321,
+ "grad_norm": 0.05607686936855316,
+ "learning_rate": 0.00013408230835407027,
+ "loss": 3.634030342102051,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083769633507856,
+ "grad_norm": 0.05493795499205589,
+ "learning_rate": 0.00013386310552896977,
+ "loss": 3.678600788116455,
+ "step": 2526
+ },
+ {
+ "epoch": 35.0977312390925,
+ "grad_norm": 0.0575743205845356,
+ "learning_rate": 0.00013364403897441016,
+ "loss": 3.6530723571777344,
+ "step": 2527
+ },
+ {
+ "epoch": 35.11169284467714,
+ "grad_norm": 0.05405096337199211,
+ "learning_rate": 0.00013342510886688944,
+ "loss": 3.633359432220459,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12565445026178,
+ "grad_norm": 0.055748239159584045,
+ "learning_rate": 0.00013320631538279548,
+ "loss": 3.647953987121582,
+ "step": 2529
+ },
+ {
+ "epoch": 35.139616055846425,
+ "grad_norm": 0.05661927908658981,
+ "learning_rate": 0.00013298765869840658,
+ "loss": 3.6305713653564453,
+ "step": 2530
+ },
+ {
+ "epoch": 35.153577661431065,
+ "grad_norm": 0.05812063813209534,
+ "learning_rate": 0.00013276913898989013,
+ "loss": 3.656938076019287,
+ "step": 2531
+ },
+ {
+ "epoch": 35.167539267015705,
+ "grad_norm": 0.057856228202581406,
+ "learning_rate": 0.00013255075643330369,
+ "loss": 3.657390594482422,
+ "step": 2532
+ },
+ {
+ "epoch": 35.181500872600346,
+ "grad_norm": 0.057047586888074875,
+ "learning_rate": 0.00013233251120459407,
+ "loss": 3.6266355514526367,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19546247818499,
+ "grad_norm": 0.05664423853158951,
+ "learning_rate": 0.0001321144034795976,
+ "loss": 3.6741762161254883,
+ "step": 2534
+ },
+ {
+ "epoch": 35.20942408376963,
+ "grad_norm": 0.056682784110307693,
+ "learning_rate": 0.00013189643343403977,
+ "loss": 3.6106696128845215,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223385689354274,
+ "grad_norm": 0.05825677514076233,
+ "learning_rate": 0.00013167860124353476,
+ "loss": 3.6535234451293945,
+ "step": 2536
+ },
+ {
+ "epoch": 35.23734729493892,
+ "grad_norm": 0.05540613457560539,
+ "learning_rate": 0.00013146090708358657,
+ "loss": 3.6676511764526367,
+ "step": 2537
+ },
+ {
+ "epoch": 35.25130890052356,
+ "grad_norm": 0.05705772340297699,
+ "learning_rate": 0.00013124335112958704,
+ "loss": 3.630685567855835,
+ "step": 2538
+ },
+ {
+ "epoch": 35.2652705061082,
+ "grad_norm": 0.056012462824583054,
+ "learning_rate": 0.0001310259335568172,
+ "loss": 3.6648166179656982,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27923211169284,
+ "grad_norm": 0.05761443451046944,
+ "learning_rate": 0.00013080865454044647,
+ "loss": 3.6521716117858887,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29319371727749,
+ "grad_norm": 0.05589524284005165,
+ "learning_rate": 0.00013059151425553263,
+ "loss": 3.663506031036377,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30715532286213,
+ "grad_norm": 0.056549277156591415,
+ "learning_rate": 0.00013037451287702183,
+ "loss": 3.653916835784912,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32111692844677,
+ "grad_norm": 0.056388892233371735,
+ "learning_rate": 0.0001301576505797479,
+ "loss": 3.6622371673583984,
+ "step": 2543
+ },
+ {
+ "epoch": 35.33507853403141,
+ "grad_norm": 0.05646223947405815,
+ "learning_rate": 0.00012994092753843293,
+ "loss": 3.697643756866455,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34904013961606,
+ "grad_norm": 0.055597055703401566,
+ "learning_rate": 0.00012972434392768687,
+ "loss": 3.657986640930176,
+ "step": 2545
+ },
+ {
+ "epoch": 35.3630017452007,
+ "grad_norm": 0.055099040269851685,
+ "learning_rate": 0.00012950789992200714,
+ "loss": 3.6413917541503906,
+ "step": 2546
+ },
+ {
+ "epoch": 35.37696335078534,
+ "grad_norm": 0.0568387471139431,
+ "learning_rate": 0.00012929159569577886,
+ "loss": 3.640183210372925,
+ "step": 2547
+ },
+ {
+ "epoch": 35.390924956369986,
+ "grad_norm": 0.05552329868078232,
+ "learning_rate": 0.0001290754314232743,
+ "loss": 3.6469502449035645,
+ "step": 2548
+ },
+ {
+ "epoch": 35.404886561954626,
+ "grad_norm": 0.05681881681084633,
+ "learning_rate": 0.00012885940727865334,
+ "loss": 3.6672420501708984,
+ "step": 2549
+ },
+ {
+ "epoch": 35.41884816753927,
+ "grad_norm": 0.05603218451142311,
+ "learning_rate": 0.00012864352343596247,
+ "loss": 3.6844213008880615,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43280977312391,
+ "grad_norm": 0.05711217224597931,
+ "learning_rate": 0.0001284277800691355,
+ "loss": 3.6611440181732178,
+ "step": 2551
+ },
+ {
+ "epoch": 35.446771378708554,
+ "grad_norm": 0.05441901832818985,
+ "learning_rate": 0.00012821217735199298,
+ "loss": 3.6248745918273926,
+ "step": 2552
+ },
+ {
+ "epoch": 35.460732984293195,
+ "grad_norm": 0.05754980444908142,
+ "learning_rate": 0.00012799671545824212,
+ "loss": 3.660444736480713,
+ "step": 2553
+ },
+ {
+ "epoch": 35.474694589877835,
+ "grad_norm": 0.055383119732141495,
+ "learning_rate": 0.0001277813945614768,
+ "loss": 3.6592612266540527,
+ "step": 2554
+ },
+ {
+ "epoch": 35.488656195462475,
+ "grad_norm": 0.05781600996851921,
+ "learning_rate": 0.00012756621483517682,
+ "loss": 3.65926456451416,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50261780104712,
+ "grad_norm": 0.05716165155172348,
+ "learning_rate": 0.00012735117645270905,
+ "loss": 3.6506943702697754,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51657940663176,
+ "grad_norm": 0.05843847990036011,
+ "learning_rate": 0.00012713627958732567,
+ "loss": 3.6542019844055176,
+ "step": 2557
+ },
+ {
+ "epoch": 35.5305410122164,
+ "grad_norm": 0.057625964283943176,
+ "learning_rate": 0.00012692152441216539,
+ "loss": 3.6501665115356445,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544502617801044,
+ "grad_norm": 0.0576319545507431,
+ "learning_rate": 0.0001267069111002525,
+ "loss": 3.647266387939453,
+ "step": 2559
+ },
+ {
+ "epoch": 35.55846422338569,
+ "grad_norm": 0.05821964517235756,
+ "learning_rate": 0.00012649243982449718,
+ "loss": 3.6559696197509766,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57242582897033,
+ "grad_norm": 0.05659124627709389,
+ "learning_rate": 0.00012627811075769502,
+ "loss": 3.6653411388397217,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58638743455497,
+ "grad_norm": 0.05749163031578064,
+ "learning_rate": 0.00012606392407252687,
+ "loss": 3.6690258979797363,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60034904013962,
+ "grad_norm": 0.056859105825424194,
+ "learning_rate": 0.00012584987994155943,
+ "loss": 3.639690637588501,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61431064572426,
+ "grad_norm": 0.0584794282913208,
+ "learning_rate": 0.00012563597853724388,
+ "loss": 3.6673288345336914,
+ "step": 2564
+ },
+ {
+ "epoch": 35.6282722513089,
+ "grad_norm": 0.057991158217191696,
+ "learning_rate": 0.000125422220031917,
+ "loss": 3.6846871376037598,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64223385689354,
+ "grad_norm": 0.05805192142724991,
+ "learning_rate": 0.00012520860459779972,
+ "loss": 3.639286518096924,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65619546247819,
+ "grad_norm": 0.05739975348114967,
+ "learning_rate": 0.0001249951324069986,
+ "loss": 3.6255669593811035,
+ "step": 2567
+ },
+ {
+ "epoch": 35.67015706806283,
+ "grad_norm": 0.057306256145238876,
+ "learning_rate": 0.0001247818036315042,
+ "loss": 3.6530256271362305,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68411867364747,
+ "grad_norm": 0.056031618267297745,
+ "learning_rate": 0.00012456861844319155,
+ "loss": 3.6340036392211914,
+ "step": 2569
+ },
+ {
+ "epoch": 35.69808027923211,
+ "grad_norm": 0.05747546628117561,
+ "learning_rate": 0.00012435557701382012,
+ "loss": 3.670499086380005,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712041884816756,
+ "grad_norm": 0.056736692786216736,
+ "learning_rate": 0.0001241426795150336,
+ "loss": 3.6485583782196045,
+ "step": 2571
+ },
+ {
+ "epoch": 35.726003490401396,
+ "grad_norm": 0.05779654160141945,
+ "learning_rate": 0.00012392992611835973,
+ "loss": 3.6740036010742188,
+ "step": 2572
+ },
+ {
+ "epoch": 35.73996509598604,
+ "grad_norm": 0.05827609449625015,
+ "learning_rate": 0.0001237173169952098,
+ "loss": 3.687256336212158,
+ "step": 2573
+ },
+ {
+ "epoch": 35.753926701570684,
+ "grad_norm": 0.0587199442088604,
+ "learning_rate": 0.00012350485231687954,
+ "loss": 3.660586357116699,
+ "step": 2574
+ },
+ {
+ "epoch": 35.767888307155324,
+ "grad_norm": 0.05669032037258148,
+ "learning_rate": 0.0001232925322545476,
+ "loss": 3.6783642768859863,
+ "step": 2575
+ },
+ {
+ "epoch": 35.781849912739965,
+ "grad_norm": 0.059275999665260315,
+ "learning_rate": 0.0001230803569792765,
+ "loss": 3.660548448562622,
+ "step": 2576
+ },
+ {
+ "epoch": 35.795811518324605,
+ "grad_norm": 0.056993551552295685,
+ "learning_rate": 0.0001228683266620121,
+ "loss": 3.6559674739837646,
+ "step": 2577
+ },
+ {
+ "epoch": 35.80977312390925,
+ "grad_norm": 0.058626480400562286,
+ "learning_rate": 0.00012265644147358326,
+ "loss": 3.679635763168335,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82373472949389,
+ "grad_norm": 0.05863586813211441,
+ "learning_rate": 0.00012244470158470226,
+ "loss": 3.657258987426758,
+ "step": 2579
+ },
+ {
+ "epoch": 35.83769633507853,
+ "grad_norm": 0.05809088423848152,
+ "learning_rate": 0.00012223310716596387,
+ "loss": 3.672903299331665,
+ "step": 2580
+ },
+ {
+ "epoch": 35.85165794066317,
+ "grad_norm": 0.05710308998823166,
+ "learning_rate": 0.00012202165838784601,
+ "loss": 3.679863929748535,
+ "step": 2581
+ },
+ {
+ "epoch": 35.86561954624782,
+ "grad_norm": 0.05867332965135574,
+ "learning_rate": 0.00012181035542070913,
+ "loss": 3.689939498901367,
+ "step": 2582
+ },
+ {
+ "epoch": 35.87958115183246,
+ "grad_norm": 0.057553596794605255,
+ "learning_rate": 0.00012159919843479617,
+ "loss": 3.6570992469787598,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8935427574171,
+ "grad_norm": 0.05948316305875778,
+ "learning_rate": 0.00012138818760023257,
+ "loss": 3.6697707176208496,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90750436300174,
+ "grad_norm": 0.058080170303583145,
+ "learning_rate": 0.00012117732308702592,
+ "loss": 3.667323589324951,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92146596858639,
+ "grad_norm": 0.05842271074652672,
+ "learning_rate": 0.00012096660506506605,
+ "loss": 3.688253164291382,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93542757417103,
+ "grad_norm": 0.058192040771245956,
+ "learning_rate": 0.00012075603370412443,
+ "loss": 3.6728200912475586,
+ "step": 2587
+ },
+ {
+ "epoch": 35.94938917975567,
+ "grad_norm": 0.0602075457572937,
+ "learning_rate": 0.00012054560917385476,
+ "loss": 3.6644749641418457,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96335078534032,
+ "grad_norm": 0.05764734745025635,
+ "learning_rate": 0.00012033533164379225,
+ "loss": 3.711341381072998,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97731239092496,
+ "grad_norm": 0.05787983164191246,
+ "learning_rate": 0.0001201252012833537,
+ "loss": 3.7075908184051514,
+ "step": 2590
+ },
+ {
+ "epoch": 35.9912739965096,
+ "grad_norm": 0.05705810338258743,
+ "learning_rate": 0.00011991521826183747,
+ "loss": 3.664104461669922,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.04269756004214287,
+ "learning_rate": 0.00011970538274842282,
+ "loss": 2.2993788719177246,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 0.6160264611244202,
+ "eval_runtime": 59.3623,
+ "eval_samples_per_second": 41.137,
+ "eval_steps_per_second": 0.657,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01396160558464,
+ "grad_norm": 0.05477261170744896,
+ "learning_rate": 0.00011949569491217073,
+ "loss": 3.6223397254943848,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02792321116929,
+ "grad_norm": 0.05240058898925781,
+ "learning_rate": 0.00011928615492202269,
+ "loss": 3.6275014877319336,
+ "step": 2594
+ },
+ {
+ "epoch": 36.04188481675393,
+ "grad_norm": 0.05488358810544014,
+ "learning_rate": 0.0001190767629468014,
+ "loss": 3.6421685218811035,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05584642233857,
+ "grad_norm": 0.05403005704283714,
+ "learning_rate": 0.00011886751915521007,
+ "loss": 3.6416306495666504,
+ "step": 2596
+ },
+ {
+ "epoch": 36.06980802792321,
+ "grad_norm": 0.05383783578872681,
+ "learning_rate": 0.00011865842371583278,
+ "loss": 3.6026291847229004,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083769633507856,
+ "grad_norm": 0.05379350110888481,
+ "learning_rate": 0.00011844947679713396,
+ "loss": 3.6012558937072754,
+ "step": 2598
+ },
+ {
+ "epoch": 36.0977312390925,
+ "grad_norm": 0.054601337760686874,
+ "learning_rate": 0.00011824067856745812,
+ "loss": 3.6351709365844727,
+ "step": 2599
+ },
+ {
+ "epoch": 36.11169284467714,
+ "grad_norm": 0.0523032620549202,
+ "learning_rate": 0.00011803202919503064,
+ "loss": 3.6155266761779785,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12565445026178,
+ "grad_norm": 0.05588298290967941,
+ "learning_rate": 0.00011782352884795615,
+ "loss": 3.6355879306793213,
+ "step": 2601
+ },
+ {
+ "epoch": 36.139616055846425,
+ "grad_norm": 0.05497356131672859,
+ "learning_rate": 0.00011761517769421983,
+ "loss": 3.5998620986938477,
+ "step": 2602
+ },
+ {
+ "epoch": 36.153577661431065,
+ "grad_norm": 0.055099356919527054,
+ "learning_rate": 0.00011740697590168635,
+ "loss": 3.6135497093200684,
+ "step": 2603
+ },
+ {
+ "epoch": 36.167539267015705,
+ "grad_norm": 0.05480733513832092,
+ "learning_rate": 0.00011719892363810018,
+ "loss": 3.631089687347412,
+ "step": 2604
+ },
+ {
+ "epoch": 36.181500872600346,
+ "grad_norm": 0.05424918979406357,
+ "learning_rate": 0.00011699102107108539,
+ "loss": 3.610616683959961,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19546247818499,
+ "grad_norm": 0.05589015781879425,
+ "learning_rate": 0.00011678326836814507,
+ "loss": 3.6014404296875,
+ "step": 2606
+ },
+ {
+ "epoch": 36.20942408376963,
+ "grad_norm": 0.05365624278783798,
+ "learning_rate": 0.00011657566569666198,
+ "loss": 3.6203207969665527,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223385689354274,
+ "grad_norm": 0.05549495667219162,
+ "learning_rate": 0.00011636821322389777,
+ "loss": 3.639944076538086,
+ "step": 2608
+ },
+ {
+ "epoch": 36.23734729493892,
+ "grad_norm": 0.05381424352526665,
+ "learning_rate": 0.00011616091111699333,
+ "loss": 3.6630518436431885,
+ "step": 2609
+ },
+ {
+ "epoch": 36.25130890052356,
+ "grad_norm": 0.05610237643122673,
+ "learning_rate": 0.00011595375954296786,
+ "loss": 3.634977102279663,
+ "step": 2610
+ },
+ {
+ "epoch": 36.2652705061082,
+ "grad_norm": 0.055497750639915466,
+ "learning_rate": 0.00011574675866871997,
+ "loss": 3.64717960357666,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27923211169284,
+ "grad_norm": 0.0543544627726078,
+ "learning_rate": 0.00011553990866102651,
+ "loss": 3.6224818229675293,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29319371727749,
+ "grad_norm": 0.056236982345581055,
+ "learning_rate": 0.00011533320968654265,
+ "loss": 3.6230814456939697,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30715532286213,
+ "grad_norm": 0.053707852959632874,
+ "learning_rate": 0.00011512666191180204,
+ "loss": 3.6355342864990234,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32111692844677,
+ "grad_norm": 0.057036347687244415,
+ "learning_rate": 0.00011492026550321653,
+ "loss": 3.6099886894226074,
+ "step": 2615
+ },
+ {
+ "epoch": 36.33507853403141,
+ "grad_norm": 0.05581443011760712,
+ "learning_rate": 0.00011471402062707607,
+ "loss": 3.6411590576171875,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34904013961606,
+ "grad_norm": 0.05495850369334221,
+ "learning_rate": 0.00011450792744954827,
+ "loss": 3.6358461380004883,
+ "step": 2617
+ },
+ {
+ "epoch": 36.3630017452007,
+ "grad_norm": 0.0549640990793705,
+ "learning_rate": 0.00011430198613667861,
+ "loss": 3.633671283721924,
+ "step": 2618
+ },
+ {
+ "epoch": 36.37696335078534,
+ "grad_norm": 0.05502061918377876,
+ "learning_rate": 0.00011409619685439064,
+ "loss": 3.632413864135742,
+ "step": 2619
+ },
+ {
+ "epoch": 36.390924956369986,
+ "grad_norm": 0.05518985912203789,
+ "learning_rate": 0.00011389055976848477,
+ "loss": 3.646486759185791,
+ "step": 2620
+ },
+ {
+ "epoch": 36.404886561954626,
+ "grad_norm": 0.053396400064229965,
+ "learning_rate": 0.00011368507504463914,
+ "loss": 3.6610465049743652,
+ "step": 2621
+ },
+ {
+ "epoch": 36.41884816753927,
+ "grad_norm": 0.05592766031622887,
+ "learning_rate": 0.00011347974284840914,
+ "loss": 3.6322021484375,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43280977312391,
+ "grad_norm": 0.054665833711624146,
+ "learning_rate": 0.00011327456334522725,
+ "loss": 3.6100504398345947,
+ "step": 2623
+ },
+ {
+ "epoch": 36.446771378708554,
+ "grad_norm": 0.05556786060333252,
+ "learning_rate": 0.00011306953670040275,
+ "loss": 3.647045135498047,
+ "step": 2624
+ },
+ {
+ "epoch": 36.460732984293195,
+ "grad_norm": 0.05619419366121292,
+ "learning_rate": 0.00011286466307912184,
+ "loss": 3.6307947635650635,
+ "step": 2625
+ },
+ {
+ "epoch": 36.474694589877835,
+ "grad_norm": 0.0551043376326561,
+ "learning_rate": 0.00011265994264644783,
+ "loss": 3.6331896781921387,
+ "step": 2626
+ },
+ {
+ "epoch": 36.488656195462475,
+ "grad_norm": 0.0548650398850441,
+ "learning_rate": 0.00011245537556732001,
+ "loss": 3.633991241455078,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50261780104712,
+ "grad_norm": 0.056781429797410965,
+ "learning_rate": 0.00011225096200655442,
+ "loss": 3.653165340423584,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51657940663176,
+ "grad_norm": 0.05489041283726692,
+ "learning_rate": 0.00011204670212884342,
+ "loss": 3.6232171058654785,
+ "step": 2629
+ },
+ {
+ "epoch": 36.5305410122164,
+ "grad_norm": 0.055730190128088,
+ "learning_rate": 0.00011184259609875557,
+ "loss": 3.644775390625,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544502617801044,
+ "grad_norm": 0.058052800595760345,
+ "learning_rate": 0.00011163864408073522,
+ "loss": 3.6141133308410645,
+ "step": 2631
+ },
+ {
+ "epoch": 36.55846422338569,
+ "grad_norm": 0.05508120357990265,
+ "learning_rate": 0.00011143484623910293,
+ "loss": 3.6275734901428223,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57242582897033,
+ "grad_norm": 0.0572693794965744,
+ "learning_rate": 0.00011123120273805496,
+ "loss": 3.631286859512329,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58638743455497,
+ "grad_norm": 0.05745089426636696,
+ "learning_rate": 0.0001110277137416632,
+ "loss": 3.6396098136901855,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60034904013962,
+ "grad_norm": 0.05624939128756523,
+ "learning_rate": 0.00011082437941387512,
+ "loss": 3.655447006225586,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61431064572426,
+ "grad_norm": 0.057037804275751114,
+ "learning_rate": 0.00011062119991851322,
+ "loss": 3.6295433044433594,
+ "step": 2636
+ },
+ {
+ "epoch": 36.6282722513089,
+ "grad_norm": 0.05740547552704811,
+ "learning_rate": 0.000110418175419276,
+ "loss": 3.6350302696228027,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64223385689354,
+ "grad_norm": 0.05586782470345497,
+ "learning_rate": 0.00011021530607973626,
+ "loss": 3.6298985481262207,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65619546247819,
+ "grad_norm": 0.05766303837299347,
+ "learning_rate": 0.00011001259206334235,
+ "loss": 3.646199941635132,
+ "step": 2639
+ },
+ {
+ "epoch": 36.67015706806283,
+ "grad_norm": 0.05768761411309242,
+ "learning_rate": 0.00010981003353341721,
+ "loss": 3.638676881790161,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68411867364747,
+ "grad_norm": 0.05806223303079605,
+ "learning_rate": 0.00010960763065315864,
+ "loss": 3.61555814743042,
+ "step": 2641
+ },
+ {
+ "epoch": 36.69808027923211,
+ "grad_norm": 0.05654353275895119,
+ "learning_rate": 0.00010940538358563906,
+ "loss": 3.6435892581939697,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712041884816756,
+ "grad_norm": 0.0572567842900753,
+ "learning_rate": 0.00010920329249380509,
+ "loss": 3.638545513153076,
+ "step": 2643
+ },
+ {
+ "epoch": 36.726003490401396,
+ "grad_norm": 0.05685669183731079,
+ "learning_rate": 0.00010900135754047799,
+ "loss": 3.6248223781585693,
+ "step": 2644
+ },
+ {
+ "epoch": 36.73996509598604,
+ "grad_norm": 0.05836780369281769,
+ "learning_rate": 0.00010879957888835307,
+ "loss": 3.6728572845458984,
+ "step": 2645
+ },
+ {
+ "epoch": 36.753926701570684,
+ "grad_norm": 0.05652812868356705,
+ "learning_rate": 0.00010859795669999981,
+ "loss": 3.6360411643981934,
+ "step": 2646
+ },
+ {
+ "epoch": 36.767888307155324,
+ "grad_norm": 0.05731404572725296,
+ "learning_rate": 0.00010839649113786137,
+ "loss": 3.624966621398926,
+ "step": 2647
+ },
+ {
+ "epoch": 36.781849912739965,
+ "grad_norm": 0.05664033442735672,
+ "learning_rate": 0.00010819518236425514,
+ "loss": 3.6407511234283447,
+ "step": 2648
+ },
+ {
+ "epoch": 36.795811518324605,
+ "grad_norm": 0.056350719183683395,
+ "learning_rate": 0.00010799403054137197,
+ "loss": 3.6436259746551514,
+ "step": 2649
+ },
+ {
+ "epoch": 36.80977312390925,
+ "grad_norm": 0.05680667608976364,
+ "learning_rate": 0.00010779303583127609,
+ "loss": 3.6506614685058594,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82373472949389,
+ "grad_norm": 0.05701072886586189,
+ "learning_rate": 0.00010759219839590535,
+ "loss": 3.643155574798584,
+ "step": 2651
+ },
+ {
+ "epoch": 36.83769633507853,
+ "grad_norm": 0.05728844553232193,
+ "learning_rate": 0.00010739151839707089,
+ "loss": 3.6734180450439453,
+ "step": 2652
+ },
+ {
+ "epoch": 36.85165794066317,
+ "grad_norm": 0.05713675543665886,
+ "learning_rate": 0.00010719099599645706,
+ "loss": 3.665465831756592,
+ "step": 2653
+ },
+ {
+ "epoch": 36.86561954624782,
+ "grad_norm": 0.05626962333917618,
+ "learning_rate": 0.0001069906313556208,
+ "loss": 3.6559338569641113,
+ "step": 2654
+ },
+ {
+ "epoch": 36.87958115183246,
+ "grad_norm": 0.05827031657099724,
+ "learning_rate": 0.00010679042463599263,
+ "loss": 3.64331316947937,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8935427574171,
+ "grad_norm": 0.05828213319182396,
+ "learning_rate": 0.00010659037599887545,
+ "loss": 3.631319046020508,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90750436300174,
+ "grad_norm": 0.057474542409181595,
+ "learning_rate": 0.0001063904856054446,
+ "loss": 3.6365761756896973,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92146596858639,
+ "grad_norm": 0.056101586669683456,
+ "learning_rate": 0.00010619075361674836,
+ "loss": 3.62568998336792,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93542757417103,
+ "grad_norm": 0.058322884142398834,
+ "learning_rate": 0.0001059911801937071,
+ "loss": 3.6452713012695312,
+ "step": 2659
+ },
+ {
+ "epoch": 36.94938917975567,
+ "grad_norm": 0.05735790356993675,
+ "learning_rate": 0.00010579176549711365,
+ "loss": 3.644145965576172,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96335078534032,
+ "grad_norm": 0.05761740356683731,
+ "learning_rate": 0.00010559250968763265,
+ "loss": 3.652322292327881,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97731239092496,
+ "grad_norm": 0.05765356495976448,
+ "learning_rate": 0.00010539341292580086,
+ "loss": 3.6571314334869385,
+ "step": 2662
+ },
+ {
+ "epoch": 36.9912739965096,
+ "grad_norm": 0.05606544017791748,
+ "learning_rate": 0.00010519447537202729,
+ "loss": 3.6158785820007324,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.042550574988126755,
+ "learning_rate": 0.00010499569718659201,
+ "loss": 2.262657403945923,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 0.6172080039978027,
+ "eval_runtime": 59.1561,
+ "eval_samples_per_second": 41.281,
+ "eval_steps_per_second": 0.659,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01396160558464,
+ "grad_norm": 0.054308392107486725,
+ "learning_rate": 0.00010479707852964713,
+ "loss": 3.6250295639038086,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02792321116929,
+ "grad_norm": 0.05218822881579399,
+ "learning_rate": 0.00010459861956121611,
+ "loss": 3.598968982696533,
+ "step": 2666
+ },
+ {
+ "epoch": 37.04188481675393,
+ "grad_norm": 0.05366450920701027,
+ "learning_rate": 0.00010440032044119383,
+ "loss": 3.5767178535461426,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05584642233857,
+ "grad_norm": 0.05169176682829857,
+ "learning_rate": 0.00010420218132934614,
+ "loss": 3.5973262786865234,
+ "step": 2668
+ },
+ {
+ "epoch": 37.06980802792321,
+ "grad_norm": 0.05438115820288658,
+ "learning_rate": 0.00010400420238531023,
+ "loss": 3.584498167037964,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083769633507856,
+ "grad_norm": 0.05290330946445465,
+ "learning_rate": 0.00010380638376859416,
+ "loss": 3.614943504333496,
+ "step": 2670
+ },
+ {
+ "epoch": 37.0977312390925,
+ "grad_norm": 0.052781153470277786,
+ "learning_rate": 0.00010360872563857682,
+ "loss": 3.5990896224975586,
+ "step": 2671
+ },
+ {
+ "epoch": 37.11169284467714,
+ "grad_norm": 0.05197159945964813,
+ "learning_rate": 0.0001034112281545079,
+ "loss": 3.579005241394043,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12565445026178,
+ "grad_norm": 0.05242403224110603,
+ "learning_rate": 0.00010321389147550725,
+ "loss": 3.608119010925293,
+ "step": 2673
+ },
+ {
+ "epoch": 37.139616055846425,
+ "grad_norm": 0.05325670540332794,
+ "learning_rate": 0.00010301671576056588,
+ "loss": 3.6093316078186035,
+ "step": 2674
+ },
+ {
+ "epoch": 37.153577661431065,
+ "grad_norm": 0.05325893685221672,
+ "learning_rate": 0.00010281970116854436,
+ "loss": 3.6037325859069824,
+ "step": 2675
+ },
+ {
+ "epoch": 37.167539267015705,
+ "grad_norm": 0.05180941894650459,
+ "learning_rate": 0.00010262284785817392,
+ "loss": 3.584935188293457,
+ "step": 2676
+ },
+ {
+ "epoch": 37.181500872600346,
+ "grad_norm": 0.05396177992224693,
+ "learning_rate": 0.00010242615598805574,
+ "loss": 3.623363494873047,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19546247818499,
+ "grad_norm": 0.05336499959230423,
+ "learning_rate": 0.00010222962571666088,
+ "loss": 3.6231229305267334,
+ "step": 2678
+ },
+ {
+ "epoch": 37.20942408376963,
+ "grad_norm": 0.052409425377845764,
+ "learning_rate": 0.00010203325720233032,
+ "loss": 3.599355697631836,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223385689354274,
+ "grad_norm": 0.0531868040561676,
+ "learning_rate": 0.00010183705060327433,
+ "loss": 3.6197028160095215,
+ "step": 2680
+ },
+ {
+ "epoch": 37.23734729493892,
+ "grad_norm": 0.053192511200904846,
+ "learning_rate": 0.00010164100607757346,
+ "loss": 3.612718105316162,
+ "step": 2681
+ },
+ {
+ "epoch": 37.25130890052356,
+ "grad_norm": 0.052938785403966904,
+ "learning_rate": 0.00010144512378317687,
+ "loss": 3.6197361946105957,
+ "step": 2682
+ },
+ {
+ "epoch": 37.2652705061082,
+ "grad_norm": 0.0538516603410244,
+ "learning_rate": 0.00010124940387790354,
+ "loss": 3.6179161071777344,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27923211169284,
+ "grad_norm": 0.05364573001861572,
+ "learning_rate": 0.00010105384651944148,
+ "loss": 3.5888404846191406,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29319371727749,
+ "grad_norm": 0.05395280569791794,
+ "learning_rate": 0.00010085845186534769,
+ "loss": 3.6174027919769287,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30715532286213,
+ "grad_norm": 0.05424579605460167,
+ "learning_rate": 0.00010066322007304819,
+ "loss": 3.6061956882476807,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32111692844677,
+ "grad_norm": 0.05412362515926361,
+ "learning_rate": 0.00010046815129983757,
+ "loss": 3.596550464630127,
+ "step": 2687
+ },
+ {
+ "epoch": 37.33507853403141,
+ "grad_norm": 0.05387300252914429,
+ "learning_rate": 0.00010027324570287925,
+ "loss": 3.627476215362549,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34904013961606,
+ "grad_norm": 0.05500417947769165,
+ "learning_rate": 0.00010007850343920519,
+ "loss": 3.601841926574707,
+ "step": 2689
+ },
+ {
+ "epoch": 37.3630017452007,
+ "grad_norm": 0.05346129834651947,
+ "learning_rate": 9.988392466571572e-05,
+ "loss": 3.6150965690612793,
+ "step": 2690
+ },
+ {
+ "epoch": 37.37696335078534,
+ "grad_norm": 0.054463278502225876,
+ "learning_rate": 9.968950953917945e-05,
+ "loss": 3.5890097618103027,
+ "step": 2691
+ },
+ {
+ "epoch": 37.390924956369986,
+ "grad_norm": 0.05312243849039078,
+ "learning_rate": 9.949525821623309e-05,
+ "loss": 3.597362518310547,
+ "step": 2692
+ },
+ {
+ "epoch": 37.404886561954626,
+ "grad_norm": 0.05424235016107559,
+ "learning_rate": 9.930117085338156e-05,
+ "loss": 3.5707716941833496,
+ "step": 2693
+ },
+ {
+ "epoch": 37.41884816753927,
+ "grad_norm": 0.054021209478378296,
+ "learning_rate": 9.910724760699731e-05,
+ "loss": 3.6366126537323,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43280977312391,
+ "grad_norm": 0.05502825230360031,
+ "learning_rate": 9.891348863332093e-05,
+ "loss": 3.6250295639038086,
+ "step": 2695
+ },
+ {
+ "epoch": 37.446771378708554,
+ "grad_norm": 0.05483090132474899,
+ "learning_rate": 9.871989408846051e-05,
+ "loss": 3.611407518386841,
+ "step": 2696
+ },
+ {
+ "epoch": 37.460732984293195,
+ "grad_norm": 0.05381562188267708,
+ "learning_rate": 9.852646412839178e-05,
+ "loss": 3.6136293411254883,
+ "step": 2697
+ },
+ {
+ "epoch": 37.474694589877835,
+ "grad_norm": 0.054287511855363846,
+ "learning_rate": 9.833319890895756e-05,
+ "loss": 3.6080033779144287,
+ "step": 2698
+ },
+ {
+ "epoch": 37.488656195462475,
+ "grad_norm": 0.05399102717638016,
+ "learning_rate": 9.814009858586817e-05,
+ "loss": 3.6026833057403564,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50261780104712,
+ "grad_norm": 0.053677961230278015,
+ "learning_rate": 9.794716331470131e-05,
+ "loss": 3.6080117225646973,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51657940663176,
+ "grad_norm": 0.055184055119752884,
+ "learning_rate": 9.77543932509012e-05,
+ "loss": 3.6028759479522705,
+ "step": 2701
+ },
+ {
+ "epoch": 37.5305410122164,
+ "grad_norm": 0.0554618164896965,
+ "learning_rate": 9.756178854977925e-05,
+ "loss": 3.6170661449432373,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544502617801044,
+ "grad_norm": 0.05491705983877182,
+ "learning_rate": 9.736934936651362e-05,
+ "loss": 3.6152079105377197,
+ "step": 2703
+ },
+ {
+ "epoch": 37.55846422338569,
+ "grad_norm": 0.055647511035203934,
+ "learning_rate": 9.717707585614916e-05,
+ "loss": 3.610424757003784,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57242582897033,
+ "grad_norm": 0.0561630018055439,
+ "learning_rate": 9.698496817359698e-05,
+ "loss": 3.6318728923797607,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58638743455497,
+ "grad_norm": 0.05447172373533249,
+ "learning_rate": 9.679302647363476e-05,
+ "loss": 3.6186347007751465,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60034904013962,
+ "grad_norm": 0.05509922280907631,
+ "learning_rate": 9.660125091090675e-05,
+ "loss": 3.6077895164489746,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61431064572426,
+ "grad_norm": 0.055178701877593994,
+ "learning_rate": 9.64096416399228e-05,
+ "loss": 3.637204647064209,
+ "step": 2708
+ },
+ {
+ "epoch": 37.6282722513089,
+ "grad_norm": 0.05637604370713234,
+ "learning_rate": 9.621819881505918e-05,
+ "loss": 3.6154603958129883,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64223385689354,
+ "grad_norm": 0.056297753006219864,
+ "learning_rate": 9.602692259055767e-05,
+ "loss": 3.6120991706848145,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65619546247819,
+ "grad_norm": 0.054915741086006165,
+ "learning_rate": 9.583581312052646e-05,
+ "loss": 3.626859188079834,
+ "step": 2711
+ },
+ {
+ "epoch": 37.67015706806283,
+ "grad_norm": 0.05664214491844177,
+ "learning_rate": 9.564487055893867e-05,
+ "loss": 3.6137616634368896,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68411867364747,
+ "grad_norm": 0.05572838336229324,
+ "learning_rate": 9.545409505963338e-05,
+ "loss": 3.622990131378174,
+ "step": 2713
+ },
+ {
+ "epoch": 37.69808027923211,
+ "grad_norm": 0.056440044194459915,
+ "learning_rate": 9.526348677631499e-05,
+ "loss": 3.6483852863311768,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712041884816756,
+ "grad_norm": 0.0582021027803421,
+ "learning_rate": 9.507304586255311e-05,
+ "loss": 3.6085047721862793,
+ "step": 2715
+ },
+ {
+ "epoch": 37.726003490401396,
+ "grad_norm": 0.055895350873470306,
+ "learning_rate": 9.488277247178267e-05,
+ "loss": 3.625001907348633,
+ "step": 2716
+ },
+ {
+ "epoch": 37.73996509598604,
+ "grad_norm": 0.055818382650613785,
+ "learning_rate": 9.469266675730319e-05,
+ "loss": 3.627051830291748,
+ "step": 2717
+ },
+ {
+ "epoch": 37.753926701570684,
+ "grad_norm": 0.05565737560391426,
+ "learning_rate": 9.450272887227983e-05,
+ "loss": 3.607292890548706,
+ "step": 2718
+ },
+ {
+ "epoch": 37.767888307155324,
+ "grad_norm": 0.05473797023296356,
+ "learning_rate": 9.431295896974182e-05,
+ "loss": 3.613346815109253,
+ "step": 2719
+ },
+ {
+ "epoch": 37.781849912739965,
+ "grad_norm": 0.056405726820230484,
+ "learning_rate": 9.412335720258341e-05,
+ "loss": 3.6148681640625,
+ "step": 2720
+ },
+ {
+ "epoch": 37.795811518324605,
+ "grad_norm": 0.0563163161277771,
+ "learning_rate": 9.393392372356335e-05,
+ "loss": 3.6225497722625732,
+ "step": 2721
+ },
+ {
+ "epoch": 37.80977312390925,
+ "grad_norm": 0.05682912468910217,
+ "learning_rate": 9.374465868530477e-05,
+ "loss": 3.620290756225586,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82373472949389,
+ "grad_norm": 0.05696539953351021,
+ "learning_rate": 9.355556224029515e-05,
+ "loss": 3.617171287536621,
+ "step": 2723
+ },
+ {
+ "epoch": 37.83769633507853,
+ "grad_norm": 0.05691574513912201,
+ "learning_rate": 9.336663454088593e-05,
+ "loss": 3.6293022632598877,
+ "step": 2724
+ },
+ {
+ "epoch": 37.85165794066317,
+ "grad_norm": 0.05566944554448128,
+ "learning_rate": 9.317787573929282e-05,
+ "loss": 3.6245670318603516,
+ "step": 2725
+ },
+ {
+ "epoch": 37.86561954624782,
+ "grad_norm": 0.05736027657985687,
+ "learning_rate": 9.298928598759541e-05,
+ "loss": 3.625857353210449,
+ "step": 2726
+ },
+ {
+ "epoch": 37.87958115183246,
+ "grad_norm": 0.055642999708652496,
+ "learning_rate": 9.280086543773698e-05,
+ "loss": 3.610138416290283,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8935427574171,
+ "grad_norm": 0.057330258190631866,
+ "learning_rate": 9.26126142415247e-05,
+ "loss": 3.593010902404785,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90750436300174,
+ "grad_norm": 0.05633636936545372,
+ "learning_rate": 9.2424532550629e-05,
+ "loss": 3.6144258975982666,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92146596858639,
+ "grad_norm": 0.05787842720746994,
+ "learning_rate": 9.223662051658408e-05,
+ "loss": 3.609016180038452,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93542757417103,
+ "grad_norm": 0.05679284781217575,
+ "learning_rate": 9.204887829078709e-05,
+ "loss": 3.5973129272460938,
+ "step": 2731
+ },
+ {
+ "epoch": 37.94938917975567,
+ "grad_norm": 0.056735891848802567,
+ "learning_rate": 9.186130602449861e-05,
+ "loss": 3.6157774925231934,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96335078534032,
+ "grad_norm": 0.05625693500041962,
+ "learning_rate": 9.167390386884224e-05,
+ "loss": 3.617924928665161,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97731239092496,
+ "grad_norm": 0.05621540546417236,
+ "learning_rate": 9.148667197480455e-05,
+ "loss": 3.597618341445923,
+ "step": 2734
+ },
+ {
+ "epoch": 37.9912739965096,
+ "grad_norm": 0.0570329986512661,
+ "learning_rate": 9.129961049323494e-05,
+ "loss": 3.6188273429870605,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.04125213250517845,
+ "learning_rate": 9.111271957484519e-05,
+ "loss": 2.2358860969543457,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 0.6186589002609253,
+ "eval_runtime": 58.8494,
+ "eval_samples_per_second": 41.496,
+ "eval_steps_per_second": 0.663,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01396160558464,
+ "grad_norm": 0.05371171608567238,
+ "learning_rate": 9.092599937021031e-05,
+ "loss": 3.5614218711853027,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02792321116929,
+ "grad_norm": 0.05166265740990639,
+ "learning_rate": 9.073945002976715e-05,
+ "loss": 3.5690364837646484,
+ "step": 2738
+ },
+ {
+ "epoch": 38.04188481675393,
+ "grad_norm": 0.053279295563697815,
+ "learning_rate": 9.055307170381522e-05,
+ "loss": 3.575930595397949,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05584642233857,
+ "grad_norm": 0.051418207585811615,
+ "learning_rate": 9.036686454251613e-05,
+ "loss": 3.568121910095215,
+ "step": 2740
+ },
+ {
+ "epoch": 38.06980802792321,
+ "grad_norm": 0.051363058388233185,
+ "learning_rate": 9.018082869589369e-05,
+ "loss": 3.5916590690612793,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083769633507856,
+ "grad_norm": 0.05149349942803383,
+ "learning_rate": 8.999496431383368e-05,
+ "loss": 3.5861239433288574,
+ "step": 2742
+ },
+ {
+ "epoch": 38.0977312390925,
+ "grad_norm": 0.05174670368432999,
+ "learning_rate": 8.980927154608341e-05,
+ "loss": 3.5812082290649414,
+ "step": 2743
+ },
+ {
+ "epoch": 38.11169284467714,
+ "grad_norm": 0.05128023028373718,
+ "learning_rate": 8.962375054225255e-05,
+ "loss": 3.5962631702423096,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12565445026178,
+ "grad_norm": 0.05153957009315491,
+ "learning_rate": 8.943840145181177e-05,
+ "loss": 3.5739307403564453,
+ "step": 2745
+ },
+ {
+ "epoch": 38.139616055846425,
+ "grad_norm": 0.0521809458732605,
+ "learning_rate": 8.925322442409361e-05,
+ "loss": 3.5964818000793457,
+ "step": 2746
+ },
+ {
+ "epoch": 38.153577661431065,
+ "grad_norm": 0.05092332884669304,
+ "learning_rate": 8.906821960829163e-05,
+ "loss": 3.5737056732177734,
+ "step": 2747
+ },
+ {
+ "epoch": 38.167539267015705,
+ "grad_norm": 0.052281491458415985,
+ "learning_rate": 8.888338715346113e-05,
+ "loss": 3.5943212509155273,
+ "step": 2748
+ },
+ {
+ "epoch": 38.181500872600346,
+ "grad_norm": 0.05170704424381256,
+ "learning_rate": 8.869872720851831e-05,
+ "loss": 3.585753917694092,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19546247818499,
+ "grad_norm": 0.05175631120800972,
+ "learning_rate": 8.851423992224012e-05,
+ "loss": 3.606010913848877,
+ "step": 2750
+ },
+ {
+ "epoch": 38.20942408376963,
+ "grad_norm": 0.0525258332490921,
+ "learning_rate": 8.832992544326473e-05,
+ "loss": 3.591367244720459,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223385689354274,
+ "grad_norm": 0.051923174411058426,
+ "learning_rate": 8.814578392009104e-05,
+ "loss": 3.563871383666992,
+ "step": 2752
+ },
+ {
+ "epoch": 38.23734729493892,
+ "grad_norm": 0.053746938705444336,
+ "learning_rate": 8.796181550107857e-05,
+ "loss": 3.576528787612915,
+ "step": 2753
+ },
+ {
+ "epoch": 38.25130890052356,
+ "grad_norm": 0.05283728986978531,
+ "learning_rate": 8.777802033444712e-05,
+ "loss": 3.600759983062744,
+ "step": 2754
+ },
+ {
+ "epoch": 38.2652705061082,
+ "grad_norm": 0.05214889720082283,
+ "learning_rate": 8.75943985682774e-05,
+ "loss": 3.582427501678467,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27923211169284,
+ "grad_norm": 0.05283285677433014,
+ "learning_rate": 8.741095035051017e-05,
+ "loss": 3.579399585723877,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29319371727749,
+ "grad_norm": 0.053677286952733994,
+ "learning_rate": 8.722767582894613e-05,
+ "loss": 3.5718324184417725,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30715532286213,
+ "grad_norm": 0.05384226143360138,
+ "learning_rate": 8.704457515124636e-05,
+ "loss": 3.593702793121338,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32111692844677,
+ "grad_norm": 0.05237982049584389,
+ "learning_rate": 8.686164846493176e-05,
+ "loss": 3.580988883972168,
+ "step": 2759
+ },
+ {
+ "epoch": 38.33507853403141,
+ "grad_norm": 0.05386205390095711,
+ "learning_rate": 8.667889591738317e-05,
+ "loss": 3.5706114768981934,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34904013961606,
+ "grad_norm": 0.05325917899608612,
+ "learning_rate": 8.649631765584083e-05,
+ "loss": 3.5707662105560303,
+ "step": 2761
+ },
+ {
+ "epoch": 38.3630017452007,
+ "grad_norm": 0.053492557257413864,
+ "learning_rate": 8.631391382740482e-05,
+ "loss": 3.618515968322754,
+ "step": 2762
+ },
+ {
+ "epoch": 38.37696335078534,
+ "grad_norm": 0.05370897799730301,
+ "learning_rate": 8.613168457903459e-05,
+ "loss": 3.5731120109558105,
+ "step": 2763
+ },
+ {
+ "epoch": 38.390924956369986,
+ "grad_norm": 0.05325158312916756,
+ "learning_rate": 8.594963005754901e-05,
+ "loss": 3.591339588165283,
+ "step": 2764
+ },
+ {
+ "epoch": 38.404886561954626,
+ "grad_norm": 0.05411994829773903,
+ "learning_rate": 8.57677504096261e-05,
+ "loss": 3.5969743728637695,
+ "step": 2765
+ },
+ {
+ "epoch": 38.41884816753927,
+ "grad_norm": 0.05363132059574127,
+ "learning_rate": 8.558604578180301e-05,
+ "loss": 3.620718479156494,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43280977312391,
+ "grad_norm": 0.0541871078312397,
+ "learning_rate": 8.540451632047593e-05,
+ "loss": 3.589287281036377,
+ "step": 2767
+ },
+ {
+ "epoch": 38.446771378708554,
+ "grad_norm": 0.05317665636539459,
+ "learning_rate": 8.522316217189972e-05,
+ "loss": 3.5677599906921387,
+ "step": 2768
+ },
+ {
+ "epoch": 38.460732984293195,
+ "grad_norm": 0.05288826301693916,
+ "learning_rate": 8.504198348218821e-05,
+ "loss": 3.5985312461853027,
+ "step": 2769
+ },
+ {
+ "epoch": 38.474694589877835,
+ "grad_norm": 0.05443241819739342,
+ "learning_rate": 8.486098039731384e-05,
+ "loss": 3.5577774047851562,
+ "step": 2770
+ },
+ {
+ "epoch": 38.488656195462475,
+ "grad_norm": 0.054173294454813004,
+ "learning_rate": 8.46801530631075e-05,
+ "loss": 3.5860495567321777,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50261780104712,
+ "grad_norm": 0.05407063290476799,
+ "learning_rate": 8.449950162525859e-05,
+ "loss": 3.604748249053955,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51657940663176,
+ "grad_norm": 0.05461187660694122,
+ "learning_rate": 8.431902622931443e-05,
+ "loss": 3.563236713409424,
+ "step": 2773
+ },
+ {
+ "epoch": 38.5305410122164,
+ "grad_norm": 0.05529208108782768,
+ "learning_rate": 8.413872702068119e-05,
+ "loss": 3.586937427520752,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544502617801044,
+ "grad_norm": 0.05481037497520447,
+ "learning_rate": 8.395860414462238e-05,
+ "loss": 3.6134698390960693,
+ "step": 2775
+ },
+ {
+ "epoch": 38.55846422338569,
+ "grad_norm": 0.054424822330474854,
+ "learning_rate": 8.377865774625985e-05,
+ "loss": 3.575363874435425,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57242582897033,
+ "grad_norm": 0.0553407222032547,
+ "learning_rate": 8.359888797057316e-05,
+ "loss": 3.5970382690429688,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58638743455497,
+ "grad_norm": 0.05625150352716446,
+ "learning_rate": 8.341929496239958e-05,
+ "loss": 3.6032121181488037,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60034904013962,
+ "grad_norm": 0.05586005747318268,
+ "learning_rate": 8.323987886643402e-05,
+ "loss": 3.602691888809204,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61431064572426,
+ "grad_norm": 0.054208606481552124,
+ "learning_rate": 8.306063982722855e-05,
+ "loss": 3.601468563079834,
+ "step": 2780
+ },
+ {
+ "epoch": 38.6282722513089,
+ "grad_norm": 0.056410644203424454,
+ "learning_rate": 8.288157798919318e-05,
+ "loss": 3.585646867752075,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64223385689354,
+ "grad_norm": 0.054060712456703186,
+ "learning_rate": 8.27026934965945e-05,
+ "loss": 3.566913604736328,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65619546247819,
+ "grad_norm": 0.0546812042593956,
+ "learning_rate": 8.25239864935566e-05,
+ "loss": 3.581869602203369,
+ "step": 2783
+ },
+ {
+ "epoch": 38.67015706806283,
+ "grad_norm": 0.05502549931406975,
+ "learning_rate": 8.23454571240605e-05,
+ "loss": 3.584887981414795,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68411867364747,
+ "grad_norm": 0.056419193744659424,
+ "learning_rate": 8.216710553194416e-05,
+ "loss": 3.5797278881073,
+ "step": 2785
+ },
+ {
+ "epoch": 38.69808027923211,
+ "grad_norm": 0.05669495090842247,
+ "learning_rate": 8.198893186090222e-05,
+ "loss": 3.604696273803711,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712041884816756,
+ "grad_norm": 0.05554523691534996,
+ "learning_rate": 8.181093625448585e-05,
+ "loss": 3.591275215148926,
+ "step": 2787
+ },
+ {
+ "epoch": 38.726003490401396,
+ "grad_norm": 0.05579648166894913,
+ "learning_rate": 8.163311885610299e-05,
+ "loss": 3.5911900997161865,
+ "step": 2788
+ },
+ {
+ "epoch": 38.73996509598604,
+ "grad_norm": 0.05564593896269798,
+ "learning_rate": 8.145547980901791e-05,
+ "loss": 3.6048314571380615,
+ "step": 2789
+ },
+ {
+ "epoch": 38.753926701570684,
+ "grad_norm": 0.05567439645528793,
+ "learning_rate": 8.127801925635126e-05,
+ "loss": 3.564915180206299,
+ "step": 2790
+ },
+ {
+ "epoch": 38.767888307155324,
+ "grad_norm": 0.055503327399492264,
+ "learning_rate": 8.110073734107954e-05,
+ "loss": 3.574024200439453,
+ "step": 2791
+ },
+ {
+ "epoch": 38.781849912739965,
+ "grad_norm": 0.05515526980161667,
+ "learning_rate": 8.092363420603584e-05,
+ "loss": 3.6017959117889404,
+ "step": 2792
+ },
+ {
+ "epoch": 38.795811518324605,
+ "grad_norm": 0.05573352053761482,
+ "learning_rate": 8.074670999390896e-05,
+ "loss": 3.5869383811950684,
+ "step": 2793
+ },
+ {
+ "epoch": 38.80977312390925,
+ "grad_norm": 0.05688846483826637,
+ "learning_rate": 8.056996484724342e-05,
+ "loss": 3.612272024154663,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82373472949389,
+ "grad_norm": 0.05551573261618614,
+ "learning_rate": 8.039339890843958e-05,
+ "loss": 3.5818700790405273,
+ "step": 2795
+ },
+ {
+ "epoch": 38.83769633507853,
+ "grad_norm": 0.05531308427453041,
+ "learning_rate": 8.02170123197535e-05,
+ "loss": 3.558943748474121,
+ "step": 2796
+ },
+ {
+ "epoch": 38.85165794066317,
+ "grad_norm": 0.05705123022198677,
+ "learning_rate": 8.004080522329674e-05,
+ "loss": 3.603858709335327,
+ "step": 2797
+ },
+ {
+ "epoch": 38.86561954624782,
+ "grad_norm": 0.055374935269355774,
+ "learning_rate": 7.986477776103601e-05,
+ "loss": 3.5933685302734375,
+ "step": 2798
+ },
+ {
+ "epoch": 38.87958115183246,
+ "grad_norm": 0.056353647261857986,
+ "learning_rate": 7.968893007479343e-05,
+ "loss": 3.598839282989502,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8935427574171,
+ "grad_norm": 0.05722062662243843,
+ "learning_rate": 7.951326230624658e-05,
+ "loss": 3.645443916320801,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90750436300174,
+ "grad_norm": 0.055990997701883316,
+ "learning_rate": 7.933777459692756e-05,
+ "loss": 3.5926828384399414,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92146596858639,
+ "grad_norm": 0.0561661571264267,
+ "learning_rate": 7.916246708822373e-05,
+ "loss": 3.6137099266052246,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93542757417103,
+ "grad_norm": 0.05624176934361458,
+ "learning_rate": 7.898733992137715e-05,
+ "loss": 3.597254991531372,
+ "step": 2803
+ },
+ {
+ "epoch": 38.94938917975567,
+ "grad_norm": 0.05730769410729408,
+ "learning_rate": 7.881239323748475e-05,
+ "loss": 3.595649242401123,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96335078534032,
+ "grad_norm": 0.05613584816455841,
+ "learning_rate": 7.863762717749771e-05,
+ "loss": 3.5786027908325195,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97731239092496,
+ "grad_norm": 0.05605073273181915,
+ "learning_rate": 7.846304188222189e-05,
+ "loss": 3.5973892211914062,
+ "step": 2806
+ },
+ {
+ "epoch": 38.9912739965096,
+ "grad_norm": 0.05703037977218628,
+ "learning_rate": 7.828863749231777e-05,
+ "loss": 3.614419460296631,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.042099807411432266,
+ "learning_rate": 7.811441414829958e-05,
+ "loss": 2.2424778938293457,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 0.620037853717804,
+ "eval_runtime": 59.1566,
+ "eval_samples_per_second": 41.28,
+ "eval_steps_per_second": 0.659,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01396160558464,
+ "grad_norm": 0.05103587359189987,
+ "learning_rate": 7.794037199053598e-05,
+ "loss": 3.556558132171631,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02792321116929,
+ "grad_norm": 0.049501027911901474,
+ "learning_rate": 7.776651115924959e-05,
+ "loss": 3.5505282878875732,
+ "step": 2810
+ },
+ {
+ "epoch": 39.04188481675393,
+ "grad_norm": 0.05127207562327385,
+ "learning_rate": 7.759283179451704e-05,
+ "loss": 3.5475456714630127,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05584642233857,
+ "grad_norm": 0.05074001103639603,
+ "learning_rate": 7.741933403626848e-05,
+ "loss": 3.5623722076416016,
+ "step": 2812
+ },
+ {
+ "epoch": 39.06980802792321,
+ "grad_norm": 0.05134766176342964,
+ "learning_rate": 7.724601802428805e-05,
+ "loss": 3.540435791015625,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083769633507856,
+ "grad_norm": 0.05144895985722542,
+ "learning_rate": 7.707288389821329e-05,
+ "loss": 3.6154632568359375,
+ "step": 2814
+ },
+ {
+ "epoch": 39.0977312390925,
+ "grad_norm": 0.0506402924656868,
+ "learning_rate": 7.689993179753519e-05,
+ "loss": 3.559809684753418,
+ "step": 2815
+ },
+ {
+ "epoch": 39.11169284467714,
+ "grad_norm": 0.05063097923994064,
+ "learning_rate": 7.672716186159829e-05,
+ "loss": 3.562814474105835,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12565445026178,
+ "grad_norm": 0.051375292241573334,
+ "learning_rate": 7.655457422959993e-05,
+ "loss": 3.5684101581573486,
+ "step": 2817
+ },
+ {
+ "epoch": 39.139616055846425,
+ "grad_norm": 0.0503469854593277,
+ "learning_rate": 7.638216904059122e-05,
+ "loss": 3.556791305541992,
+ "step": 2818
+ },
+ {
+ "epoch": 39.153577661431065,
+ "grad_norm": 0.050908755511045456,
+ "learning_rate": 7.620994643347559e-05,
+ "loss": 3.575982093811035,
+ "step": 2819
+ },
+ {
+ "epoch": 39.167539267015705,
+ "grad_norm": 0.049920931458473206,
+ "learning_rate": 7.60379065470098e-05,
+ "loss": 3.5394392013549805,
+ "step": 2820
+ },
+ {
+ "epoch": 39.181500872600346,
+ "grad_norm": 0.052037835121154785,
+ "learning_rate": 7.586604951980326e-05,
+ "loss": 3.5330114364624023,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19546247818499,
+ "grad_norm": 0.05024134740233421,
+ "learning_rate": 7.569437549031813e-05,
+ "loss": 3.555091142654419,
+ "step": 2822
+ },
+ {
+ "epoch": 39.20942408376963,
+ "grad_norm": 0.05144693702459335,
+ "learning_rate": 7.55228845968691e-05,
+ "loss": 3.5827114582061768,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223385689354274,
+ "grad_norm": 0.052199963480234146,
+ "learning_rate": 7.535157697762313e-05,
+ "loss": 3.559880256652832,
+ "step": 2824
+ },
+ {
+ "epoch": 39.23734729493892,
+ "grad_norm": 0.05210355296730995,
+ "learning_rate": 7.518045277059979e-05,
+ "loss": 3.5808815956115723,
+ "step": 2825
+ },
+ {
+ "epoch": 39.25130890052356,
+ "grad_norm": 0.05106557160615921,
+ "learning_rate": 7.500951211367068e-05,
+ "loss": 3.5711770057678223,
+ "step": 2826
+ },
+ {
+ "epoch": 39.2652705061082,
+ "grad_norm": 0.051606904715299606,
+ "learning_rate": 7.483875514455979e-05,
+ "loss": 3.5610320568084717,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27923211169284,
+ "grad_norm": 0.05146080255508423,
+ "learning_rate": 7.466818200084264e-05,
+ "loss": 3.558434247970581,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29319371727749,
+ "grad_norm": 0.05288682132959366,
+ "learning_rate": 7.449779281994712e-05,
+ "loss": 3.6013455390930176,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30715532286213,
+ "grad_norm": 0.05138479545712471,
+ "learning_rate": 7.43275877391528e-05,
+ "loss": 3.5789742469787598,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32111692844677,
+ "grad_norm": 0.052338190376758575,
+ "learning_rate": 7.415756689559061e-05,
+ "loss": 3.5699806213378906,
+ "step": 2831
+ },
+ {
+ "epoch": 39.33507853403141,
+ "grad_norm": 0.051882997155189514,
+ "learning_rate": 7.398773042624341e-05,
+ "loss": 3.54583740234375,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34904013961606,
+ "grad_norm": 0.05100737139582634,
+ "learning_rate": 7.381807846794532e-05,
+ "loss": 3.5269007682800293,
+ "step": 2833
+ },
+ {
+ "epoch": 39.3630017452007,
+ "grad_norm": 0.05167330801486969,
+ "learning_rate": 7.3648611157382e-05,
+ "loss": 3.547883987426758,
+ "step": 2834
+ },
+ {
+ "epoch": 39.37696335078534,
+ "grad_norm": 0.052119553089141846,
+ "learning_rate": 7.34793286310899e-05,
+ "loss": 3.570747137069702,
+ "step": 2835
+ },
+ {
+ "epoch": 39.390924956369986,
+ "grad_norm": 0.05193249508738518,
+ "learning_rate": 7.331023102545716e-05,
+ "loss": 3.5619454383850098,
+ "step": 2836
+ },
+ {
+ "epoch": 39.404886561954626,
+ "grad_norm": 0.05142345651984215,
+ "learning_rate": 7.314131847672268e-05,
+ "loss": 3.568817138671875,
+ "step": 2837
+ },
+ {
+ "epoch": 39.41884816753927,
+ "grad_norm": 0.05171925202012062,
+ "learning_rate": 7.297259112097608e-05,
+ "loss": 3.5891952514648438,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43280977312391,
+ "grad_norm": 0.051284465938806534,
+ "learning_rate": 7.280404909415801e-05,
+ "loss": 3.56705379486084,
+ "step": 2839
+ },
+ {
+ "epoch": 39.446771378708554,
+ "grad_norm": 0.05284653604030609,
+ "learning_rate": 7.263569253205973e-05,
+ "loss": 3.5296974182128906,
+ "step": 2840
+ },
+ {
+ "epoch": 39.460732984293195,
+ "grad_norm": 0.05274311453104019,
+ "learning_rate": 7.24675215703232e-05,
+ "loss": 3.5653305053710938,
+ "step": 2841
+ },
+ {
+ "epoch": 39.474694589877835,
+ "grad_norm": 0.05247433856129646,
+ "learning_rate": 7.229953634444055e-05,
+ "loss": 3.584923267364502,
+ "step": 2842
+ },
+ {
+ "epoch": 39.488656195462475,
+ "grad_norm": 0.05246812105178833,
+ "learning_rate": 7.213173698975444e-05,
+ "loss": 3.604335069656372,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50261780104712,
+ "grad_norm": 0.052823301404714584,
+ "learning_rate": 7.1964123641458e-05,
+ "loss": 3.5415382385253906,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51657940663176,
+ "grad_norm": 0.05378150939941406,
+ "learning_rate": 7.179669643459405e-05,
+ "loss": 3.5816025733947754,
+ "step": 2845
+ },
+ {
+ "epoch": 39.5305410122164,
+ "grad_norm": 0.053023822605609894,
+ "learning_rate": 7.162945550405575e-05,
+ "loss": 3.542984962463379,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544502617801044,
+ "grad_norm": 0.0532502755522728,
+ "learning_rate": 7.146240098458613e-05,
+ "loss": 3.5629453659057617,
+ "step": 2847
+ },
+ {
+ "epoch": 39.55846422338569,
+ "grad_norm": 0.05303101986646652,
+ "learning_rate": 7.129553301077801e-05,
+ "loss": 3.5725502967834473,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57242582897033,
+ "grad_norm": 0.052930813282728195,
+ "learning_rate": 7.112885171707378e-05,
+ "loss": 3.5748348236083984,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58638743455497,
+ "grad_norm": 0.0558759830892086,
+ "learning_rate": 7.096235723776562e-05,
+ "loss": 3.569337844848633,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60034904013962,
+ "grad_norm": 0.05266093835234642,
+ "learning_rate": 7.079604970699514e-05,
+ "loss": 3.579071044921875,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61431064572426,
+ "grad_norm": 0.0530242957174778,
+ "learning_rate": 7.062992925875332e-05,
+ "loss": 3.569580078125,
+ "step": 2852
+ },
+ {
+ "epoch": 39.6282722513089,
+ "grad_norm": 0.05434603616595268,
+ "learning_rate": 7.046399602688045e-05,
+ "loss": 3.590569496154785,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64223385689354,
+ "grad_norm": 0.05442731827497482,
+ "learning_rate": 7.029825014506572e-05,
+ "loss": 3.584219455718994,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65619546247819,
+ "grad_norm": 0.05413390323519707,
+ "learning_rate": 7.013269174684795e-05,
+ "loss": 3.5826475620269775,
+ "step": 2855
+ },
+ {
+ "epoch": 39.67015706806283,
+ "grad_norm": 0.053916264325380325,
+ "learning_rate": 6.996732096561422e-05,
+ "loss": 3.590533971786499,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68411867364747,
+ "grad_norm": 0.05374522507190704,
+ "learning_rate": 6.980213793460092e-05,
+ "loss": 3.5486674308776855,
+ "step": 2857
+ },
+ {
+ "epoch": 39.69808027923211,
+ "grad_norm": 0.05496302619576454,
+ "learning_rate": 6.963714278689304e-05,
+ "loss": 3.585171699523926,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712041884816756,
+ "grad_norm": 0.05460911989212036,
+ "learning_rate": 6.947233565542417e-05,
+ "loss": 3.545654296875,
+ "step": 2859
+ },
+ {
+ "epoch": 39.726003490401396,
+ "grad_norm": 0.055129971355199814,
+ "learning_rate": 6.930771667297653e-05,
+ "loss": 3.5461995601654053,
+ "step": 2860
+ },
+ {
+ "epoch": 39.73996509598604,
+ "grad_norm": 0.05326374992728233,
+ "learning_rate": 6.914328597218042e-05,
+ "loss": 3.5750136375427246,
+ "step": 2861
+ },
+ {
+ "epoch": 39.753926701570684,
+ "grad_norm": 0.05515550076961517,
+ "learning_rate": 6.8979043685515e-05,
+ "loss": 3.6057279109954834,
+ "step": 2862
+ },
+ {
+ "epoch": 39.767888307155324,
+ "grad_norm": 0.05500713735818863,
+ "learning_rate": 6.881498994530708e-05,
+ "loss": 3.5784642696380615,
+ "step": 2863
+ },
+ {
+ "epoch": 39.781849912739965,
+ "grad_norm": 0.05474477633833885,
+ "learning_rate": 6.865112488373186e-05,
+ "loss": 3.562279224395752,
+ "step": 2864
+ },
+ {
+ "epoch": 39.795811518324605,
+ "grad_norm": 0.05467807129025459,
+ "learning_rate": 6.84874486328125e-05,
+ "loss": 3.582568645477295,
+ "step": 2865
+ },
+ {
+ "epoch": 39.80977312390925,
+ "grad_norm": 0.05313905328512192,
+ "learning_rate": 6.832396132441993e-05,
+ "loss": 3.5913379192352295,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82373472949389,
+ "grad_norm": 0.05476069450378418,
+ "learning_rate": 6.816066309027309e-05,
+ "loss": 3.588045835494995,
+ "step": 2867
+ },
+ {
+ "epoch": 39.83769633507853,
+ "grad_norm": 0.05372064188122749,
+ "learning_rate": 6.799755406193815e-05,
+ "loss": 3.5555458068847656,
+ "step": 2868
+ },
+ {
+ "epoch": 39.85165794066317,
+ "grad_norm": 0.05371057987213135,
+ "learning_rate": 6.783463437082929e-05,
+ "loss": 3.5642948150634766,
+ "step": 2869
+ },
+ {
+ "epoch": 39.86561954624782,
+ "grad_norm": 0.05320272222161293,
+ "learning_rate": 6.767190414820792e-05,
+ "loss": 3.5728759765625,
+ "step": 2870
+ },
+ {
+ "epoch": 39.87958115183246,
+ "grad_norm": 0.054637160152196884,
+ "learning_rate": 6.750936352518284e-05,
+ "loss": 3.5528564453125,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8935427574171,
+ "grad_norm": 0.05465191975235939,
+ "learning_rate": 6.734701263271011e-05,
+ "loss": 3.5658421516418457,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90750436300174,
+ "grad_norm": 0.05452556163072586,
+ "learning_rate": 6.718485160159289e-05,
+ "loss": 3.5749800205230713,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92146596858639,
+ "grad_norm": 0.05434390529990196,
+ "learning_rate": 6.702288056248145e-05,
+ "loss": 3.5784506797790527,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93542757417103,
+ "grad_norm": 0.05458179861307144,
+ "learning_rate": 6.686109964587285e-05,
+ "loss": 3.5742459297180176,
+ "step": 2875
+ },
+ {
+ "epoch": 39.94938917975567,
+ "grad_norm": 0.05434086546301842,
+ "learning_rate": 6.669950898211106e-05,
+ "loss": 3.5479142665863037,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96335078534032,
+ "grad_norm": 0.054262660443782806,
+ "learning_rate": 6.653810870138676e-05,
+ "loss": 3.5872840881347656,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97731239092496,
+ "grad_norm": 0.05524756386876106,
+ "learning_rate": 6.637689893373729e-05,
+ "loss": 3.569241523742676,
+ "step": 2878
+ },
+ {
+ "epoch": 39.9912739965096,
+ "grad_norm": 0.05461305379867554,
+ "learning_rate": 6.621587980904651e-05,
+ "loss": 3.579705238342285,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.04095758497714996,
+ "learning_rate": 6.605505145704437e-05,
+ "loss": 2.2304811477661133,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 0.6212776303291321,
+ "eval_runtime": 58.55,
+ "eval_samples_per_second": 41.708,
+ "eval_steps_per_second": 0.666,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01396160558464,
+ "grad_norm": 0.0490451343357563,
+ "learning_rate": 6.58944140073077e-05,
+ "loss": 3.56355357170105,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02792321116929,
+ "grad_norm": 0.04885198175907135,
+ "learning_rate": 6.573396758925891e-05,
+ "loss": 3.563727855682373,
+ "step": 2882
+ },
+ {
+ "epoch": 40.04188481675393,
+ "grad_norm": 0.04925692081451416,
+ "learning_rate": 6.557371233216692e-05,
+ "loss": 3.5260658264160156,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05584642233857,
+ "grad_norm": 0.048543550074100494,
+ "learning_rate": 6.541364836514646e-05,
+ "loss": 3.520245313644409,
+ "step": 2884
+ },
+ {
+ "epoch": 40.06980802792321,
+ "grad_norm": 0.04830699786543846,
+ "learning_rate": 6.525377581715829e-05,
+ "loss": 3.5621728897094727,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083769633507856,
+ "grad_norm": 0.050245992839336395,
+ "learning_rate": 6.509409481700864e-05,
+ "loss": 3.5443291664123535,
+ "step": 2886
+ },
+ {
+ "epoch": 40.0977312390925,
+ "grad_norm": 0.050024352967739105,
+ "learning_rate": 6.49346054933496e-05,
+ "loss": 3.5327181816101074,
+ "step": 2887
+ },
+ {
+ "epoch": 40.11169284467714,
+ "grad_norm": 0.05024973675608635,
+ "learning_rate": 6.477530797467911e-05,
+ "loss": 3.547147035598755,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12565445026178,
+ "grad_norm": 0.049933336675167084,
+ "learning_rate": 6.461620238934006e-05,
+ "loss": 3.544022560119629,
+ "step": 2889
+ },
+ {
+ "epoch": 40.139616055846425,
+ "grad_norm": 0.0494471937417984,
+ "learning_rate": 6.445728886552109e-05,
+ "loss": 3.5606048107147217,
+ "step": 2890
+ },
+ {
+ "epoch": 40.153577661431065,
+ "grad_norm": 0.05088911950588226,
+ "learning_rate": 6.429856753125573e-05,
+ "loss": 3.5984342098236084,
+ "step": 2891
+ },
+ {
+ "epoch": 40.167539267015705,
+ "grad_norm": 0.05072201415896416,
+ "learning_rate": 6.414003851442318e-05,
+ "loss": 3.546821117401123,
+ "step": 2892
+ },
+ {
+ "epoch": 40.181500872600346,
+ "grad_norm": 0.04955511912703514,
+ "learning_rate": 6.398170194274722e-05,
+ "loss": 3.531683921813965,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19546247818499,
+ "grad_norm": 0.05047255754470825,
+ "learning_rate": 6.38235579437968e-05,
+ "loss": 3.540769100189209,
+ "step": 2894
+ },
+ {
+ "epoch": 40.20942408376963,
+ "grad_norm": 0.04991615191102028,
+ "learning_rate": 6.366560664498572e-05,
+ "loss": 3.5645358562469482,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223385689354274,
+ "grad_norm": 0.04939914494752884,
+ "learning_rate": 6.350784817357242e-05,
+ "loss": 3.500366687774658,
+ "step": 2896
+ },
+ {
+ "epoch": 40.23734729493892,
+ "grad_norm": 0.05038188025355339,
+ "learning_rate": 6.335028265666023e-05,
+ "loss": 3.547734022140503,
+ "step": 2897
+ },
+ {
+ "epoch": 40.25130890052356,
+ "grad_norm": 0.04952634498476982,
+ "learning_rate": 6.319291022119653e-05,
+ "loss": 3.525299549102783,
+ "step": 2898
+ },
+ {
+ "epoch": 40.2652705061082,
+ "grad_norm": 0.050515759736299515,
+ "learning_rate": 6.303573099397378e-05,
+ "loss": 3.5263519287109375,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27923211169284,
+ "grad_norm": 0.05072543025016785,
+ "learning_rate": 6.287874510162821e-05,
+ "loss": 3.5518910884857178,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29319371727749,
+ "grad_norm": 0.04986647143959999,
+ "learning_rate": 6.272195267064063e-05,
+ "loss": 3.5375378131866455,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30715532286213,
+ "grad_norm": 0.05132763460278511,
+ "learning_rate": 6.25653538273358e-05,
+ "loss": 3.561056613922119,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32111692844677,
+ "grad_norm": 0.050235629081726074,
+ "learning_rate": 6.240894869788267e-05,
+ "loss": 3.5394344329833984,
+ "step": 2903
+ },
+ {
+ "epoch": 40.33507853403141,
+ "grad_norm": 0.0514477975666523,
+ "learning_rate": 6.225273740829404e-05,
+ "loss": 3.5648441314697266,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34904013961606,
+ "grad_norm": 0.05128472298383713,
+ "learning_rate": 6.209672008442635e-05,
+ "loss": 3.5410401821136475,
+ "step": 2905
+ },
+ {
+ "epoch": 40.3630017452007,
+ "grad_norm": 0.05111101642251015,
+ "learning_rate": 6.19408968519801e-05,
+ "loss": 3.549811840057373,
+ "step": 2906
+ },
+ {
+ "epoch": 40.37696335078534,
+ "grad_norm": 0.05159356817603111,
+ "learning_rate": 6.178526783649916e-05,
+ "loss": 3.5218734741210938,
+ "step": 2907
+ },
+ {
+ "epoch": 40.390924956369986,
+ "grad_norm": 0.049884531646966934,
+ "learning_rate": 6.162983316337109e-05,
+ "loss": 3.5478882789611816,
+ "step": 2908
+ },
+ {
+ "epoch": 40.404886561954626,
+ "grad_norm": 0.051223963499069214,
+ "learning_rate": 6.147459295782676e-05,
+ "loss": 3.5509533882141113,
+ "step": 2909
+ },
+ {
+ "epoch": 40.41884816753927,
+ "grad_norm": 0.05251265689730644,
+ "learning_rate": 6.131954734494045e-05,
+ "loss": 3.5500032901763916,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43280977312391,
+ "grad_norm": 0.05150502547621727,
+ "learning_rate": 6.11646964496296e-05,
+ "loss": 3.551039218902588,
+ "step": 2911
+ },
+ {
+ "epoch": 40.446771378708554,
+ "grad_norm": 0.050975680351257324,
+ "learning_rate": 6.101004039665471e-05,
+ "loss": 3.5218753814697266,
+ "step": 2912
+ },
+ {
+ "epoch": 40.460732984293195,
+ "grad_norm": 0.052832040935754776,
+ "learning_rate": 6.085557931061937e-05,
+ "loss": 3.5624842643737793,
+ "step": 2913
+ },
+ {
+ "epoch": 40.474694589877835,
+ "grad_norm": 0.051763877272605896,
+ "learning_rate": 6.070131331597015e-05,
+ "loss": 3.572378158569336,
+ "step": 2914
+ },
+ {
+ "epoch": 40.488656195462475,
+ "grad_norm": 0.04998628795146942,
+ "learning_rate": 6.054724253699636e-05,
+ "loss": 3.5298023223876953,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50261780104712,
+ "grad_norm": 0.05136828497052193,
+ "learning_rate": 6.03933670978301e-05,
+ "loss": 3.541879653930664,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51657940663176,
+ "grad_norm": 0.05399318039417267,
+ "learning_rate": 6.0239687122445826e-05,
+ "loss": 3.5703022480010986,
+ "step": 2917
+ },
+ {
+ "epoch": 40.5305410122164,
+ "grad_norm": 0.050659872591495514,
+ "learning_rate": 6.0086202734661e-05,
+ "loss": 3.555227518081665,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544502617801044,
+ "grad_norm": 0.051574256271123886,
+ "learning_rate": 5.9932914058135004e-05,
+ "loss": 3.549437999725342,
+ "step": 2919
+ },
+ {
+ "epoch": 40.55846422338569,
+ "grad_norm": 0.05199757218360901,
+ "learning_rate": 5.977982121636984e-05,
+ "loss": 3.564350128173828,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57242582897033,
+ "grad_norm": 0.05227932706475258,
+ "learning_rate": 5.962692433270962e-05,
+ "loss": 3.538482666015625,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58638743455497,
+ "grad_norm": 0.05216033384203911,
+ "learning_rate": 5.9474223530340634e-05,
+ "loss": 3.5592684745788574,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60034904013962,
+ "grad_norm": 0.051728200167417526,
+ "learning_rate": 5.932171893229124e-05,
+ "loss": 3.516632556915283,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61431064572426,
+ "grad_norm": 0.05290914699435234,
+ "learning_rate": 5.916941066143137e-05,
+ "loss": 3.5496959686279297,
+ "step": 2924
+ },
+ {
+ "epoch": 40.6282722513089,
+ "grad_norm": 0.053392428904771805,
+ "learning_rate": 5.9017298840473364e-05,
+ "loss": 3.548393964767456,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64223385689354,
+ "grad_norm": 0.052064698189496994,
+ "learning_rate": 5.8865383591970775e-05,
+ "loss": 3.5496573448181152,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65619546247819,
+ "grad_norm": 0.05260622501373291,
+ "learning_rate": 5.8713665038319e-05,
+ "loss": 3.556190252304077,
+ "step": 2927
+ },
+ {
+ "epoch": 40.67015706806283,
+ "grad_norm": 0.052858006209135056,
+ "learning_rate": 5.856214330175498e-05,
+ "loss": 3.550811290740967,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68411867364747,
+ "grad_norm": 0.052420660853385925,
+ "learning_rate": 5.841081850435704e-05,
+ "loss": 3.536261796951294,
+ "step": 2929
+ },
+ {
+ "epoch": 40.69808027923211,
+ "grad_norm": 0.053714022040367126,
+ "learning_rate": 5.825969076804488e-05,
+ "loss": 3.587533473968506,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712041884816756,
+ "grad_norm": 0.05208775773644447,
+ "learning_rate": 5.810876021457925e-05,
+ "loss": 3.540727376937866,
+ "step": 2931
+ },
+ {
+ "epoch": 40.726003490401396,
+ "grad_norm": 0.05275597423315048,
+ "learning_rate": 5.795802696556222e-05,
+ "loss": 3.574899196624756,
+ "step": 2932
+ },
+ {
+ "epoch": 40.73996509598604,
+ "grad_norm": 0.05299631133675575,
+ "learning_rate": 5.78074911424369e-05,
+ "loss": 3.530336380004883,
+ "step": 2933
+ },
+ {
+ "epoch": 40.753926701570684,
+ "grad_norm": 0.054256338626146317,
+ "learning_rate": 5.765715286648734e-05,
+ "loss": 3.5379605293273926,
+ "step": 2934
+ },
+ {
+ "epoch": 40.767888307155324,
+ "grad_norm": 0.05158054083585739,
+ "learning_rate": 5.7507012258838154e-05,
+ "loss": 3.562894821166992,
+ "step": 2935
+ },
+ {
+ "epoch": 40.781849912739965,
+ "grad_norm": 0.05369063839316368,
+ "learning_rate": 5.7357069440455095e-05,
+ "loss": 3.529047966003418,
+ "step": 2936
+ },
+ {
+ "epoch": 40.795811518324605,
+ "grad_norm": 0.05377575755119324,
+ "learning_rate": 5.7207324532144454e-05,
+ "loss": 3.5563199520111084,
+ "step": 2937
+ },
+ {
+ "epoch": 40.80977312390925,
+ "grad_norm": 0.05389560014009476,
+ "learning_rate": 5.7057777654552785e-05,
+ "loss": 3.5682358741760254,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82373472949389,
+ "grad_norm": 0.053750719875097275,
+ "learning_rate": 5.690842892816741e-05,
+ "loss": 3.5546717643737793,
+ "step": 2939
+ },
+ {
+ "epoch": 40.83769633507853,
+ "grad_norm": 0.053277388215065,
+ "learning_rate": 5.675927847331593e-05,
+ "loss": 3.5211615562438965,
+ "step": 2940
+ },
+ {
+ "epoch": 40.85165794066317,
+ "grad_norm": 0.05360420420765877,
+ "learning_rate": 5.661032641016619e-05,
+ "loss": 3.5586609840393066,
+ "step": 2941
+ },
+ {
+ "epoch": 40.86561954624782,
+ "grad_norm": 0.05353206396102905,
+ "learning_rate": 5.6461572858726086e-05,
+ "loss": 3.5891380310058594,
+ "step": 2942
+ },
+ {
+ "epoch": 40.87958115183246,
+ "grad_norm": 0.05361996591091156,
+ "learning_rate": 5.6313017938843595e-05,
+ "loss": 3.5477938652038574,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8935427574171,
+ "grad_norm": 0.053566232323646545,
+ "learning_rate": 5.6164661770207026e-05,
+ "loss": 3.560500144958496,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90750436300174,
+ "grad_norm": 0.05357549339532852,
+ "learning_rate": 5.6016504472344e-05,
+ "loss": 3.557918071746826,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92146596858639,
+ "grad_norm": 0.05396264046430588,
+ "learning_rate": 5.5868546164622284e-05,
+ "loss": 3.5511131286621094,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93542757417103,
+ "grad_norm": 0.05372125282883644,
+ "learning_rate": 5.572078696624917e-05,
+ "loss": 3.529221534729004,
+ "step": 2947
+ },
+ {
+ "epoch": 40.94938917975567,
+ "grad_norm": 0.05311017110943794,
+ "learning_rate": 5.557322699627168e-05,
+ "loss": 3.5863747596740723,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96335078534032,
+ "grad_norm": 0.05271926149725914,
+ "learning_rate": 5.542586637357607e-05,
+ "loss": 3.5291237831115723,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97731239092496,
+ "grad_norm": 0.05377885699272156,
+ "learning_rate": 5.527870521688817e-05,
+ "loss": 3.5661916732788086,
+ "step": 2950
+ },
+ {
+ "epoch": 40.9912739965096,
+ "grad_norm": 0.05412346124649048,
+ "learning_rate": 5.513174364477311e-05,
+ "loss": 3.5414018630981445,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.041567444801330566,
+ "learning_rate": 5.498498177563518e-05,
+ "loss": 2.210968494415283,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 0.6227477192878723,
+ "eval_runtime": 59.639,
+ "eval_samples_per_second": 40.946,
+ "eval_steps_per_second": 0.654,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01396160558464,
+ "grad_norm": 0.04824776574969292,
+ "learning_rate": 5.4838419727717716e-05,
+ "loss": 3.5330162048339844,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02792321116929,
+ "grad_norm": 0.04834216460585594,
+ "learning_rate": 5.4692057619103144e-05,
+ "loss": 3.506112575531006,
+ "step": 2954
+ },
+ {
+ "epoch": 41.04188481675393,
+ "grad_norm": 0.048684537410736084,
+ "learning_rate": 5.454589556771286e-05,
+ "loss": 3.552931785583496,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05584642233857,
+ "grad_norm": 0.04952389374375343,
+ "learning_rate": 5.439993369130681e-05,
+ "loss": 3.522508144378662,
+ "step": 2956
+ },
+ {
+ "epoch": 41.06980802792321,
+ "grad_norm": 0.047509536147117615,
+ "learning_rate": 5.4254172107483924e-05,
+ "loss": 3.521543502807617,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083769633507856,
+ "grad_norm": 0.04766285791993141,
+ "learning_rate": 5.410861093368169e-05,
+ "loss": 3.5325450897216797,
+ "step": 2958
+ },
+ {
+ "epoch": 41.0977312390925,
+ "grad_norm": 0.04850609600543976,
+ "learning_rate": 5.3963250287176094e-05,
+ "loss": 3.5616042613983154,
+ "step": 2959
+ },
+ {
+ "epoch": 41.11169284467714,
+ "grad_norm": 0.04834820702672005,
+ "learning_rate": 5.381809028508168e-05,
+ "loss": 3.5167665481567383,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12565445026178,
+ "grad_norm": 0.04798751696944237,
+ "learning_rate": 5.367313104435104e-05,
+ "loss": 3.5004749298095703,
+ "step": 2961
+ },
+ {
+ "epoch": 41.139616055846425,
+ "grad_norm": 0.047878291457891464,
+ "learning_rate": 5.352837268177544e-05,
+ "loss": 3.5392799377441406,
+ "step": 2962
+ },
+ {
+ "epoch": 41.153577661431065,
+ "grad_norm": 0.048691265285015106,
+ "learning_rate": 5.3383815313983916e-05,
+ "loss": 3.525770664215088,
+ "step": 2963
+ },
+ {
+ "epoch": 41.167539267015705,
+ "grad_norm": 0.048624299466609955,
+ "learning_rate": 5.323945905744379e-05,
+ "loss": 3.5122787952423096,
+ "step": 2964
+ },
+ {
+ "epoch": 41.181500872600346,
+ "grad_norm": 0.0488092415034771,
+ "learning_rate": 5.3095304028460284e-05,
+ "loss": 3.50136661529541,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19546247818499,
+ "grad_norm": 0.04929783567786217,
+ "learning_rate": 5.2951350343176495e-05,
+ "loss": 3.51039981842041,
+ "step": 2966
+ },
+ {
+ "epoch": 41.20942408376963,
+ "grad_norm": 0.049221813678741455,
+ "learning_rate": 5.2807598117573384e-05,
+ "loss": 3.530416965484619,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223385689354274,
+ "grad_norm": 0.04767163097858429,
+ "learning_rate": 5.266404746746939e-05,
+ "loss": 3.5467047691345215,
+ "step": 2968
+ },
+ {
+ "epoch": 41.23734729493892,
+ "grad_norm": 0.048979729413986206,
+ "learning_rate": 5.252069850852072e-05,
+ "loss": 3.557248115539551,
+ "step": 2969
+ },
+ {
+ "epoch": 41.25130890052356,
+ "grad_norm": 0.048910439014434814,
+ "learning_rate": 5.237755135622107e-05,
+ "loss": 3.5098860263824463,
+ "step": 2970
+ },
+ {
+ "epoch": 41.2652705061082,
+ "grad_norm": 0.048609018325805664,
+ "learning_rate": 5.223460612590159e-05,
+ "loss": 3.5389151573181152,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27923211169284,
+ "grad_norm": 0.04871147871017456,
+ "learning_rate": 5.209186293273044e-05,
+ "loss": 3.5385420322418213,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29319371727749,
+ "grad_norm": 0.04857845976948738,
+ "learning_rate": 5.1949321891713446e-05,
+ "loss": 3.507431983947754,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30715532286213,
+ "grad_norm": 0.05005535110831261,
+ "learning_rate": 5.180698311769338e-05,
+ "loss": 3.5485854148864746,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32111692844677,
+ "grad_norm": 0.05051431804895401,
+ "learning_rate": 5.166484672534983e-05,
+ "loss": 3.5140175819396973,
+ "step": 2975
+ },
+ {
+ "epoch": 41.33507853403141,
+ "grad_norm": 0.04972303658723831,
+ "learning_rate": 5.152291282919961e-05,
+ "loss": 3.541937828063965,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34904013961606,
+ "grad_norm": 0.04871612414717674,
+ "learning_rate": 5.138118154359632e-05,
+ "loss": 3.543015956878662,
+ "step": 2977
+ },
+ {
+ "epoch": 41.3630017452007,
+ "grad_norm": 0.05032794922590256,
+ "learning_rate": 5.1239652982730335e-05,
+ "loss": 3.4958419799804688,
+ "step": 2978
+ },
+ {
+ "epoch": 41.37696335078534,
+ "grad_norm": 0.05026672035455704,
+ "learning_rate": 5.1098327260628466e-05,
+ "loss": 3.519773244857788,
+ "step": 2979
+ },
+ {
+ "epoch": 41.390924956369986,
+ "grad_norm": 0.04959705099463463,
+ "learning_rate": 5.09572044911545e-05,
+ "loss": 3.508209466934204,
+ "step": 2980
+ },
+ {
+ "epoch": 41.404886561954626,
+ "grad_norm": 0.04974748566746712,
+ "learning_rate": 5.0816284788008527e-05,
+ "loss": 3.5200767517089844,
+ "step": 2981
+ },
+ {
+ "epoch": 41.41884816753927,
+ "grad_norm": 0.050430137664079666,
+ "learning_rate": 5.067556826472684e-05,
+ "loss": 3.528503656387329,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43280977312391,
+ "grad_norm": 0.05002054572105408,
+ "learning_rate": 5.053505503468228e-05,
+ "loss": 3.544261932373047,
+ "step": 2983
+ },
+ {
+ "epoch": 41.446771378708554,
+ "grad_norm": 0.050833702087402344,
+ "learning_rate": 5.0394745211083785e-05,
+ "loss": 3.528820276260376,
+ "step": 2984
+ },
+ {
+ "epoch": 41.460732984293195,
+ "grad_norm": 0.05177270993590355,
+ "learning_rate": 5.025463890697655e-05,
+ "loss": 3.5320916175842285,
+ "step": 2985
+ },
+ {
+ "epoch": 41.474694589877835,
+ "grad_norm": 0.05068309232592583,
+ "learning_rate": 5.011473623524159e-05,
+ "loss": 3.535667896270752,
+ "step": 2986
+ },
+ {
+ "epoch": 41.488656195462475,
+ "grad_norm": 0.05038442835211754,
+ "learning_rate": 4.9975037308595864e-05,
+ "loss": 3.544569969177246,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50261780104712,
+ "grad_norm": 0.051365263760089874,
+ "learning_rate": 4.983554223959257e-05,
+ "loss": 3.5355050563812256,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51657940663176,
+ "grad_norm": 0.05044831335544586,
+ "learning_rate": 4.9696251140620134e-05,
+ "loss": 3.511427879333496,
+ "step": 2989
+ },
+ {
+ "epoch": 41.5305410122164,
+ "grad_norm": 0.050490397959947586,
+ "learning_rate": 4.9557164123902924e-05,
+ "loss": 3.544853687286377,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544502617801044,
+ "grad_norm": 0.05034386366605759,
+ "learning_rate": 4.9418281301500844e-05,
+ "loss": 3.5631227493286133,
+ "step": 2991
+ },
+ {
+ "epoch": 41.55846422338569,
+ "grad_norm": 0.0505383126437664,
+ "learning_rate": 4.9279602785309365e-05,
+ "loss": 3.5400047302246094,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57242582897033,
+ "grad_norm": 0.05098932981491089,
+ "learning_rate": 4.914112868705911e-05,
+ "loss": 3.5390548706054688,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58638743455497,
+ "grad_norm": 0.05151797831058502,
+ "learning_rate": 4.900285911831624e-05,
+ "loss": 3.535597324371338,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60034904013962,
+ "grad_norm": 0.05120394378900528,
+ "learning_rate": 4.886479419048201e-05,
+ "loss": 3.5225820541381836,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61431064572426,
+ "grad_norm": 0.05033888295292854,
+ "learning_rate": 4.872693401479292e-05,
+ "loss": 3.4912185668945312,
+ "step": 2996
+ },
+ {
+ "epoch": 41.6282722513089,
+ "grad_norm": 0.05037662014365196,
+ "learning_rate": 4.858927870232039e-05,
+ "loss": 3.5338728427886963,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64223385689354,
+ "grad_norm": 0.05047066509723663,
+ "learning_rate": 4.845182836397068e-05,
+ "loss": 3.537113666534424,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65619546247819,
+ "grad_norm": 0.051437340676784515,
+ "learning_rate": 4.831458311048532e-05,
+ "loss": 3.537376880645752,
+ "step": 2999
+ },
+ {
+ "epoch": 41.67015706806283,
+ "grad_norm": 0.05054743215441704,
+ "learning_rate": 4.817754305244008e-05,
+ "loss": 3.558231830596924,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68411867364747,
+ "grad_norm": 0.050861239433288574,
+ "learning_rate": 4.804070830024577e-05,
+ "loss": 3.5442352294921875,
+ "step": 3001
+ },
+ {
+ "epoch": 41.69808027923211,
+ "grad_norm": 0.05110339820384979,
+ "learning_rate": 4.7904078964147654e-05,
+ "loss": 3.537139892578125,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712041884816756,
+ "grad_norm": 0.05177166685461998,
+ "learning_rate": 4.776765515422557e-05,
+ "loss": 3.5539965629577637,
+ "step": 3003
+ },
+ {
+ "epoch": 41.726003490401396,
+ "grad_norm": 0.051353856921195984,
+ "learning_rate": 4.763143698039371e-05,
+ "loss": 3.5276026725769043,
+ "step": 3004
+ },
+ {
+ "epoch": 41.73996509598604,
+ "grad_norm": 0.05237841606140137,
+ "learning_rate": 4.749542455240043e-05,
+ "loss": 3.5436954498291016,
+ "step": 3005
+ },
+ {
+ "epoch": 41.753926701570684,
+ "grad_norm": 0.051556285470724106,
+ "learning_rate": 4.735961797982876e-05,
+ "loss": 3.525407552719116,
+ "step": 3006
+ },
+ {
+ "epoch": 41.767888307155324,
+ "grad_norm": 0.0522376224398613,
+ "learning_rate": 4.7224017372095414e-05,
+ "loss": 3.5100879669189453,
+ "step": 3007
+ },
+ {
+ "epoch": 41.781849912739965,
+ "grad_norm": 0.05480589717626572,
+ "learning_rate": 4.708862283845143e-05,
+ "loss": 3.5373239517211914,
+ "step": 3008
+ },
+ {
+ "epoch": 41.795811518324605,
+ "grad_norm": 0.05243408679962158,
+ "learning_rate": 4.695343448798163e-05,
+ "loss": 3.564474105834961,
+ "step": 3009
+ },
+ {
+ "epoch": 41.80977312390925,
+ "grad_norm": 0.05195145308971405,
+ "learning_rate": 4.681845242960492e-05,
+ "loss": 3.5219478607177734,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82373472949389,
+ "grad_norm": 0.051901329308748245,
+ "learning_rate": 4.668367677207398e-05,
+ "loss": 3.5563879013061523,
+ "step": 3011
+ },
+ {
+ "epoch": 41.83769633507853,
+ "grad_norm": 0.05103657394647598,
+ "learning_rate": 4.654910762397499e-05,
+ "loss": 3.5318007469177246,
+ "step": 3012
+ },
+ {
+ "epoch": 41.85165794066317,
+ "grad_norm": 0.05273298919200897,
+ "learning_rate": 4.6414745093727913e-05,
+ "loss": 3.5337777137756348,
+ "step": 3013
+ },
+ {
+ "epoch": 41.86561954624782,
+ "grad_norm": 0.05128652974963188,
+ "learning_rate": 4.6280589289586255e-05,
+ "loss": 3.523869037628174,
+ "step": 3014
+ },
+ {
+ "epoch": 41.87958115183246,
+ "grad_norm": 0.051192507147789,
+ "learning_rate": 4.614664031963692e-05,
+ "loss": 3.5141286849975586,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8935427574171,
+ "grad_norm": 0.052560918033123016,
+ "learning_rate": 4.601289829180004e-05,
+ "loss": 3.542026996612549,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90750436300174,
+ "grad_norm": 0.05215048789978027,
+ "learning_rate": 4.587936331382934e-05,
+ "loss": 3.5189595222473145,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92146596858639,
+ "grad_norm": 0.05149101838469505,
+ "learning_rate": 4.574603549331151e-05,
+ "loss": 3.559493064880371,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93542757417103,
+ "grad_norm": 0.05164817348122597,
+ "learning_rate": 4.561291493766625e-05,
+ "loss": 3.553772211074829,
+ "step": 3019
+ },
+ {
+ "epoch": 41.94938917975567,
+ "grad_norm": 0.05225958675146103,
+ "learning_rate": 4.5480001754146455e-05,
+ "loss": 3.5521295070648193,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96335078534032,
+ "grad_norm": 0.051618315279483795,
+ "learning_rate": 4.5347296049837875e-05,
+ "loss": 3.5303773880004883,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97731239092496,
+ "grad_norm": 0.051267798990011215,
+ "learning_rate": 4.52147979316592e-05,
+ "loss": 3.5165226459503174,
+ "step": 3022
+ },
+ {
+ "epoch": 41.9912739965096,
+ "grad_norm": 0.05264551192522049,
+ "learning_rate": 4.5082507506361574e-05,
+ "loss": 3.5303573608398438,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.03939991816878319,
+ "learning_rate": 4.495042488052904e-05,
+ "loss": 2.205373764038086,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 0.6238990426063538,
+ "eval_runtime": 59.3761,
+ "eval_samples_per_second": 41.128,
+ "eval_steps_per_second": 0.657,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01396160558464,
+ "grad_norm": 0.04582864046096802,
+ "learning_rate": 4.481855016057839e-05,
+ "loss": 3.55916690826416,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02792321116929,
+ "grad_norm": 0.045121654868125916,
+ "learning_rate": 4.468688345275848e-05,
+ "loss": 3.525993585586548,
+ "step": 3026
+ },
+ {
+ "epoch": 42.04188481675393,
+ "grad_norm": 0.045851755887269974,
+ "learning_rate": 4.455542486315086e-05,
+ "loss": 3.536099433898926,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05584642233857,
+ "grad_norm": 0.04772236570715904,
+ "learning_rate": 4.442417449766938e-05,
+ "loss": 3.4842875003814697,
+ "step": 3028
+ },
+ {
+ "epoch": 42.06980802792321,
+ "grad_norm": 0.04591905325651169,
+ "learning_rate": 4.429313246206014e-05,
+ "loss": 3.4977431297302246,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083769633507856,
+ "grad_norm": 0.04562445729970932,
+ "learning_rate": 4.416229886190119e-05,
+ "loss": 3.5109801292419434,
+ "step": 3030
+ },
+ {
+ "epoch": 42.0977312390925,
+ "grad_norm": 0.046725720167160034,
+ "learning_rate": 4.403167380260281e-05,
+ "loss": 3.507308006286621,
+ "step": 3031
+ },
+ {
+ "epoch": 42.11169284467714,
+ "grad_norm": 0.047091979533433914,
+ "learning_rate": 4.3901257389407475e-05,
+ "loss": 3.5205769538879395,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12565445026178,
+ "grad_norm": 0.04644174128770828,
+ "learning_rate": 4.3771049727389075e-05,
+ "loss": 3.519360065460205,
+ "step": 3033
+ },
+ {
+ "epoch": 42.139616055846425,
+ "grad_norm": 0.0471586212515831,
+ "learning_rate": 4.364105092145377e-05,
+ "loss": 3.511544704437256,
+ "step": 3034
+ },
+ {
+ "epoch": 42.153577661431065,
+ "grad_norm": 0.04756023734807968,
+ "learning_rate": 4.3511261076339e-05,
+ "loss": 3.5082921981811523,
+ "step": 3035
+ },
+ {
+ "epoch": 42.167539267015705,
+ "grad_norm": 0.04774578660726547,
+ "learning_rate": 4.3381680296614394e-05,
+ "loss": 3.514894485473633,
+ "step": 3036
+ },
+ {
+ "epoch": 42.181500872600346,
+ "grad_norm": 0.047490909695625305,
+ "learning_rate": 4.3252308686680626e-05,
+ "loss": 3.5040950775146484,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19546247818499,
+ "grad_norm": 0.04813152179121971,
+ "learning_rate": 4.3123146350770146e-05,
+ "loss": 3.5205535888671875,
+ "step": 3038
+ },
+ {
+ "epoch": 42.20942408376963,
+ "grad_norm": 0.047403670847415924,
+ "learning_rate": 4.2994193392946726e-05,
+ "loss": 3.4987292289733887,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223385689354274,
+ "grad_norm": 0.047732383012771606,
+ "learning_rate": 4.286544991710541e-05,
+ "loss": 3.5132713317871094,
+ "step": 3040
+ },
+ {
+ "epoch": 42.23734729493892,
+ "grad_norm": 0.0471004955470562,
+ "learning_rate": 4.2736916026972576e-05,
+ "loss": 3.4943442344665527,
+ "step": 3041
+ },
+ {
+ "epoch": 42.25130890052356,
+ "grad_norm": 0.04769458621740341,
+ "learning_rate": 4.260859182610542e-05,
+ "loss": 3.488661527633667,
+ "step": 3042
+ },
+ {
+ "epoch": 42.2652705061082,
+ "grad_norm": 0.04852530360221863,
+ "learning_rate": 4.2480477417892776e-05,
+ "loss": 3.502333641052246,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27923211169284,
+ "grad_norm": 0.047681331634521484,
+ "learning_rate": 4.235257290555381e-05,
+ "loss": 3.5339419841766357,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29319371727749,
+ "grad_norm": 0.04790053516626358,
+ "learning_rate": 4.222487839213903e-05,
+ "loss": 3.531029462814331,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30715532286213,
+ "grad_norm": 0.04872751981019974,
+ "learning_rate": 4.209739398052956e-05,
+ "loss": 3.5269131660461426,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32111692844677,
+ "grad_norm": 0.0475989505648613,
+ "learning_rate": 4.1970119773437316e-05,
+ "loss": 3.520803213119507,
+ "step": 3047
+ },
+ {
+ "epoch": 42.33507853403141,
+ "grad_norm": 0.047820597887039185,
+ "learning_rate": 4.184305587340483e-05,
+ "loss": 3.516744613647461,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34904013961606,
+ "grad_norm": 0.046507153660058975,
+ "learning_rate": 4.171620238280511e-05,
+ "loss": 3.4996590614318848,
+ "step": 3049
+ },
+ {
+ "epoch": 42.3630017452007,
+ "grad_norm": 0.047835614532232285,
+ "learning_rate": 4.158955940384179e-05,
+ "loss": 3.5125174522399902,
+ "step": 3050
+ },
+ {
+ "epoch": 42.37696335078534,
+ "grad_norm": 0.048144467175006866,
+ "learning_rate": 4.146312703854881e-05,
+ "loss": 3.5353198051452637,
+ "step": 3051
+ },
+ {
+ "epoch": 42.390924956369986,
+ "grad_norm": 0.04883087798953056,
+ "learning_rate": 4.133690538879046e-05,
+ "loss": 3.518767833709717,
+ "step": 3052
+ },
+ {
+ "epoch": 42.404886561954626,
+ "grad_norm": 0.047930024564266205,
+ "learning_rate": 4.1210894556261226e-05,
+ "loss": 3.5251386165618896,
+ "step": 3053
+ },
+ {
+ "epoch": 42.41884816753927,
+ "grad_norm": 0.04800357297062874,
+ "learning_rate": 4.108509464248578e-05,
+ "loss": 3.5164432525634766,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43280977312391,
+ "grad_norm": 0.048447951674461365,
+ "learning_rate": 4.095950574881891e-05,
+ "loss": 3.4852211475372314,
+ "step": 3055
+ },
+ {
+ "epoch": 42.446771378708554,
+ "grad_norm": 0.04907943308353424,
+ "learning_rate": 4.083412797644514e-05,
+ "loss": 3.513625383377075,
+ "step": 3056
+ },
+ {
+ "epoch": 42.460732984293195,
+ "grad_norm": 0.04915059357881546,
+ "learning_rate": 4.0708961426379214e-05,
+ "loss": 3.5324220657348633,
+ "step": 3057
+ },
+ {
+ "epoch": 42.474694589877835,
+ "grad_norm": 0.04837046191096306,
+ "learning_rate": 4.0584006199465484e-05,
+ "loss": 3.513761520385742,
+ "step": 3058
+ },
+ {
+ "epoch": 42.488656195462475,
+ "grad_norm": 0.04837717115879059,
+ "learning_rate": 4.0459262396378165e-05,
+ "loss": 3.4778947830200195,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50261780104712,
+ "grad_norm": 0.048395294696092606,
+ "learning_rate": 4.033473011762116e-05,
+ "loss": 3.5122289657592773,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51657940663176,
+ "grad_norm": 0.04952006787061691,
+ "learning_rate": 4.0210409463527656e-05,
+ "loss": 3.5174784660339355,
+ "step": 3061
+ },
+ {
+ "epoch": 42.5305410122164,
+ "grad_norm": 0.050250094383955,
+ "learning_rate": 4.008630053426082e-05,
+ "loss": 3.481564521789551,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544502617801044,
+ "grad_norm": 0.04852081835269928,
+ "learning_rate": 3.996240342981279e-05,
+ "loss": 3.5288257598876953,
+ "step": 3063
+ },
+ {
+ "epoch": 42.55846422338569,
+ "grad_norm": 0.048682864755392075,
+ "learning_rate": 3.983871825000529e-05,
+ "loss": 3.517641067504883,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57242582897033,
+ "grad_norm": 0.049304574728012085,
+ "learning_rate": 3.971524509448925e-05,
+ "loss": 3.5146350860595703,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58638743455497,
+ "grad_norm": 0.049619562923908234,
+ "learning_rate": 3.959198406274472e-05,
+ "loss": 3.5137906074523926,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60034904013962,
+ "grad_norm": 0.04923011362552643,
+ "learning_rate": 3.946893525408097e-05,
+ "loss": 3.5143535137176514,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61431064572426,
+ "grad_norm": 0.0497874915599823,
+ "learning_rate": 3.934609876763604e-05,
+ "loss": 3.5229015350341797,
+ "step": 3068
+ },
+ {
+ "epoch": 42.6282722513089,
+ "grad_norm": 0.050324805080890656,
+ "learning_rate": 3.922347470237728e-05,
+ "loss": 3.5317938327789307,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64223385689354,
+ "grad_norm": 0.049399957060813904,
+ "learning_rate": 3.910106315710053e-05,
+ "loss": 3.535451650619507,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65619546247819,
+ "grad_norm": 0.04937320575118065,
+ "learning_rate": 3.897886423043061e-05,
+ "loss": 3.510342836380005,
+ "step": 3071
+ },
+ {
+ "epoch": 42.67015706806283,
+ "grad_norm": 0.04943464323878288,
+ "learning_rate": 3.8856878020820975e-05,
+ "loss": 3.53397798538208,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68411867364747,
+ "grad_norm": 0.04982450604438782,
+ "learning_rate": 3.873510462655375e-05,
+ "loss": 3.5372724533081055,
+ "step": 3073
+ },
+ {
+ "epoch": 42.69808027923211,
+ "grad_norm": 0.04879437014460564,
+ "learning_rate": 3.861354414573954e-05,
+ "loss": 3.5235369205474854,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712041884816756,
+ "grad_norm": 0.049403633922338486,
+ "learning_rate": 3.849219667631735e-05,
+ "loss": 3.523263931274414,
+ "step": 3075
+ },
+ {
+ "epoch": 42.726003490401396,
+ "grad_norm": 0.04999256134033203,
+ "learning_rate": 3.8371062316054764e-05,
+ "loss": 3.5183444023132324,
+ "step": 3076
+ },
+ {
+ "epoch": 42.73996509598604,
+ "grad_norm": 0.04949156567454338,
+ "learning_rate": 3.8250141162547466e-05,
+ "loss": 3.5372426509857178,
+ "step": 3077
+ },
+ {
+ "epoch": 42.753926701570684,
+ "grad_norm": 0.05010758712887764,
+ "learning_rate": 3.812943331321956e-05,
+ "loss": 3.494075298309326,
+ "step": 3078
+ },
+ {
+ "epoch": 42.767888307155324,
+ "grad_norm": 0.04966307431459427,
+ "learning_rate": 3.8008938865322955e-05,
+ "loss": 3.518535614013672,
+ "step": 3079
+ },
+ {
+ "epoch": 42.781849912739965,
+ "grad_norm": 0.050056494772434235,
+ "learning_rate": 3.788865791593813e-05,
+ "loss": 3.5020852088928223,
+ "step": 3080
+ },
+ {
+ "epoch": 42.795811518324605,
+ "grad_norm": 0.04934854060411453,
+ "learning_rate": 3.7768590561973064e-05,
+ "loss": 3.5209922790527344,
+ "step": 3081
+ },
+ {
+ "epoch": 42.80977312390925,
+ "grad_norm": 0.049946680665016174,
+ "learning_rate": 3.764873690016394e-05,
+ "loss": 3.517817974090576,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82373472949389,
+ "grad_norm": 0.04958394914865494,
+ "learning_rate": 3.7529097027074654e-05,
+ "loss": 3.5093369483947754,
+ "step": 3083
+ },
+ {
+ "epoch": 42.83769633507853,
+ "grad_norm": 0.04957697167992592,
+ "learning_rate": 3.7409671039096936e-05,
+ "loss": 3.4855117797851562,
+ "step": 3084
+ },
+ {
+ "epoch": 42.85165794066317,
+ "grad_norm": 0.05044497177004814,
+ "learning_rate": 3.7290459032450184e-05,
+ "loss": 3.518657684326172,
+ "step": 3085
+ },
+ {
+ "epoch": 42.86561954624782,
+ "grad_norm": 0.049479082226753235,
+ "learning_rate": 3.717146110318123e-05,
+ "loss": 3.5081276893615723,
+ "step": 3086
+ },
+ {
+ "epoch": 42.87958115183246,
+ "grad_norm": 0.0499165914952755,
+ "learning_rate": 3.7052677347164614e-05,
+ "loss": 3.525416374206543,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8935427574171,
+ "grad_norm": 0.05002507194876671,
+ "learning_rate": 3.693410786010231e-05,
+ "loss": 3.5287156105041504,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90750436300174,
+ "grad_norm": 0.04990137368440628,
+ "learning_rate": 3.6815752737523537e-05,
+ "loss": 3.511509895324707,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92146596858639,
+ "grad_norm": 0.05140728875994682,
+ "learning_rate": 3.6697612074784944e-05,
+ "loss": 3.529341220855713,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93542757417103,
+ "grad_norm": 0.049925290048122406,
+ "learning_rate": 3.657968596707029e-05,
+ "loss": 3.538410186767578,
+ "step": 3091
+ },
+ {
+ "epoch": 42.94938917975567,
+ "grad_norm": 0.04992802441120148,
+ "learning_rate": 3.64619745093906e-05,
+ "loss": 3.502620220184326,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96335078534032,
+ "grad_norm": 0.051207154989242554,
+ "learning_rate": 3.6344477796583714e-05,
+ "loss": 3.50933575630188,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97731239092496,
+ "grad_norm": 0.05062158778309822,
+ "learning_rate": 3.622719592331472e-05,
+ "loss": 3.5215516090393066,
+ "step": 3094
+ },
+ {
+ "epoch": 42.9912739965096,
+ "grad_norm": 0.05091328173875809,
+ "learning_rate": 3.611012898407544e-05,
+ "loss": 3.523796796798706,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.03939380124211311,
+ "learning_rate": 3.599327707318464e-05,
+ "loss": 2.227039337158203,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 0.6250734925270081,
+ "eval_runtime": 59.919,
+ "eval_samples_per_second": 40.755,
+ "eval_steps_per_second": 0.651,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01396160558464,
+ "grad_norm": 0.045106735080480576,
+ "learning_rate": 3.587664028478779e-05,
+ "loss": 3.484560489654541,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02792321116929,
+ "grad_norm": 0.04480452463030815,
+ "learning_rate": 3.5760218712857046e-05,
+ "loss": 3.4960923194885254,
+ "step": 3098
+ },
+ {
+ "epoch": 43.04188481675393,
+ "grad_norm": 0.04462869092822075,
+ "learning_rate": 3.564401245119119e-05,
+ "loss": 3.509406089782715,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05584642233857,
+ "grad_norm": 0.045312751084566116,
+ "learning_rate": 3.552802159341543e-05,
+ "loss": 3.508052349090576,
+ "step": 3100
+ },
+ {
+ "epoch": 43.06980802792321,
+ "grad_norm": 0.046693965792655945,
+ "learning_rate": 3.541224623298155e-05,
+ "loss": 3.482706069946289,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083769633507856,
+ "grad_norm": 0.04617954418063164,
+ "learning_rate": 3.529668646316765e-05,
+ "loss": 3.505126953125,
+ "step": 3102
+ },
+ {
+ "epoch": 43.0977312390925,
+ "grad_norm": 0.045126017183065414,
+ "learning_rate": 3.5181342377078166e-05,
+ "loss": 3.511366367340088,
+ "step": 3103
+ },
+ {
+ "epoch": 43.11169284467714,
+ "grad_norm": 0.045670486986637115,
+ "learning_rate": 3.5066214067643785e-05,
+ "loss": 3.5024423599243164,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12565445026178,
+ "grad_norm": 0.04518337547779083,
+ "learning_rate": 3.495130162762119e-05,
+ "loss": 3.4837448596954346,
+ "step": 3105
+ },
+ {
+ "epoch": 43.139616055846425,
+ "grad_norm": 0.04559924453496933,
+ "learning_rate": 3.483660514959343e-05,
+ "loss": 3.505150556564331,
+ "step": 3106
+ },
+ {
+ "epoch": 43.153577661431065,
+ "grad_norm": 0.04574387148022652,
+ "learning_rate": 3.4722124725969235e-05,
+ "loss": 3.4935598373413086,
+ "step": 3107
+ },
+ {
+ "epoch": 43.167539267015705,
+ "grad_norm": 0.045334119349718094,
+ "learning_rate": 3.460786044898346e-05,
+ "loss": 3.498568534851074,
+ "step": 3108
+ },
+ {
+ "epoch": 43.181500872600346,
+ "grad_norm": 0.04662643373012543,
+ "learning_rate": 3.4493812410696784e-05,
+ "loss": 3.5291733741760254,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19546247818499,
+ "grad_norm": 0.04649493470788002,
+ "learning_rate": 3.437998070299564e-05,
+ "loss": 3.5123510360717773,
+ "step": 3110
+ },
+ {
+ "epoch": 43.20942408376963,
+ "grad_norm": 0.047649726271629333,
+ "learning_rate": 3.426636541759225e-05,
+ "loss": 3.4946069717407227,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223385689354274,
+ "grad_norm": 0.045752767473459244,
+ "learning_rate": 3.415296664602425e-05,
+ "loss": 3.5374412536621094,
+ "step": 3112
+ },
+ {
+ "epoch": 43.23734729493892,
+ "grad_norm": 0.04584295302629471,
+ "learning_rate": 3.403978447965507e-05,
+ "loss": 3.5236153602600098,
+ "step": 3113
+ },
+ {
+ "epoch": 43.25130890052356,
+ "grad_norm": 0.04679722711443901,
+ "learning_rate": 3.392681900967355e-05,
+ "loss": 3.4972286224365234,
+ "step": 3114
+ },
+ {
+ "epoch": 43.2652705061082,
+ "grad_norm": 0.04629402980208397,
+ "learning_rate": 3.381407032709395e-05,
+ "loss": 3.5219178199768066,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27923211169284,
+ "grad_norm": 0.046149227768182755,
+ "learning_rate": 3.3701538522755714e-05,
+ "loss": 3.4896912574768066,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29319371727749,
+ "grad_norm": 0.04659546539187431,
+ "learning_rate": 3.35892236873238e-05,
+ "loss": 3.4911556243896484,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30715532286213,
+ "grad_norm": 0.0461997464299202,
+ "learning_rate": 3.3477125911288284e-05,
+ "loss": 3.499098777770996,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32111692844677,
+ "grad_norm": 0.0467575341463089,
+ "learning_rate": 3.3365245284964225e-05,
+ "loss": 3.4635555744171143,
+ "step": 3119
+ },
+ {
+ "epoch": 43.33507853403141,
+ "grad_norm": 0.04900452122092247,
+ "learning_rate": 3.3253581898491836e-05,
+ "loss": 3.50935697555542,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34904013961606,
+ "grad_norm": 0.0469595268368721,
+ "learning_rate": 3.3142135841836276e-05,
+ "loss": 3.508852958679199,
+ "step": 3121
+ },
+ {
+ "epoch": 43.3630017452007,
+ "grad_norm": 0.048511046916246414,
+ "learning_rate": 3.303090720478773e-05,
+ "loss": 3.5232086181640625,
+ "step": 3122
+ },
+ {
+ "epoch": 43.37696335078534,
+ "grad_norm": 0.047680724412202835,
+ "learning_rate": 3.291989607696092e-05,
+ "loss": 3.510446786880493,
+ "step": 3123
+ },
+ {
+ "epoch": 43.390924956369986,
+ "grad_norm": 0.045223165303468704,
+ "learning_rate": 3.280910254779565e-05,
+ "loss": 3.477311611175537,
+ "step": 3124
+ },
+ {
+ "epoch": 43.404886561954626,
+ "grad_norm": 0.047112707048654556,
+ "learning_rate": 3.269852670655629e-05,
+ "loss": 3.488718032836914,
+ "step": 3125
+ },
+ {
+ "epoch": 43.41884816753927,
+ "grad_norm": 0.04957739636301994,
+ "learning_rate": 3.258816864233169e-05,
+ "loss": 3.511599540710449,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43280977312391,
+ "grad_norm": 0.04708652198314667,
+ "learning_rate": 3.247802844403536e-05,
+ "loss": 3.5076985359191895,
+ "step": 3127
+ },
+ {
+ "epoch": 43.446771378708554,
+ "grad_norm": 0.04767759144306183,
+ "learning_rate": 3.236810620040532e-05,
+ "loss": 3.5122997760772705,
+ "step": 3128
+ },
+ {
+ "epoch": 43.460732984293195,
+ "grad_norm": 0.04700035974383354,
+ "learning_rate": 3.225840200000396e-05,
+ "loss": 3.479099988937378,
+ "step": 3129
+ },
+ {
+ "epoch": 43.474694589877835,
+ "grad_norm": 0.046844471246004105,
+ "learning_rate": 3.214891593121788e-05,
+ "loss": 3.4828433990478516,
+ "step": 3130
+ },
+ {
+ "epoch": 43.488656195462475,
+ "grad_norm": 0.04689041152596474,
+ "learning_rate": 3.203964808225801e-05,
+ "loss": 3.49288010597229,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50261780104712,
+ "grad_norm": 0.04734606295824051,
+ "learning_rate": 3.193059854115968e-05,
+ "loss": 3.50508975982666,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51657940663176,
+ "grad_norm": 0.04855719581246376,
+ "learning_rate": 3.182176739578195e-05,
+ "loss": 3.5019211769104004,
+ "step": 3133
+ },
+ {
+ "epoch": 43.5305410122164,
+ "grad_norm": 0.047114770859479904,
+ "learning_rate": 3.171315473380815e-05,
+ "loss": 3.522956371307373,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544502617801044,
+ "grad_norm": 0.04759597033262253,
+ "learning_rate": 3.160476064274555e-05,
+ "loss": 3.488766670227051,
+ "step": 3135
+ },
+ {
+ "epoch": 43.55846422338569,
+ "grad_norm": 0.04906241595745087,
+ "learning_rate": 3.149658520992537e-05,
+ "loss": 3.507312536239624,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57242582897033,
+ "grad_norm": 0.047610968351364136,
+ "learning_rate": 3.138862852250249e-05,
+ "loss": 3.5200676918029785,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58638743455497,
+ "grad_norm": 0.048370856791734695,
+ "learning_rate": 3.1280890667455694e-05,
+ "loss": 3.488644599914551,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60034904013962,
+ "grad_norm": 0.04862125962972641,
+ "learning_rate": 3.1173371731587496e-05,
+ "loss": 3.5156755447387695,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61431064572426,
+ "grad_norm": 0.04806689918041229,
+ "learning_rate": 3.1066071801523905e-05,
+ "loss": 3.500375747680664,
+ "step": 3140
+ },
+ {
+ "epoch": 43.6282722513089,
+ "grad_norm": 0.04813141003251076,
+ "learning_rate": 3.095899096371456e-05,
+ "loss": 3.489372730255127,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64223385689354,
+ "grad_norm": 0.047211259603500366,
+ "learning_rate": 3.0852129304432506e-05,
+ "loss": 3.51399564743042,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65619546247819,
+ "grad_norm": 0.04881766065955162,
+ "learning_rate": 3.074548690977434e-05,
+ "loss": 3.5099027156829834,
+ "step": 3143
+ },
+ {
+ "epoch": 43.67015706806283,
+ "grad_norm": 0.04750249162316322,
+ "learning_rate": 3.0639063865659834e-05,
+ "loss": 3.5211377143859863,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68411867364747,
+ "grad_norm": 0.04968058690428734,
+ "learning_rate": 3.0532860257832144e-05,
+ "loss": 3.4967284202575684,
+ "step": 3145
+ },
+ {
+ "epoch": 43.69808027923211,
+ "grad_norm": 0.04763082042336464,
+ "learning_rate": 3.04268761718576e-05,
+ "loss": 3.510603666305542,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712041884816756,
+ "grad_norm": 0.04788519814610481,
+ "learning_rate": 3.0321111693125648e-05,
+ "loss": 3.486903190612793,
+ "step": 3147
+ },
+ {
+ "epoch": 43.726003490401396,
+ "grad_norm": 0.04858475551009178,
+ "learning_rate": 3.0215566906848906e-05,
+ "loss": 3.5200934410095215,
+ "step": 3148
+ },
+ {
+ "epoch": 43.73996509598604,
+ "grad_norm": 0.04838474094867706,
+ "learning_rate": 3.0110241898062726e-05,
+ "loss": 3.4970016479492188,
+ "step": 3149
+ },
+ {
+ "epoch": 43.753926701570684,
+ "grad_norm": 0.0477188378572464,
+ "learning_rate": 3.0005136751625748e-05,
+ "loss": 3.5032782554626465,
+ "step": 3150
+ },
+ {
+ "epoch": 43.767888307155324,
+ "grad_norm": 0.048580750823020935,
+ "learning_rate": 2.9900251552219178e-05,
+ "loss": 3.5062713623046875,
+ "step": 3151
+ },
+ {
+ "epoch": 43.781849912739965,
+ "grad_norm": 0.048856496810913086,
+ "learning_rate": 2.9795586384347202e-05,
+ "loss": 3.5173544883728027,
+ "step": 3152
+ },
+ {
+ "epoch": 43.795811518324605,
+ "grad_norm": 0.04810957610607147,
+ "learning_rate": 2.9691141332336494e-05,
+ "loss": 3.5213584899902344,
+ "step": 3153
+ },
+ {
+ "epoch": 43.80977312390925,
+ "grad_norm": 0.048595767468214035,
+ "learning_rate": 2.9586916480336735e-05,
+ "loss": 3.5045084953308105,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82373472949389,
+ "grad_norm": 0.048057105392217636,
+ "learning_rate": 2.9482911912320006e-05,
+ "loss": 3.4990477561950684,
+ "step": 3155
+ },
+ {
+ "epoch": 43.83769633507853,
+ "grad_norm": 0.047996897250413895,
+ "learning_rate": 2.9379127712080756e-05,
+ "loss": 3.5060601234436035,
+ "step": 3156
+ },
+ {
+ "epoch": 43.85165794066317,
+ "grad_norm": 0.0492832325398922,
+ "learning_rate": 2.9275563963236107e-05,
+ "loss": 3.4959583282470703,
+ "step": 3157
+ },
+ {
+ "epoch": 43.86561954624782,
+ "grad_norm": 0.047811295837163925,
+ "learning_rate": 2.9172220749225548e-05,
+ "loss": 3.5086965560913086,
+ "step": 3158
+ },
+ {
+ "epoch": 43.87958115183246,
+ "grad_norm": 0.04762179031968117,
+ "learning_rate": 2.906909815331085e-05,
+ "loss": 3.5065364837646484,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8935427574171,
+ "grad_norm": 0.04817652702331543,
+ "learning_rate": 2.896619625857588e-05,
+ "loss": 3.4908013343811035,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90750436300174,
+ "grad_norm": 0.04766019433736801,
+ "learning_rate": 2.8863515147927003e-05,
+ "loss": 3.4984240531921387,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92146596858639,
+ "grad_norm": 0.04822462424635887,
+ "learning_rate": 2.8761054904092514e-05,
+ "loss": 3.4717109203338623,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93542757417103,
+ "grad_norm": 0.048152536153793335,
+ "learning_rate": 2.8658815609622665e-05,
+ "loss": 3.4566991329193115,
+ "step": 3163
+ },
+ {
+ "epoch": 43.94938917975567,
+ "grad_norm": 0.04917222261428833,
+ "learning_rate": 2.855679734688993e-05,
+ "loss": 3.5119752883911133,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96335078534032,
+ "grad_norm": 0.04876438155770302,
+ "learning_rate": 2.8455000198088502e-05,
+ "loss": 3.500932216644287,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97731239092496,
+ "grad_norm": 0.048159267753362656,
+ "learning_rate": 2.8353424245234582e-05,
+ "loss": 3.4993038177490234,
+ "step": 3166
+ },
+ {
+ "epoch": 43.9912739965096,
+ "grad_norm": 0.04845968633890152,
+ "learning_rate": 2.825206957016599e-05,
+ "loss": 3.5007855892181396,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.03740447014570236,
+ "learning_rate": 2.8150936254542376e-05,
+ "loss": 2.204704523086548,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 0.6261638402938843,
+ "eval_runtime": 60.2,
+ "eval_samples_per_second": 40.565,
+ "eval_steps_per_second": 0.648,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01396160558464,
+ "grad_norm": 0.042757242918014526,
+ "learning_rate": 2.8050024379845118e-05,
+ "loss": 3.4896841049194336,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02792321116929,
+ "grad_norm": 0.04291922226548195,
+ "learning_rate": 2.7949334027376986e-05,
+ "loss": 3.519742965698242,
+ "step": 3170
+ },
+ {
+ "epoch": 44.04188481675393,
+ "grad_norm": 0.042688220739364624,
+ "learning_rate": 2.7848865278262427e-05,
+ "loss": 3.468482255935669,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05584642233857,
+ "grad_norm": 0.044173792004585266,
+ "learning_rate": 2.7748618213447268e-05,
+ "loss": 3.50119686126709,
+ "step": 3172
+ },
+ {
+ "epoch": 44.06980802792321,
+ "grad_norm": 0.04376482963562012,
+ "learning_rate": 2.764859291369885e-05,
+ "loss": 3.466043472290039,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083769633507856,
+ "grad_norm": 0.044028062373399734,
+ "learning_rate": 2.7548789459605603e-05,
+ "loss": 3.505507469177246,
+ "step": 3174
+ },
+ {
+ "epoch": 44.0977312390925,
+ "grad_norm": 0.04578253999352455,
+ "learning_rate": 2.744920793157743e-05,
+ "loss": 3.498349189758301,
+ "step": 3175
+ },
+ {
+ "epoch": 44.11169284467714,
+ "grad_norm": 0.044063616544008255,
+ "learning_rate": 2.7349848409845462e-05,
+ "loss": 3.4928536415100098,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12565445026178,
+ "grad_norm": 0.04500304162502289,
+ "learning_rate": 2.725071097446174e-05,
+ "loss": 3.5163066387176514,
+ "step": 3177
+ },
+ {
+ "epoch": 44.139616055846425,
+ "grad_norm": 0.0449041984975338,
+ "learning_rate": 2.715179570529959e-05,
+ "loss": 3.512826442718506,
+ "step": 3178
+ },
+ {
+ "epoch": 44.153577661431065,
+ "grad_norm": 0.0450286865234375,
+ "learning_rate": 2.7053102682053134e-05,
+ "loss": 3.508207082748413,
+ "step": 3179
+ },
+ {
+ "epoch": 44.167539267015705,
+ "grad_norm": 0.04426885396242142,
+ "learning_rate": 2.6954631984237766e-05,
+ "loss": 3.4822616577148438,
+ "step": 3180
+ },
+ {
+ "epoch": 44.181500872600346,
+ "grad_norm": 0.04422073811292648,
+ "learning_rate": 2.6856383691189347e-05,
+ "loss": 3.490192413330078,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19546247818499,
+ "grad_norm": 0.04403862729668617,
+ "learning_rate": 2.675835788206485e-05,
+ "loss": 3.477935552597046,
+ "step": 3182
+ },
+ {
+ "epoch": 44.20942408376963,
+ "grad_norm": 0.0447356291115284,
+ "learning_rate": 2.6660554635841912e-05,
+ "loss": 3.4865713119506836,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223385689354274,
+ "grad_norm": 0.04489988833665848,
+ "learning_rate": 2.6562974031318788e-05,
+ "loss": 3.4577198028564453,
+ "step": 3184
+ },
+ {
+ "epoch": 44.23734729493892,
+ "grad_norm": 0.046116411685943604,
+ "learning_rate": 2.6465616147114485e-05,
+ "loss": 3.4821128845214844,
+ "step": 3185
+ },
+ {
+ "epoch": 44.25130890052356,
+ "grad_norm": 0.04555881768465042,
+ "learning_rate": 2.6368481061668344e-05,
+ "loss": 3.5149850845336914,
+ "step": 3186
+ },
+ {
+ "epoch": 44.2652705061082,
+ "grad_norm": 0.04373980686068535,
+ "learning_rate": 2.6271568853240557e-05,
+ "loss": 3.49794602394104,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27923211169284,
+ "grad_norm": 0.04471353814005852,
+ "learning_rate": 2.6174879599911347e-05,
+ "loss": 3.4806714057922363,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29319371727749,
+ "grad_norm": 0.044246856123209,
+ "learning_rate": 2.6078413379581627e-05,
+ "loss": 3.481912851333618,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30715532286213,
+ "grad_norm": 0.04349849745631218,
+ "learning_rate": 2.5982170269972397e-05,
+ "loss": 3.478154182434082,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32111692844677,
+ "grad_norm": 0.044867731630802155,
+ "learning_rate": 2.588615034862507e-05,
+ "loss": 3.497056007385254,
+ "step": 3191
+ },
+ {
+ "epoch": 44.33507853403141,
+ "grad_norm": 0.04483356326818466,
+ "learning_rate": 2.5790353692901158e-05,
+ "loss": 3.4905591011047363,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34904013961606,
+ "grad_norm": 0.044564153999090195,
+ "learning_rate": 2.5694780379982238e-05,
+ "loss": 3.488785743713379,
+ "step": 3193
+ },
+ {
+ "epoch": 44.3630017452007,
+ "grad_norm": 0.044888898730278015,
+ "learning_rate": 2.5599430486870055e-05,
+ "loss": 3.5219860076904297,
+ "step": 3194
+ },
+ {
+ "epoch": 44.37696335078534,
+ "grad_norm": 0.04566695913672447,
+ "learning_rate": 2.5504304090386225e-05,
+ "loss": 3.5113654136657715,
+ "step": 3195
+ },
+ {
+ "epoch": 44.390924956369986,
+ "grad_norm": 0.045218855142593384,
+ "learning_rate": 2.5409401267172474e-05,
+ "loss": 3.4437203407287598,
+ "step": 3196
+ },
+ {
+ "epoch": 44.404886561954626,
+ "grad_norm": 0.04407033324241638,
+ "learning_rate": 2.5314722093690176e-05,
+ "loss": 3.4498770236968994,
+ "step": 3197
+ },
+ {
+ "epoch": 44.41884816753927,
+ "grad_norm": 0.04537077248096466,
+ "learning_rate": 2.522026664622071e-05,
+ "loss": 3.492107391357422,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43280977312391,
+ "grad_norm": 0.04648765176534653,
+ "learning_rate": 2.5126035000865124e-05,
+ "loss": 3.4614241123199463,
+ "step": 3199
+ },
+ {
+ "epoch": 44.446771378708554,
+ "grad_norm": 0.044741127640008926,
+ "learning_rate": 2.503202723354413e-05,
+ "loss": 3.488483428955078,
+ "step": 3200
+ },
+ {
+ "epoch": 44.460732984293195,
+ "grad_norm": 0.04496583715081215,
+ "learning_rate": 2.4938243419998047e-05,
+ "loss": 3.5051095485687256,
+ "step": 3201
+ },
+ {
+ "epoch": 44.474694589877835,
+ "grad_norm": 0.044309262186288834,
+ "learning_rate": 2.484468363578689e-05,
+ "loss": 3.503817081451416,
+ "step": 3202
+ },
+ {
+ "epoch": 44.488656195462475,
+ "grad_norm": 0.04608091711997986,
+ "learning_rate": 2.475134795629004e-05,
+ "loss": 3.4899988174438477,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50261780104712,
+ "grad_norm": 0.044740043580532074,
+ "learning_rate": 2.4658236456706357e-05,
+ "loss": 3.5216774940490723,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51657940663176,
+ "grad_norm": 0.04505334049463272,
+ "learning_rate": 2.456534921205403e-05,
+ "loss": 3.4898223876953125,
+ "step": 3205
+ },
+ {
+ "epoch": 44.5305410122164,
+ "grad_norm": 0.04470878466963768,
+ "learning_rate": 2.4472686297170822e-05,
+ "loss": 3.5085020065307617,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544502617801044,
+ "grad_norm": 0.04671593755483627,
+ "learning_rate": 2.4380247786713403e-05,
+ "loss": 3.4924004077911377,
+ "step": 3207
+ },
+ {
+ "epoch": 44.55846422338569,
+ "grad_norm": 0.04502643644809723,
+ "learning_rate": 2.428803375515788e-05,
+ "loss": 3.4717538356781006,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57242582897033,
+ "grad_norm": 0.0448530837893486,
+ "learning_rate": 2.4196044276799397e-05,
+ "loss": 3.489748954772949,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58638743455497,
+ "grad_norm": 0.04592438042163849,
+ "learning_rate": 2.410427942575228e-05,
+ "loss": 3.49373722076416,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60034904013962,
+ "grad_norm": 0.04584595933556557,
+ "learning_rate": 2.4012739275949746e-05,
+ "loss": 3.5074329376220703,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61431064572426,
+ "grad_norm": 0.04531612619757652,
+ "learning_rate": 2.3921423901144008e-05,
+ "loss": 3.4626989364624023,
+ "step": 3212
+ },
+ {
+ "epoch": 44.6282722513089,
+ "grad_norm": 0.045012522488832474,
+ "learning_rate": 2.3830333374906342e-05,
+ "loss": 3.4816949367523193,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64223385689354,
+ "grad_norm": 0.04681449756026268,
+ "learning_rate": 2.3739467770626637e-05,
+ "loss": 3.5316720008850098,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65619546247819,
+ "grad_norm": 0.04523782432079315,
+ "learning_rate": 2.3648827161513686e-05,
+ "loss": 3.5114080905914307,
+ "step": 3215
+ },
+ {
+ "epoch": 44.67015706806283,
+ "grad_norm": 0.046586692333221436,
+ "learning_rate": 2.3558411620595038e-05,
+ "loss": 3.5085229873657227,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68411867364747,
+ "grad_norm": 0.046556103974580765,
+ "learning_rate": 2.3468221220716844e-05,
+ "loss": 3.496047019958496,
+ "step": 3217
+ },
+ {
+ "epoch": 44.69808027923211,
+ "grad_norm": 0.046004518866539,
+ "learning_rate": 2.33782560345438e-05,
+ "loss": 3.498654365539551,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712041884816756,
+ "grad_norm": 0.046531420201063156,
+ "learning_rate": 2.328851613455932e-05,
+ "loss": 3.4796559810638428,
+ "step": 3219
+ },
+ {
+ "epoch": 44.726003490401396,
+ "grad_norm": 0.04683070257306099,
+ "learning_rate": 2.3199001593065204e-05,
+ "loss": 3.494418144226074,
+ "step": 3220
+ },
+ {
+ "epoch": 44.73996509598604,
+ "grad_norm": 0.04620381072163582,
+ "learning_rate": 2.3109712482181684e-05,
+ "loss": 3.488778591156006,
+ "step": 3221
+ },
+ {
+ "epoch": 44.753926701570684,
+ "grad_norm": 0.04523707926273346,
+ "learning_rate": 2.3020648873847414e-05,
+ "loss": 3.48232364654541,
+ "step": 3222
+ },
+ {
+ "epoch": 44.767888307155324,
+ "grad_norm": 0.04752018675208092,
+ "learning_rate": 2.2931810839819235e-05,
+ "loss": 3.5018577575683594,
+ "step": 3223
+ },
+ {
+ "epoch": 44.781849912739965,
+ "grad_norm": 0.04529130086302757,
+ "learning_rate": 2.284319845167253e-05,
+ "loss": 3.478651523590088,
+ "step": 3224
+ },
+ {
+ "epoch": 44.795811518324605,
+ "grad_norm": 0.04641016200184822,
+ "learning_rate": 2.2754811780800532e-05,
+ "loss": 3.5006914138793945,
+ "step": 3225
+ },
+ {
+ "epoch": 44.80977312390925,
+ "grad_norm": 0.0468735434114933,
+ "learning_rate": 2.2666650898414827e-05,
+ "loss": 3.481553554534912,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82373472949389,
+ "grad_norm": 0.04699835553765297,
+ "learning_rate": 2.2578715875545095e-05,
+ "loss": 3.496619939804077,
+ "step": 3227
+ },
+ {
+ "epoch": 44.83769633507853,
+ "grad_norm": 0.045977186411619186,
+ "learning_rate": 2.2491006783038976e-05,
+ "loss": 3.4861462116241455,
+ "step": 3228
+ },
+ {
+ "epoch": 44.85165794066317,
+ "grad_norm": 0.04640476405620575,
+ "learning_rate": 2.2403523691562092e-05,
+ "loss": 3.518702983856201,
+ "step": 3229
+ },
+ {
+ "epoch": 44.86561954624782,
+ "grad_norm": 0.04592562094330788,
+ "learning_rate": 2.2316266671598003e-05,
+ "loss": 3.47056245803833,
+ "step": 3230
+ },
+ {
+ "epoch": 44.87958115183246,
+ "grad_norm": 0.04678195342421532,
+ "learning_rate": 2.222923579344805e-05,
+ "loss": 3.4597339630126953,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8935427574171,
+ "grad_norm": 0.04635465890169144,
+ "learning_rate": 2.2142431127231502e-05,
+ "loss": 3.4888925552368164,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90750436300174,
+ "grad_norm": 0.04632442072033882,
+ "learning_rate": 2.2055852742885322e-05,
+ "loss": 3.5017807483673096,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92146596858639,
+ "grad_norm": 0.04697349667549133,
+ "learning_rate": 2.196950071016409e-05,
+ "loss": 3.4915904998779297,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93542757417103,
+ "grad_norm": 0.04719756916165352,
+ "learning_rate": 2.1883375098640134e-05,
+ "loss": 3.4640939235687256,
+ "step": 3235
+ },
+ {
+ "epoch": 44.94938917975567,
+ "grad_norm": 0.046704139560461044,
+ "learning_rate": 2.1797475977703352e-05,
+ "loss": 3.4803948402404785,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96335078534032,
+ "grad_norm": 0.04726865515112877,
+ "learning_rate": 2.1711803416560964e-05,
+ "loss": 3.4580843448638916,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97731239092496,
+ "grad_norm": 0.04642590507864952,
+ "learning_rate": 2.16263574842379e-05,
+ "loss": 3.457815170288086,
+ "step": 3238
+ },
+ {
+ "epoch": 44.9912739965096,
+ "grad_norm": 0.047316454350948334,
+ "learning_rate": 2.15411382495764e-05,
+ "loss": 3.5036163330078125,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.035441044718027115,
+ "learning_rate": 2.1456145781236057e-05,
+ "loss": 2.2000532150268555,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 0.6272639632225037,
+ "eval_runtime": 58.8916,
+ "eval_samples_per_second": 41.466,
+ "eval_steps_per_second": 0.662,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01396160558464,
+ "grad_norm": 0.04144950956106186,
+ "learning_rate": 2.1371380147693752e-05,
+ "loss": 3.467620849609375,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02792321116929,
+ "grad_norm": 0.041062530130147934,
+ "learning_rate": 2.1286841417243662e-05,
+ "loss": 3.456000328063965,
+ "step": 3242
+ },
+ {
+ "epoch": 45.04188481675393,
+ "grad_norm": 0.0414804182946682,
+ "learning_rate": 2.1202529657997137e-05,
+ "loss": 3.489531993865967,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05584642233857,
+ "grad_norm": 0.041725676506757736,
+ "learning_rate": 2.1118444937882538e-05,
+ "loss": 3.4876179695129395,
+ "step": 3244
+ },
+ {
+ "epoch": 45.06980802792321,
+ "grad_norm": 0.04175892472267151,
+ "learning_rate": 2.1034587324645456e-05,
+ "loss": 3.4610490798950195,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083769633507856,
+ "grad_norm": 0.04214112088084221,
+ "learning_rate": 2.095095688584846e-05,
+ "loss": 3.456325054168701,
+ "step": 3246
+ },
+ {
+ "epoch": 45.0977312390925,
+ "grad_norm": 0.04136132821440697,
+ "learning_rate": 2.086755368887107e-05,
+ "loss": 3.4657845497131348,
+ "step": 3247
+ },
+ {
+ "epoch": 45.11169284467714,
+ "grad_norm": 0.04285932704806328,
+ "learning_rate": 2.0784377800909802e-05,
+ "loss": 3.464888572692871,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12565445026178,
+ "grad_norm": 0.041763875633478165,
+ "learning_rate": 2.0701429288977847e-05,
+ "loss": 3.4659647941589355,
+ "step": 3249
+ },
+ {
+ "epoch": 45.139616055846425,
+ "grad_norm": 0.04164697974920273,
+ "learning_rate": 2.0618708219905455e-05,
+ "loss": 3.487064838409424,
+ "step": 3250
+ },
+ {
+ "epoch": 45.153577661431065,
+ "grad_norm": 0.04224378988146782,
+ "learning_rate": 2.0536214660339398e-05,
+ "loss": 3.458814859390259,
+ "step": 3251
+ },
+ {
+ "epoch": 45.167539267015705,
+ "grad_norm": 0.043333012610673904,
+ "learning_rate": 2.0453948676743267e-05,
+ "loss": 3.473972797393799,
+ "step": 3252
+ },
+ {
+ "epoch": 45.181500872600346,
+ "grad_norm": 0.04221031069755554,
+ "learning_rate": 2.037191033539731e-05,
+ "loss": 3.4818778038024902,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19546247818499,
+ "grad_norm": 0.0420684814453125,
+ "learning_rate": 2.029009970239831e-05,
+ "loss": 3.4924895763397217,
+ "step": 3254
+ },
+ {
+ "epoch": 45.20942408376963,
+ "grad_norm": 0.04335429146885872,
+ "learning_rate": 2.0208516843659678e-05,
+ "loss": 3.4944746494293213,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223385689354274,
+ "grad_norm": 0.04257744923233986,
+ "learning_rate": 2.012716182491114e-05,
+ "loss": 3.5212855339050293,
+ "step": 3256
+ },
+ {
+ "epoch": 45.23734729493892,
+ "grad_norm": 0.04177210107445717,
+ "learning_rate": 2.0046034711699048e-05,
+ "loss": 3.473543643951416,
+ "step": 3257
+ },
+ {
+ "epoch": 45.25130890052356,
+ "grad_norm": 0.04274517297744751,
+ "learning_rate": 1.9965135569386032e-05,
+ "loss": 3.4608006477355957,
+ "step": 3258
+ },
+ {
+ "epoch": 45.2652705061082,
+ "grad_norm": 0.04179602861404419,
+ "learning_rate": 1.9884464463151077e-05,
+ "loss": 3.490656614303589,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27923211169284,
+ "grad_norm": 0.043795786798000336,
+ "learning_rate": 1.980402145798933e-05,
+ "loss": 3.4702229499816895,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29319371727749,
+ "grad_norm": 0.043217770755290985,
+ "learning_rate": 1.9723806618712484e-05,
+ "loss": 3.502239942550659,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30715532286213,
+ "grad_norm": 0.043023768812417984,
+ "learning_rate": 1.9643820009947983e-05,
+ "loss": 3.4787240028381348,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32111692844677,
+ "grad_norm": 0.0435432493686676,
+ "learning_rate": 1.9564061696139687e-05,
+ "loss": 3.4872756004333496,
+ "step": 3263
+ },
+ {
+ "epoch": 45.33507853403141,
+ "grad_norm": 0.04377902299165726,
+ "learning_rate": 1.948453174154744e-05,
+ "loss": 3.498331069946289,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34904013961606,
+ "grad_norm": 0.04359947890043259,
+ "learning_rate": 1.9405230210247042e-05,
+ "loss": 3.4837467670440674,
+ "step": 3265
+ },
+ {
+ "epoch": 45.3630017452007,
+ "grad_norm": 0.04306787624955177,
+ "learning_rate": 1.9326157166130365e-05,
+ "loss": 3.510972023010254,
+ "step": 3266
+ },
+ {
+ "epoch": 45.37696335078534,
+ "grad_norm": 0.04290027916431427,
+ "learning_rate": 1.9247312672905017e-05,
+ "loss": 3.5001373291015625,
+ "step": 3267
+ },
+ {
+ "epoch": 45.390924956369986,
+ "grad_norm": 0.04314560815691948,
+ "learning_rate": 1.9168696794094758e-05,
+ "loss": 3.4687185287475586,
+ "step": 3268
+ },
+ {
+ "epoch": 45.404886561954626,
+ "grad_norm": 0.0426696240901947,
+ "learning_rate": 1.9090309593038812e-05,
+ "loss": 3.452179431915283,
+ "step": 3269
+ },
+ {
+ "epoch": 45.41884816753927,
+ "grad_norm": 0.041740886867046356,
+ "learning_rate": 1.9012151132892425e-05,
+ "loss": 3.4812817573547363,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43280977312391,
+ "grad_norm": 0.04257669672369957,
+ "learning_rate": 1.893422147662644e-05,
+ "loss": 3.466768264770508,
+ "step": 3271
+ },
+ {
+ "epoch": 45.446771378708554,
+ "grad_norm": 0.04383594170212746,
+ "learning_rate": 1.8856520687027325e-05,
+ "loss": 3.4809494018554688,
+ "step": 3272
+ },
+ {
+ "epoch": 45.460732984293195,
+ "grad_norm": 0.04269306734204292,
+ "learning_rate": 1.8779048826697318e-05,
+ "loss": 3.5032958984375,
+ "step": 3273
+ },
+ {
+ "epoch": 45.474694589877835,
+ "grad_norm": 0.04308982565999031,
+ "learning_rate": 1.8701805958054e-05,
+ "loss": 3.4806809425354004,
+ "step": 3274
+ },
+ {
+ "epoch": 45.488656195462475,
+ "grad_norm": 0.044433314353227615,
+ "learning_rate": 1.8624792143330544e-05,
+ "loss": 3.4718875885009766,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50261780104712,
+ "grad_norm": 0.044048234820365906,
+ "learning_rate": 1.8548007444575656e-05,
+ "loss": 3.476128578186035,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51657940663176,
+ "grad_norm": 0.04227549210190773,
+ "learning_rate": 1.8471451923653322e-05,
+ "loss": 3.5020196437835693,
+ "step": 3277
+ },
+ {
+ "epoch": 45.5305410122164,
+ "grad_norm": 0.043861839920282364,
+ "learning_rate": 1.8395125642242987e-05,
+ "loss": 3.4762532711029053,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544502617801044,
+ "grad_norm": 0.04279392585158348,
+ "learning_rate": 1.8319028661839343e-05,
+ "loss": 3.4641947746276855,
+ "step": 3279
+ },
+ {
+ "epoch": 45.55846422338569,
+ "grad_norm": 0.04272545874118805,
+ "learning_rate": 1.8243161043752384e-05,
+ "loss": 3.4937636852264404,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57242582897033,
+ "grad_norm": 0.04243306443095207,
+ "learning_rate": 1.8167522849107228e-05,
+ "loss": 3.4754786491394043,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58638743455497,
+ "grad_norm": 0.04236120358109474,
+ "learning_rate": 1.809211413884418e-05,
+ "loss": 3.480517625808716,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60034904013962,
+ "grad_norm": 0.04358774796128273,
+ "learning_rate": 1.8016934973718762e-05,
+ "loss": 3.4826560020446777,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61431064572426,
+ "grad_norm": 0.04353169724345207,
+ "learning_rate": 1.7941985414301403e-05,
+ "loss": 3.4996914863586426,
+ "step": 3284
+ },
+ {
+ "epoch": 45.6282722513089,
+ "grad_norm": 0.043377768248319626,
+ "learning_rate": 1.786726552097766e-05,
+ "loss": 3.4741501808166504,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64223385689354,
+ "grad_norm": 0.04338441044092178,
+ "learning_rate": 1.7792775353947908e-05,
+ "loss": 3.4612033367156982,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65619546247819,
+ "grad_norm": 0.04326530918478966,
+ "learning_rate": 1.7718514973227634e-05,
+ "loss": 3.4714088439941406,
+ "step": 3287
+ },
+ {
+ "epoch": 45.67015706806283,
+ "grad_norm": 0.04477553442120552,
+ "learning_rate": 1.7644484438647057e-05,
+ "loss": 3.5052547454833984,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68411867364747,
+ "grad_norm": 0.043149422854185104,
+ "learning_rate": 1.75706838098512e-05,
+ "loss": 3.482375383377075,
+ "step": 3289
+ },
+ {
+ "epoch": 45.69808027923211,
+ "grad_norm": 0.04328175261616707,
+ "learning_rate": 1.749711314629992e-05,
+ "loss": 3.4968152046203613,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712041884816756,
+ "grad_norm": 0.04341139271855354,
+ "learning_rate": 1.742377250726779e-05,
+ "loss": 3.503188133239746,
+ "step": 3291
+ },
+ {
+ "epoch": 45.726003490401396,
+ "grad_norm": 0.0437319278717041,
+ "learning_rate": 1.7350661951844065e-05,
+ "loss": 3.4781315326690674,
+ "step": 3292
+ },
+ {
+ "epoch": 45.73996509598604,
+ "grad_norm": 0.04307844117283821,
+ "learning_rate": 1.727778153893248e-05,
+ "loss": 3.4852523803710938,
+ "step": 3293
+ },
+ {
+ "epoch": 45.753926701570684,
+ "grad_norm": 0.04390977323055267,
+ "learning_rate": 1.720513132725162e-05,
+ "loss": 3.4855167865753174,
+ "step": 3294
+ },
+ {
+ "epoch": 45.767888307155324,
+ "grad_norm": 0.044143348932266235,
+ "learning_rate": 1.713271137533433e-05,
+ "loss": 3.448042392730713,
+ "step": 3295
+ },
+ {
+ "epoch": 45.781849912739965,
+ "grad_norm": 0.04376598075032234,
+ "learning_rate": 1.7060521741528127e-05,
+ "loss": 3.4677300453186035,
+ "step": 3296
+ },
+ {
+ "epoch": 45.795811518324605,
+ "grad_norm": 0.044589050114154816,
+ "learning_rate": 1.698856248399482e-05,
+ "loss": 3.4752211570739746,
+ "step": 3297
+ },
+ {
+ "epoch": 45.80977312390925,
+ "grad_norm": 0.045438025146722794,
+ "learning_rate": 1.6916833660710785e-05,
+ "loss": 3.4890098571777344,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82373472949389,
+ "grad_norm": 0.04418959841132164,
+ "learning_rate": 1.6845335329466606e-05,
+ "loss": 3.4844353199005127,
+ "step": 3299
+ },
+ {
+ "epoch": 45.83769633507853,
+ "grad_norm": 0.043791044503450394,
+ "learning_rate": 1.6774067547867113e-05,
+ "loss": 3.462052822113037,
+ "step": 3300
+ },
+ {
+ "epoch": 45.85165794066317,
+ "grad_norm": 0.04390951991081238,
+ "learning_rate": 1.6703030373331585e-05,
+ "loss": 3.484663486480713,
+ "step": 3301
+ },
+ {
+ "epoch": 45.86561954624782,
+ "grad_norm": 0.04421361908316612,
+ "learning_rate": 1.6632223863093328e-05,
+ "loss": 3.4872078895568848,
+ "step": 3302
+ },
+ {
+ "epoch": 45.87958115183246,
+ "grad_norm": 0.043854694813489914,
+ "learning_rate": 1.656164807419988e-05,
+ "loss": 3.460891008377075,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8935427574171,
+ "grad_norm": 0.04413064941763878,
+ "learning_rate": 1.6491303063512866e-05,
+ "loss": 3.5196800231933594,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90750436300174,
+ "grad_norm": 0.04355199635028839,
+ "learning_rate": 1.6421188887708013e-05,
+ "loss": 3.480886936187744,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92146596858639,
+ "grad_norm": 0.04401418939232826,
+ "learning_rate": 1.6351305603275067e-05,
+ "loss": 3.488598585128784,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93542757417103,
+ "grad_norm": 0.043825335800647736,
+ "learning_rate": 1.6281653266517658e-05,
+ "loss": 3.4577674865722656,
+ "step": 3307
+ },
+ {
+ "epoch": 45.94938917975567,
+ "grad_norm": 0.04364679008722305,
+ "learning_rate": 1.6212231933553485e-05,
+ "loss": 3.466571569442749,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96335078534032,
+ "grad_norm": 0.04436296969652176,
+ "learning_rate": 1.6143041660313997e-05,
+ "loss": 3.5187854766845703,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97731239092496,
+ "grad_norm": 0.043687671422958374,
+ "learning_rate": 1.6074082502544668e-05,
+ "loss": 3.4564049243927,
+ "step": 3310
+ },
+ {
+ "epoch": 45.9912739965096,
+ "grad_norm": 0.0440901480615139,
+ "learning_rate": 1.6005354515804503e-05,
+ "loss": 3.471485137939453,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.03345930203795433,
+ "learning_rate": 1.59368577554664e-05,
+ "loss": 2.173523426055908,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 0.6281663775444031,
+ "eval_runtime": 59.4568,
+ "eval_samples_per_second": 41.072,
+ "eval_steps_per_second": 0.656,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01396160558464,
+ "grad_norm": 0.03809294104576111,
+ "learning_rate": 1.5868592276717137e-05,
+ "loss": 3.446911096572876,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02792321116929,
+ "grad_norm": 0.038114506751298904,
+ "learning_rate": 1.5800558134556806e-05,
+ "loss": 3.480736494064331,
+ "step": 3314
+ },
+ {
+ "epoch": 46.04188481675393,
+ "grad_norm": 0.03834784775972366,
+ "learning_rate": 1.5732755383799315e-05,
+ "loss": 3.460118055343628,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05584642233857,
+ "grad_norm": 0.03990248590707779,
+ "learning_rate": 1.5665184079072162e-05,
+ "loss": 3.4752254486083984,
+ "step": 3316
+ },
+ {
+ "epoch": 46.06980802792321,
+ "grad_norm": 0.039307426661252975,
+ "learning_rate": 1.5597844274816352e-05,
+ "loss": 3.4845035076141357,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083769633507856,
+ "grad_norm": 0.04001433029770851,
+ "learning_rate": 1.553073602528626e-05,
+ "loss": 3.48492693901062,
+ "step": 3318
+ },
+ {
+ "epoch": 46.0977312390925,
+ "grad_norm": 0.039653975516557693,
+ "learning_rate": 1.546385938454982e-05,
+ "loss": 3.474198818206787,
+ "step": 3319
+ },
+ {
+ "epoch": 46.11169284467714,
+ "grad_norm": 0.04127629101276398,
+ "learning_rate": 1.5397214406488374e-05,
+ "loss": 3.463090419769287,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12565445026178,
+ "grad_norm": 0.03933945298194885,
+ "learning_rate": 1.533080114479657e-05,
+ "loss": 3.430025577545166,
+ "step": 3321
+ },
+ {
+ "epoch": 46.139616055846425,
+ "grad_norm": 0.039435796439647675,
+ "learning_rate": 1.5264619652982352e-05,
+ "loss": 3.497743844985962,
+ "step": 3322
+ },
+ {
+ "epoch": 46.153577661431065,
+ "grad_norm": 0.0395926795899868,
+ "learning_rate": 1.5198669984366876e-05,
+ "loss": 3.4905428886413574,
+ "step": 3323
+ },
+ {
+ "epoch": 46.167539267015705,
+ "grad_norm": 0.03973625600337982,
+ "learning_rate": 1.5132952192084793e-05,
+ "loss": 3.4599156379699707,
+ "step": 3324
+ },
+ {
+ "epoch": 46.181500872600346,
+ "grad_norm": 0.04009953886270523,
+ "learning_rate": 1.5067466329083533e-05,
+ "loss": 3.4526820182800293,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19546247818499,
+ "grad_norm": 0.04180147871375084,
+ "learning_rate": 1.5002212448123997e-05,
+ "loss": 3.462456226348877,
+ "step": 3326
+ },
+ {
+ "epoch": 46.20942408376963,
+ "grad_norm": 0.040219224989414215,
+ "learning_rate": 1.4937190601780052e-05,
+ "loss": 3.460576057434082,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223385689354274,
+ "grad_norm": 0.03909081220626831,
+ "learning_rate": 1.4872400842438576e-05,
+ "loss": 3.4756855964660645,
+ "step": 3328
+ },
+ {
+ "epoch": 46.23734729493892,
+ "grad_norm": 0.03988939896225929,
+ "learning_rate": 1.4807843222299557e-05,
+ "loss": 3.489511013031006,
+ "step": 3329
+ },
+ {
+ "epoch": 46.25130890052356,
+ "grad_norm": 0.04111015051603317,
+ "learning_rate": 1.4743517793375814e-05,
+ "loss": 3.4622678756713867,
+ "step": 3330
+ },
+ {
+ "epoch": 46.2652705061082,
+ "grad_norm": 0.039236005395650864,
+ "learning_rate": 1.4679424607493315e-05,
+ "loss": 3.4623537063598633,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27923211169284,
+ "grad_norm": 0.04015125706791878,
+ "learning_rate": 1.4615563716290698e-05,
+ "loss": 3.462616443634033,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29319371727749,
+ "grad_norm": 0.039570603519678116,
+ "learning_rate": 1.455193517121951e-05,
+ "loss": 3.4670767784118652,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30715532286213,
+ "grad_norm": 0.040151674300432205,
+ "learning_rate": 1.448853902354421e-05,
+ "loss": 3.4493308067321777,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32111692844677,
+ "grad_norm": 0.039877235889434814,
+ "learning_rate": 1.4425375324341827e-05,
+ "loss": 3.485646963119507,
+ "step": 3335
+ },
+ {
+ "epoch": 46.33507853403141,
+ "grad_norm": 0.041507795453071594,
+ "learning_rate": 1.4362444124502309e-05,
+ "loss": 3.455681562423706,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34904013961606,
+ "grad_norm": 0.04023195058107376,
+ "learning_rate": 1.429974547472814e-05,
+ "loss": 3.4749460220336914,
+ "step": 3337
+ },
+ {
+ "epoch": 46.3630017452007,
+ "grad_norm": 0.04012970253825188,
+ "learning_rate": 1.4237279425534442e-05,
+ "loss": 3.4900565147399902,
+ "step": 3338
+ },
+ {
+ "epoch": 46.37696335078534,
+ "grad_norm": 0.0395909808576107,
+ "learning_rate": 1.4175046027249065e-05,
+ "loss": 3.4859256744384766,
+ "step": 3339
+ },
+ {
+ "epoch": 46.390924956369986,
+ "grad_norm": 0.040099069476127625,
+ "learning_rate": 1.4113045330012324e-05,
+ "loss": 3.490790843963623,
+ "step": 3340
+ },
+ {
+ "epoch": 46.404886561954626,
+ "grad_norm": 0.041033532470464706,
+ "learning_rate": 1.4051277383776953e-05,
+ "loss": 3.4861276149749756,
+ "step": 3341
+ },
+ {
+ "epoch": 46.41884816753927,
+ "grad_norm": 0.03906581178307533,
+ "learning_rate": 1.398974223830839e-05,
+ "loss": 3.4604973793029785,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43280977312391,
+ "grad_norm": 0.04063129052519798,
+ "learning_rate": 1.3928439943184394e-05,
+ "loss": 3.47312593460083,
+ "step": 3343
+ },
+ {
+ "epoch": 46.446771378708554,
+ "grad_norm": 0.039977047592401505,
+ "learning_rate": 1.3867370547795065e-05,
+ "loss": 3.4339423179626465,
+ "step": 3344
+ },
+ {
+ "epoch": 46.460732984293195,
+ "grad_norm": 0.04055241495370865,
+ "learning_rate": 1.3806534101342891e-05,
+ "loss": 3.4682364463806152,
+ "step": 3345
+ },
+ {
+ "epoch": 46.474694589877835,
+ "grad_norm": 0.04057244211435318,
+ "learning_rate": 1.3745930652842728e-05,
+ "loss": 3.4726850986480713,
+ "step": 3346
+ },
+ {
+ "epoch": 46.488656195462475,
+ "grad_norm": 0.03994397819042206,
+ "learning_rate": 1.368556025112176e-05,
+ "loss": 3.4811201095581055,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50261780104712,
+ "grad_norm": 0.04000184312462807,
+ "learning_rate": 1.362542294481918e-05,
+ "loss": 3.4635982513427734,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51657940663176,
+ "grad_norm": 0.04023074731230736,
+ "learning_rate": 1.356551878238656e-05,
+ "loss": 3.4687578678131104,
+ "step": 3349
+ },
+ {
+ "epoch": 46.5305410122164,
+ "grad_norm": 0.041356779634952545,
+ "learning_rate": 1.3505847812087724e-05,
+ "loss": 3.4677982330322266,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544502617801044,
+ "grad_norm": 0.04142126813530922,
+ "learning_rate": 1.3446410081998384e-05,
+ "loss": 3.4522178173065186,
+ "step": 3351
+ },
+ {
+ "epoch": 46.55846422338569,
+ "grad_norm": 0.04021862894296646,
+ "learning_rate": 1.3387205640006494e-05,
+ "loss": 3.486114025115967,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57242582897033,
+ "grad_norm": 0.04077965393662453,
+ "learning_rate": 1.332823453381196e-05,
+ "loss": 3.4675588607788086,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58638743455497,
+ "grad_norm": 0.040349189192056656,
+ "learning_rate": 1.3269496810926807e-05,
+ "loss": 3.472097635269165,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60034904013962,
+ "grad_norm": 0.04097423329949379,
+ "learning_rate": 1.3210992518674848e-05,
+ "loss": 3.4809305667877197,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61431064572426,
+ "grad_norm": 0.0396139919757843,
+ "learning_rate": 1.315272170419201e-05,
+ "loss": 3.4606056213378906,
+ "step": 3356
+ },
+ {
+ "epoch": 46.6282722513089,
+ "grad_norm": 0.04106489196419716,
+ "learning_rate": 1.3094684414426073e-05,
+ "loss": 3.4697275161743164,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64223385689354,
+ "grad_norm": 0.041042376309633255,
+ "learning_rate": 1.3036880696136532e-05,
+ "loss": 3.4602208137512207,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65619546247819,
+ "grad_norm": 0.04126123711466789,
+ "learning_rate": 1.2979310595894882e-05,
+ "loss": 3.4826760292053223,
+ "step": 3359
+ },
+ {
+ "epoch": 46.67015706806283,
+ "grad_norm": 0.04302288964390755,
+ "learning_rate": 1.2921974160084251e-05,
+ "loss": 3.4608049392700195,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68411867364747,
+ "grad_norm": 0.04080858826637268,
+ "learning_rate": 1.2864871434899672e-05,
+ "loss": 3.46588134765625,
+ "step": 3361
+ },
+ {
+ "epoch": 46.69808027923211,
+ "grad_norm": 0.041789159178733826,
+ "learning_rate": 1.2808002466347654e-05,
+ "loss": 3.475581645965576,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712041884816756,
+ "grad_norm": 0.04102788120508194,
+ "learning_rate": 1.2751367300246535e-05,
+ "loss": 3.4973440170288086,
+ "step": 3363
+ },
+ {
+ "epoch": 46.726003490401396,
+ "grad_norm": 0.041402388364076614,
+ "learning_rate": 1.2694965982226301e-05,
+ "loss": 3.4613161087036133,
+ "step": 3364
+ },
+ {
+ "epoch": 46.73996509598604,
+ "grad_norm": 0.04148413985967636,
+ "learning_rate": 1.2638798557728402e-05,
+ "loss": 3.496325969696045,
+ "step": 3365
+ },
+ {
+ "epoch": 46.753926701570684,
+ "grad_norm": 0.04156513512134552,
+ "learning_rate": 1.2582865072006e-05,
+ "loss": 3.478461742401123,
+ "step": 3366
+ },
+ {
+ "epoch": 46.767888307155324,
+ "grad_norm": 0.040885698050260544,
+ "learning_rate": 1.2527165570123586e-05,
+ "loss": 3.4652328491210938,
+ "step": 3367
+ },
+ {
+ "epoch": 46.781849912739965,
+ "grad_norm": 0.040900446474552155,
+ "learning_rate": 1.2471700096957376e-05,
+ "loss": 3.4657950401306152,
+ "step": 3368
+ },
+ {
+ "epoch": 46.795811518324605,
+ "grad_norm": 0.04140219837427139,
+ "learning_rate": 1.2416468697194767e-05,
+ "loss": 3.4554483890533447,
+ "step": 3369
+ },
+ {
+ "epoch": 46.80977312390925,
+ "grad_norm": 0.041410576552152634,
+ "learning_rate": 1.2361471415334795e-05,
+ "loss": 3.47219181060791,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82373472949389,
+ "grad_norm": 0.0410308837890625,
+ "learning_rate": 1.2306708295687706e-05,
+ "loss": 3.472898244857788,
+ "step": 3371
+ },
+ {
+ "epoch": 46.83769633507853,
+ "grad_norm": 0.041487667709589005,
+ "learning_rate": 1.2252179382375211e-05,
+ "loss": 3.4671497344970703,
+ "step": 3372
+ },
+ {
+ "epoch": 46.85165794066317,
+ "grad_norm": 0.04129431024193764,
+ "learning_rate": 1.2197884719330281e-05,
+ "loss": 3.452998161315918,
+ "step": 3373
+ },
+ {
+ "epoch": 46.86561954624782,
+ "grad_norm": 0.04150335118174553,
+ "learning_rate": 1.2143824350297104e-05,
+ "loss": 3.463134765625,
+ "step": 3374
+ },
+ {
+ "epoch": 46.87958115183246,
+ "grad_norm": 0.04116620495915413,
+ "learning_rate": 1.2089998318831137e-05,
+ "loss": 3.4834423065185547,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8935427574171,
+ "grad_norm": 0.040423303842544556,
+ "learning_rate": 1.2036406668299062e-05,
+ "loss": 3.471663236618042,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90750436300174,
+ "grad_norm": 0.04132043942809105,
+ "learning_rate": 1.198304944187868e-05,
+ "loss": 3.469839572906494,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92146596858639,
+ "grad_norm": 0.040267933160066605,
+ "learning_rate": 1.1929926682558995e-05,
+ "loss": 3.4881057739257812,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93542757417103,
+ "grad_norm": 0.040919527411460876,
+ "learning_rate": 1.1877038433139998e-05,
+ "loss": 3.4716835021972656,
+ "step": 3379
+ },
+ {
+ "epoch": 46.94938917975567,
+ "grad_norm": 0.040698472410440445,
+ "learning_rate": 1.1824384736232879e-05,
+ "loss": 3.491934061050415,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96335078534032,
+ "grad_norm": 0.04163767397403717,
+ "learning_rate": 1.1771965634259653e-05,
+ "loss": 3.519442081451416,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97731239092496,
+ "grad_norm": 0.04191676527261734,
+ "learning_rate": 1.1719781169453536e-05,
+ "loss": 3.4748454093933105,
+ "step": 3382
+ },
+ {
+ "epoch": 46.9912739965096,
+ "grad_norm": 0.04132457450032234,
+ "learning_rate": 1.1667831383858562e-05,
+ "loss": 3.490064859390259,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.03242214769124985,
+ "learning_rate": 1.1616116319329808e-05,
+ "loss": 2.1782138347625732,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 0.6289563179016113,
+ "eval_runtime": 59.4678,
+ "eval_samples_per_second": 41.064,
+ "eval_steps_per_second": 0.656,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01396160558464,
+ "grad_norm": 0.03715526685118675,
+ "learning_rate": 1.1564636017533036e-05,
+ "loss": 3.465269088745117,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02792321116929,
+ "grad_norm": 0.03720983862876892,
+ "learning_rate": 1.151339051994508e-05,
+ "loss": 3.4568798542022705,
+ "step": 3386
+ },
+ {
+ "epoch": 47.04188481675393,
+ "grad_norm": 0.03712873533368111,
+ "learning_rate": 1.146237986785356e-05,
+ "loss": 3.455738067626953,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05584642233857,
+ "grad_norm": 0.03695226460695267,
+ "learning_rate": 1.141160410235674e-05,
+ "loss": 3.4416942596435547,
+ "step": 3388
+ },
+ {
+ "epoch": 47.06980802792321,
+ "grad_norm": 0.03724801167845726,
+ "learning_rate": 1.1361063264363776e-05,
+ "loss": 3.47540545463562,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083769633507856,
+ "grad_norm": 0.03712484985589981,
+ "learning_rate": 1.1310757394594514e-05,
+ "loss": 3.476196765899658,
+ "step": 3390
+ },
+ {
+ "epoch": 47.0977312390925,
+ "grad_norm": 0.037230268120765686,
+ "learning_rate": 1.1260686533579481e-05,
+ "loss": 3.4795098304748535,
+ "step": 3391
+ },
+ {
+ "epoch": 47.11169284467714,
+ "grad_norm": 0.037289027124643326,
+ "learning_rate": 1.121085072165986e-05,
+ "loss": 3.446183681488037,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12565445026178,
+ "grad_norm": 0.03767047077417374,
+ "learning_rate": 1.1161249998987429e-05,
+ "loss": 3.459652900695801,
+ "step": 3393
+ },
+ {
+ "epoch": 47.139616055846425,
+ "grad_norm": 0.037236738950014114,
+ "learning_rate": 1.111188440552471e-05,
+ "loss": 3.479717254638672,
+ "step": 3394
+ },
+ {
+ "epoch": 47.153577661431065,
+ "grad_norm": 0.03738565370440483,
+ "learning_rate": 1.106275398104455e-05,
+ "loss": 3.472529172897339,
+ "step": 3395
+ },
+ {
+ "epoch": 47.167539267015705,
+ "grad_norm": 0.038441140204668045,
+ "learning_rate": 1.1013858765130468e-05,
+ "loss": 3.4527950286865234,
+ "step": 3396
+ },
+ {
+ "epoch": 47.181500872600346,
+ "grad_norm": 0.03788449615240097,
+ "learning_rate": 1.0965198797176513e-05,
+ "loss": 3.474555492401123,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19546247818499,
+ "grad_norm": 0.03705836832523346,
+ "learning_rate": 1.0916774116387103e-05,
+ "loss": 3.459512710571289,
+ "step": 3398
+ },
+ {
+ "epoch": 47.20942408376963,
+ "grad_norm": 0.03746683895587921,
+ "learning_rate": 1.0868584761777125e-05,
+ "loss": 3.467989921569824,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223385689354274,
+ "grad_norm": 0.03808161988854408,
+ "learning_rate": 1.0820630772171869e-05,
+ "loss": 3.472649574279785,
+ "step": 3400
+ },
+ {
+ "epoch": 47.23734729493892,
+ "grad_norm": 0.03828210383653641,
+ "learning_rate": 1.0772912186206987e-05,
+ "loss": 3.4917690753936768,
+ "step": 3401
+ },
+ {
+ "epoch": 47.25130890052356,
+ "grad_norm": 0.039966095238924026,
+ "learning_rate": 1.0725429042328542e-05,
+ "loss": 3.474637985229492,
+ "step": 3402
+ },
+ {
+ "epoch": 47.2652705061082,
+ "grad_norm": 0.03823436051607132,
+ "learning_rate": 1.0678181378792794e-05,
+ "loss": 3.4760427474975586,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27923211169284,
+ "grad_norm": 0.03723641857504845,
+ "learning_rate": 1.0631169233666306e-05,
+ "loss": 3.4606034755706787,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29319371727749,
+ "grad_norm": 0.038485877215862274,
+ "learning_rate": 1.0584392644826016e-05,
+ "loss": 3.468674898147583,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30715532286213,
+ "grad_norm": 0.03701888024806976,
+ "learning_rate": 1.053785164995886e-05,
+ "loss": 3.4768829345703125,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32111692844677,
+ "grad_norm": 0.03713027015328407,
+ "learning_rate": 1.0491546286562148e-05,
+ "loss": 3.469245672225952,
+ "step": 3407
+ },
+ {
+ "epoch": 47.33507853403141,
+ "grad_norm": 0.03890090435743332,
+ "learning_rate": 1.0445476591943259e-05,
+ "loss": 3.454719305038452,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34904013961606,
+ "grad_norm": 0.03849492967128754,
+ "learning_rate": 1.0399642603219712e-05,
+ "loss": 3.4725534915924072,
+ "step": 3409
+ },
+ {
+ "epoch": 47.3630017452007,
+ "grad_norm": 0.03847581893205643,
+ "learning_rate": 1.0354044357319129e-05,
+ "loss": 3.451141357421875,
+ "step": 3410
+ },
+ {
+ "epoch": 47.37696335078534,
+ "grad_norm": 0.03812417387962341,
+ "learning_rate": 1.0308681890979173e-05,
+ "loss": 3.4664249420166016,
+ "step": 3411
+ },
+ {
+ "epoch": 47.390924956369986,
+ "grad_norm": 0.0375078059732914,
+ "learning_rate": 1.026355524074758e-05,
+ "loss": 3.469621181488037,
+ "step": 3412
+ },
+ {
+ "epoch": 47.404886561954626,
+ "grad_norm": 0.03773224353790283,
+ "learning_rate": 1.0218664442982084e-05,
+ "loss": 3.459860324859619,
+ "step": 3413
+ },
+ {
+ "epoch": 47.41884816753927,
+ "grad_norm": 0.038476865738630295,
+ "learning_rate": 1.0174009533850368e-05,
+ "loss": 3.4900689125061035,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43280977312391,
+ "grad_norm": 0.03848643600940704,
+ "learning_rate": 1.0129590549330119e-05,
+ "loss": 3.4760494232177734,
+ "step": 3415
+ },
+ {
+ "epoch": 47.446771378708554,
+ "grad_norm": 0.03840518742799759,
+ "learning_rate": 1.0085407525208865e-05,
+ "loss": 3.4690675735473633,
+ "step": 3416
+ },
+ {
+ "epoch": 47.460732984293195,
+ "grad_norm": 0.03839331865310669,
+ "learning_rate": 1.0041460497084135e-05,
+ "loss": 3.465109348297119,
+ "step": 3417
+ },
+ {
+ "epoch": 47.474694589877835,
+ "grad_norm": 0.03813324123620987,
+ "learning_rate": 9.997749500363172e-06,
+ "loss": 3.44476318359375,
+ "step": 3418
+ },
+ {
+ "epoch": 47.488656195462475,
+ "grad_norm": 0.03780321031808853,
+ "learning_rate": 9.954274570263191e-06,
+ "loss": 3.4609241485595703,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50261780104712,
+ "grad_norm": 0.03767361119389534,
+ "learning_rate": 9.911035741811153e-06,
+ "loss": 3.466444492340088,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51657940663176,
+ "grad_norm": 0.0383882075548172,
+ "learning_rate": 9.868033049843787e-06,
+ "loss": 3.4332656860351562,
+ "step": 3421
+ },
+ {
+ "epoch": 47.5305410122164,
+ "grad_norm": 0.03806808590888977,
+ "learning_rate": 9.825266529007601e-06,
+ "loss": 3.4749398231506348,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544502617801044,
+ "grad_norm": 0.03775869309902191,
+ "learning_rate": 9.782736213758782e-06,
+ "loss": 3.501807928085327,
+ "step": 3423
+ },
+ {
+ "epoch": 47.55846422338569,
+ "grad_norm": 0.03878234326839447,
+ "learning_rate": 9.740442138363317e-06,
+ "loss": 3.469111442565918,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57242582897033,
+ "grad_norm": 0.0384051613509655,
+ "learning_rate": 9.698384336896679e-06,
+ "loss": 3.4669487476348877,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58638743455497,
+ "grad_norm": 0.038485217839479446,
+ "learning_rate": 9.656562843244114e-06,
+ "loss": 3.4800267219543457,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60034904013962,
+ "grad_norm": 0.03876911848783493,
+ "learning_rate": 9.614977691100445e-06,
+ "loss": 3.487262725830078,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61431064572426,
+ "grad_norm": 0.03795599192380905,
+ "learning_rate": 9.57362891397007e-06,
+ "loss": 3.470116376876831,
+ "step": 3428
+ },
+ {
+ "epoch": 47.6282722513089,
+ "grad_norm": 0.03837175294756889,
+ "learning_rate": 9.532516545167004e-06,
+ "loss": 3.465224266052246,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64223385689354,
+ "grad_norm": 0.03832276538014412,
+ "learning_rate": 9.491640617814634e-06,
+ "loss": 3.485874652862549,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65619546247819,
+ "grad_norm": 0.039029840379953384,
+ "learning_rate": 9.451001164846092e-06,
+ "loss": 3.473172426223755,
+ "step": 3431
+ },
+ {
+ "epoch": 47.67015706806283,
+ "grad_norm": 0.03817339241504669,
+ "learning_rate": 9.410598219003725e-06,
+ "loss": 3.458969831466675,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68411867364747,
+ "grad_norm": 0.03954322636127472,
+ "learning_rate": 9.370431812839556e-06,
+ "loss": 3.456747055053711,
+ "step": 3433
+ },
+ {
+ "epoch": 47.69808027923211,
+ "grad_norm": 0.03809211775660515,
+ "learning_rate": 9.330501978714857e-06,
+ "loss": 3.469237804412842,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712041884816756,
+ "grad_norm": 0.038770176470279694,
+ "learning_rate": 9.290808748800408e-06,
+ "loss": 3.4691624641418457,
+ "step": 3435
+ },
+ {
+ "epoch": 47.726003490401396,
+ "grad_norm": 0.03788571432232857,
+ "learning_rate": 9.25135215507634e-06,
+ "loss": 3.463486671447754,
+ "step": 3436
+ },
+ {
+ "epoch": 47.73996509598604,
+ "grad_norm": 0.03861233592033386,
+ "learning_rate": 9.212132229332094e-06,
+ "loss": 3.443244457244873,
+ "step": 3437
+ },
+ {
+ "epoch": 47.753926701570684,
+ "grad_norm": 0.03825637698173523,
+ "learning_rate": 9.173149003166458e-06,
+ "loss": 3.4462857246398926,
+ "step": 3438
+ },
+ {
+ "epoch": 47.767888307155324,
+ "grad_norm": 0.03837190195918083,
+ "learning_rate": 9.134402507987533e-06,
+ "loss": 3.4718427658081055,
+ "step": 3439
+ },
+ {
+ "epoch": 47.781849912739965,
+ "grad_norm": 0.039119500666856766,
+ "learning_rate": 9.095892775012672e-06,
+ "loss": 3.462841510772705,
+ "step": 3440
+ },
+ {
+ "epoch": 47.795811518324605,
+ "grad_norm": 0.038461871445178986,
+ "learning_rate": 9.057619835268437e-06,
+ "loss": 3.4429988861083984,
+ "step": 3441
+ },
+ {
+ "epoch": 47.80977312390925,
+ "grad_norm": 0.039531826972961426,
+ "learning_rate": 9.019583719590641e-06,
+ "loss": 3.4569945335388184,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82373472949389,
+ "grad_norm": 0.03858548030257225,
+ "learning_rate": 8.981784458624375e-06,
+ "loss": 3.4473373889923096,
+ "step": 3443
+ },
+ {
+ "epoch": 47.83769633507853,
+ "grad_norm": 0.03812285140156746,
+ "learning_rate": 8.944222082823717e-06,
+ "loss": 3.482055187225342,
+ "step": 3444
+ },
+ {
+ "epoch": 47.85165794066317,
+ "grad_norm": 0.03824259713292122,
+ "learning_rate": 8.906896622452055e-06,
+ "loss": 3.4854540824890137,
+ "step": 3445
+ },
+ {
+ "epoch": 47.86561954624782,
+ "grad_norm": 0.03781299293041229,
+ "learning_rate": 8.86980810758183e-06,
+ "loss": 3.443784713745117,
+ "step": 3446
+ },
+ {
+ "epoch": 47.87958115183246,
+ "grad_norm": 0.03863195702433586,
+ "learning_rate": 8.832956568094596e-06,
+ "loss": 3.4715123176574707,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8935427574171,
+ "grad_norm": 0.037651095539331436,
+ "learning_rate": 8.796342033680928e-06,
+ "loss": 3.434251546859741,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90750436300174,
+ "grad_norm": 0.03879479691386223,
+ "learning_rate": 8.759964533840574e-06,
+ "loss": 3.4661810398101807,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92146596858639,
+ "grad_norm": 0.03911615535616875,
+ "learning_rate": 8.723824097882145e-06,
+ "loss": 3.459430694580078,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93542757417103,
+ "grad_norm": 0.039879582822322845,
+ "learning_rate": 8.687920754923393e-06,
+ "loss": 3.4611387252807617,
+ "step": 3451
+ },
+ {
+ "epoch": 47.94938917975567,
+ "grad_norm": 0.03886502608656883,
+ "learning_rate": 8.652254533890956e-06,
+ "loss": 3.473893642425537,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96335078534032,
+ "grad_norm": 0.03822801634669304,
+ "learning_rate": 8.616825463520493e-06,
+ "loss": 3.4915146827697754,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97731239092496,
+ "grad_norm": 0.03881949558854103,
+ "learning_rate": 8.581633572356576e-06,
+ "loss": 3.43632173538208,
+ "step": 3454
+ },
+ {
+ "epoch": 47.9912739965096,
+ "grad_norm": 0.03907693922519684,
+ "learning_rate": 8.546678888752629e-06,
+ "loss": 3.464989185333252,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.02964114584028721,
+ "learning_rate": 8.51196144087103e-06,
+ "loss": 2.1612966060638428,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 0.6296182870864868,
+ "eval_runtime": 59.3275,
+ "eval_samples_per_second": 41.161,
+ "eval_steps_per_second": 0.657,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01396160558464,
+ "grad_norm": 0.034902241080999374,
+ "learning_rate": 8.477481256683005e-06,
+ "loss": 3.4160664081573486,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02792321116929,
+ "grad_norm": 0.03473009541630745,
+ "learning_rate": 8.44323836396864e-06,
+ "loss": 3.464644432067871,
+ "step": 3458
+ },
+ {
+ "epoch": 48.04188481675393,
+ "grad_norm": 0.035172175616025925,
+ "learning_rate": 8.409232790316763e-06,
+ "loss": 3.471442699432373,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05584642233857,
+ "grad_norm": 0.03598090633749962,
+ "learning_rate": 8.375464563125094e-06,
+ "loss": 3.475680351257324,
+ "step": 3460
+ },
+ {
+ "epoch": 48.06980802792321,
+ "grad_norm": 0.03536432608962059,
+ "learning_rate": 8.341933709600102e-06,
+ "loss": 3.4751956462860107,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083769633507856,
+ "grad_norm": 0.03533550724387169,
+ "learning_rate": 8.308640256756916e-06,
+ "loss": 3.4708917140960693,
+ "step": 3462
+ },
+ {
+ "epoch": 48.0977312390925,
+ "grad_norm": 0.03569445013999939,
+ "learning_rate": 8.275584231419512e-06,
+ "loss": 3.449456214904785,
+ "step": 3463
+ },
+ {
+ "epoch": 48.11169284467714,
+ "grad_norm": 0.03571323677897453,
+ "learning_rate": 8.24276566022055e-06,
+ "loss": 3.445202350616455,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12565445026178,
+ "grad_norm": 0.0362127348780632,
+ "learning_rate": 8.21018456960132e-06,
+ "loss": 3.446835517883301,
+ "step": 3465
+ },
+ {
+ "epoch": 48.139616055846425,
+ "grad_norm": 0.03627890348434448,
+ "learning_rate": 8.177840985811858e-06,
+ "loss": 3.4842529296875,
+ "step": 3466
+ },
+ {
+ "epoch": 48.153577661431065,
+ "grad_norm": 0.03570439666509628,
+ "learning_rate": 8.145734934910757e-06,
+ "loss": 3.460944652557373,
+ "step": 3467
+ },
+ {
+ "epoch": 48.167539267015705,
+ "grad_norm": 0.03506971150636673,
+ "learning_rate": 8.113866442765364e-06,
+ "loss": 3.4649758338928223,
+ "step": 3468
+ },
+ {
+ "epoch": 48.181500872600346,
+ "grad_norm": 0.0358048677444458,
+ "learning_rate": 8.082235535051449e-06,
+ "loss": 3.4629063606262207,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19546247818499,
+ "grad_norm": 0.0363197922706604,
+ "learning_rate": 8.050842237253533e-06,
+ "loss": 3.4885497093200684,
+ "step": 3470
+ },
+ {
+ "epoch": 48.20942408376963,
+ "grad_norm": 0.03464783728122711,
+ "learning_rate": 8.019686574664564e-06,
+ "loss": 3.470947504043579,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223385689354274,
+ "grad_norm": 0.03518608957529068,
+ "learning_rate": 7.988768572386171e-06,
+ "loss": 3.476457118988037,
+ "step": 3472
+ },
+ {
+ "epoch": 48.23734729493892,
+ "grad_norm": 0.03605307266116142,
+ "learning_rate": 7.958088255328415e-06,
+ "loss": 3.4698140621185303,
+ "step": 3473
+ },
+ {
+ "epoch": 48.25130890052356,
+ "grad_norm": 0.03509247303009033,
+ "learning_rate": 7.927645648209837e-06,
+ "loss": 3.4488396644592285,
+ "step": 3474
+ },
+ {
+ "epoch": 48.2652705061082,
+ "grad_norm": 0.03533833846449852,
+ "learning_rate": 7.897440775557506e-06,
+ "loss": 3.460695266723633,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27923211169284,
+ "grad_norm": 0.03508375212550163,
+ "learning_rate": 7.867473661706946e-06,
+ "loss": 3.4774932861328125,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29319371727749,
+ "grad_norm": 0.03569795936346054,
+ "learning_rate": 7.837744330802136e-06,
+ "loss": 3.461393356323242,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30715532286213,
+ "grad_norm": 0.035462040454149246,
+ "learning_rate": 7.808252806795446e-06,
+ "loss": 3.4403514862060547,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32111692844677,
+ "grad_norm": 0.03597921505570412,
+ "learning_rate": 7.778999113447673e-06,
+ "loss": 3.4533801078796387,
+ "step": 3479
+ },
+ {
+ "epoch": 48.33507853403141,
+ "grad_norm": 0.035972610116004944,
+ "learning_rate": 7.749983274327994e-06,
+ "loss": 3.452761173248291,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34904013961606,
+ "grad_norm": 0.03618500754237175,
+ "learning_rate": 7.721205312813921e-06,
+ "loss": 3.4626686573028564,
+ "step": 3481
+ },
+ {
+ "epoch": 48.3630017452007,
+ "grad_norm": 0.03674949333071709,
+ "learning_rate": 7.69266525209132e-06,
+ "loss": 3.442460060119629,
+ "step": 3482
+ },
+ {
+ "epoch": 48.37696335078534,
+ "grad_norm": 0.0355762355029583,
+ "learning_rate": 7.664363115154476e-06,
+ "loss": 3.4195423126220703,
+ "step": 3483
+ },
+ {
+ "epoch": 48.390924956369986,
+ "grad_norm": 0.03706294670701027,
+ "learning_rate": 7.636298924805872e-06,
+ "loss": 3.4368388652801514,
+ "step": 3484
+ },
+ {
+ "epoch": 48.404886561954626,
+ "grad_norm": 0.03530324995517731,
+ "learning_rate": 7.608472703656282e-06,
+ "loss": 3.4791674613952637,
+ "step": 3485
+ },
+ {
+ "epoch": 48.41884816753927,
+ "grad_norm": 0.03684843331575394,
+ "learning_rate": 7.580884474124833e-06,
+ "loss": 3.470335006713867,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43280977312391,
+ "grad_norm": 0.03622817248106003,
+ "learning_rate": 7.55353425843888e-06,
+ "loss": 3.4495296478271484,
+ "step": 3487
+ },
+ {
+ "epoch": 48.446771378708554,
+ "grad_norm": 0.035566296428442,
+ "learning_rate": 7.526422078633939e-06,
+ "loss": 3.456986904144287,
+ "step": 3488
+ },
+ {
+ "epoch": 48.460732984293195,
+ "grad_norm": 0.03579388186335564,
+ "learning_rate": 7.4995479565538476e-06,
+ "loss": 3.463121175765991,
+ "step": 3489
+ },
+ {
+ "epoch": 48.474694589877835,
+ "grad_norm": 0.03644305467605591,
+ "learning_rate": 7.4729119138506044e-06,
+ "loss": 3.4775054454803467,
+ "step": 3490
+ },
+ {
+ "epoch": 48.488656195462475,
+ "grad_norm": 0.03613374009728432,
+ "learning_rate": 7.446513971984366e-06,
+ "loss": 3.424849271774292,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50261780104712,
+ "grad_norm": 0.035841211676597595,
+ "learning_rate": 7.420354152223485e-06,
+ "loss": 3.4898815155029297,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51657940663176,
+ "grad_norm": 0.03662864491343498,
+ "learning_rate": 7.394432475644402e-06,
+ "loss": 3.4759573936462402,
+ "step": 3493
+ },
+ {
+ "epoch": 48.5305410122164,
+ "grad_norm": 0.0366714708507061,
+ "learning_rate": 7.368748963131854e-06,
+ "loss": 3.4445905685424805,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544502617801044,
+ "grad_norm": 0.036630287766456604,
+ "learning_rate": 7.343303635378537e-06,
+ "loss": 3.4678235054016113,
+ "step": 3495
+ },
+ {
+ "epoch": 48.55846422338569,
+ "grad_norm": 0.036535147577524185,
+ "learning_rate": 7.31809651288524e-06,
+ "loss": 3.442391872406006,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57242582897033,
+ "grad_norm": 0.036027126014232635,
+ "learning_rate": 7.2931276159609475e-06,
+ "loss": 3.465264320373535,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58638743455497,
+ "grad_norm": 0.03622591868042946,
+ "learning_rate": 7.268396964722602e-06,
+ "loss": 3.450774669647217,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60034904013962,
+ "grad_norm": 0.03653610870242119,
+ "learning_rate": 7.243904579095308e-06,
+ "loss": 3.43194842338562,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61431064572426,
+ "grad_norm": 0.0361340306699276,
+ "learning_rate": 7.219650478812065e-06,
+ "loss": 3.453685998916626,
+ "step": 3500
+ },
+ {
+ "epoch": 48.6282722513089,
+ "grad_norm": 0.03608343005180359,
+ "learning_rate": 7.195634683413996e-06,
+ "loss": 3.4311797618865967,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64223385689354,
+ "grad_norm": 0.036149926483631134,
+ "learning_rate": 7.171857212250225e-06,
+ "loss": 3.4822096824645996,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65619546247819,
+ "grad_norm": 0.037198860198259354,
+ "learning_rate": 7.148318084477766e-06,
+ "loss": 3.4502971172332764,
+ "step": 3503
+ },
+ {
+ "epoch": 48.67015706806283,
+ "grad_norm": 0.03657691553235054,
+ "learning_rate": 7.125017319061728e-06,
+ "loss": 3.469975233078003,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68411867364747,
+ "grad_norm": 0.03717451170086861,
+ "learning_rate": 7.101954934775084e-06,
+ "loss": 3.4513626098632812,
+ "step": 3505
+ },
+ {
+ "epoch": 48.69808027923211,
+ "grad_norm": 0.03665406256914139,
+ "learning_rate": 7.079130950198802e-06,
+ "loss": 3.466418981552124,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712041884816756,
+ "grad_norm": 0.03546656295657158,
+ "learning_rate": 7.05654538372171e-06,
+ "loss": 3.46168851852417,
+ "step": 3507
+ },
+ {
+ "epoch": 48.726003490401396,
+ "grad_norm": 0.03597790002822876,
+ "learning_rate": 7.034198253540665e-06,
+ "loss": 3.4690957069396973,
+ "step": 3508
+ },
+ {
+ "epoch": 48.73996509598604,
+ "grad_norm": 0.03684461861848831,
+ "learning_rate": 7.012089577660255e-06,
+ "loss": 3.4837193489074707,
+ "step": 3509
+ },
+ {
+ "epoch": 48.753926701570684,
+ "grad_norm": 0.037007417529821396,
+ "learning_rate": 6.990219373893128e-06,
+ "loss": 3.4740185737609863,
+ "step": 3510
+ },
+ {
+ "epoch": 48.767888307155324,
+ "grad_norm": 0.03697473928332329,
+ "learning_rate": 6.968587659859665e-06,
+ "loss": 3.451822519302368,
+ "step": 3511
+ },
+ {
+ "epoch": 48.781849912739965,
+ "grad_norm": 0.03721608966588974,
+ "learning_rate": 6.9471944529881376e-06,
+ "loss": 3.449659824371338,
+ "step": 3512
+ },
+ {
+ "epoch": 48.795811518324605,
+ "grad_norm": 0.03738904371857643,
+ "learning_rate": 6.926039770514686e-06,
+ "loss": 3.4743895530700684,
+ "step": 3513
+ },
+ {
+ "epoch": 48.80977312390925,
+ "grad_norm": 0.0370398610830307,
+ "learning_rate": 6.905123629483277e-06,
+ "loss": 3.469454050064087,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82373472949389,
+ "grad_norm": 0.0374385230243206,
+ "learning_rate": 6.884446046745608e-06,
+ "loss": 3.446925640106201,
+ "step": 3515
+ },
+ {
+ "epoch": 48.83769633507853,
+ "grad_norm": 0.0373171791434288,
+ "learning_rate": 6.8640070389613055e-06,
+ "loss": 3.4784493446350098,
+ "step": 3516
+ },
+ {
+ "epoch": 48.85165794066317,
+ "grad_norm": 0.03629579022526741,
+ "learning_rate": 6.843806622597661e-06,
+ "loss": 3.467413902282715,
+ "step": 3517
+ },
+ {
+ "epoch": 48.86561954624782,
+ "grad_norm": 0.036762531846761703,
+ "learning_rate": 6.823844813929796e-06,
+ "loss": 3.4539122581481934,
+ "step": 3518
+ },
+ {
+ "epoch": 48.87958115183246,
+ "grad_norm": 0.0365823470056057,
+ "learning_rate": 6.804121629040597e-06,
+ "loss": 3.4581236839294434,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8935427574171,
+ "grad_norm": 0.036474503576755524,
+ "learning_rate": 6.7846370838206455e-06,
+ "loss": 3.435303211212158,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90750436300174,
+ "grad_norm": 0.03734104335308075,
+ "learning_rate": 6.765391193968287e-06,
+ "loss": 3.4876582622528076,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92146596858639,
+ "grad_norm": 0.03731166943907738,
+ "learning_rate": 6.7463839749896335e-06,
+ "loss": 3.468780279159546,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93542757417103,
+ "grad_norm": 0.03610360249876976,
+ "learning_rate": 6.727615442198394e-06,
+ "loss": 3.4627249240875244,
+ "step": 3523
+ },
+ {
+ "epoch": 48.94938917975567,
+ "grad_norm": 0.037647709250450134,
+ "learning_rate": 6.709085610716107e-06,
+ "loss": 3.444429874420166,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96335078534032,
+ "grad_norm": 0.03720395639538765,
+ "learning_rate": 6.690794495471878e-06,
+ "loss": 3.451779365539551,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97731239092496,
+ "grad_norm": 0.036114346235990524,
+ "learning_rate": 6.672742111202543e-06,
+ "loss": 3.4656484127044678,
+ "step": 3526
+ },
+ {
+ "epoch": 48.9912739965096,
+ "grad_norm": 0.03711980581283569,
+ "learning_rate": 6.654928472452569e-06,
+ "loss": 3.4410219192504883,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.029349278658628464,
+ "learning_rate": 6.637353593574088e-06,
+ "loss": 2.167140245437622,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 0.630150556564331,
+ "eval_runtime": 59.6586,
+ "eval_samples_per_second": 40.933,
+ "eval_steps_per_second": 0.654,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01396160558464,
+ "grad_norm": 0.03393097594380379,
+ "learning_rate": 6.620017488726867e-06,
+ "loss": 3.4833078384399414,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02792321116929,
+ "grad_norm": 0.03360513225197792,
+ "learning_rate": 6.602920171878265e-06,
+ "loss": 3.4467785358428955,
+ "step": 3530
+ },
+ {
+ "epoch": 49.04188481675393,
+ "grad_norm": 0.03462687507271767,
+ "learning_rate": 6.5860616568033465e-06,
+ "loss": 3.4624109268188477,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05584642233857,
+ "grad_norm": 0.0335378535091877,
+ "learning_rate": 6.569441957084671e-06,
+ "loss": 3.445650339126587,
+ "step": 3532
+ },
+ {
+ "epoch": 49.06980802792321,
+ "grad_norm": 0.033280305564403534,
+ "learning_rate": 6.5530610861124e-06,
+ "loss": 3.4626822471618652,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083769633507856,
+ "grad_norm": 0.033499252051115036,
+ "learning_rate": 6.536919057084357e-06,
+ "loss": 3.464099168777466,
+ "step": 3534
+ },
+ {
+ "epoch": 49.0977312390925,
+ "grad_norm": 0.03338196128606796,
+ "learning_rate": 6.521015883005866e-06,
+ "loss": 3.450887441635132,
+ "step": 3535
+ },
+ {
+ "epoch": 49.11169284467714,
+ "grad_norm": 0.034279800951480865,
+ "learning_rate": 6.5053515766898195e-06,
+ "loss": 3.478257656097412,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12565445026178,
+ "grad_norm": 0.034303855150938034,
+ "learning_rate": 6.489926150756674e-06,
+ "loss": 3.483400344848633,
+ "step": 3537
+ },
+ {
+ "epoch": 49.139616055846425,
+ "grad_norm": 0.03399111330509186,
+ "learning_rate": 6.4747396176344176e-06,
+ "loss": 3.4596505165100098,
+ "step": 3538
+ },
+ {
+ "epoch": 49.153577661431065,
+ "grad_norm": 0.0343659445643425,
+ "learning_rate": 6.459791989558542e-06,
+ "loss": 3.481147050857544,
+ "step": 3539
+ },
+ {
+ "epoch": 49.167539267015705,
+ "grad_norm": 0.03448395058512688,
+ "learning_rate": 6.445083278572134e-06,
+ "loss": 3.4279236793518066,
+ "step": 3540
+ },
+ {
+ "epoch": 49.181500872600346,
+ "grad_norm": 0.034383561462163925,
+ "learning_rate": 6.430613496525685e-06,
+ "loss": 3.4385128021240234,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19546247818499,
+ "grad_norm": 0.03362111747264862,
+ "learning_rate": 6.416382655077248e-06,
+ "loss": 3.476166009902954,
+ "step": 3542
+ },
+ {
+ "epoch": 49.20942408376963,
+ "grad_norm": 0.03427999094128609,
+ "learning_rate": 6.402390765692381e-06,
+ "loss": 3.4807822704315186,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223385689354274,
+ "grad_norm": 0.034991223365068436,
+ "learning_rate": 6.388637839644073e-06,
+ "loss": 3.469905376434326,
+ "step": 3544
+ },
+ {
+ "epoch": 49.23734729493892,
+ "grad_norm": 0.0341213159263134,
+ "learning_rate": 6.375123888012812e-06,
+ "loss": 3.439946174621582,
+ "step": 3545
+ },
+ {
+ "epoch": 49.25130890052356,
+ "grad_norm": 0.03451673313975334,
+ "learning_rate": 6.361848921686523e-06,
+ "loss": 3.458463191986084,
+ "step": 3546
+ },
+ {
+ "epoch": 49.2652705061082,
+ "grad_norm": 0.034910883754491806,
+ "learning_rate": 6.348812951360662e-06,
+ "loss": 3.459704637527466,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27923211169284,
+ "grad_norm": 0.03429499641060829,
+ "learning_rate": 6.33601598753802e-06,
+ "loss": 3.443666934967041,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29319371727749,
+ "grad_norm": 0.03437327221035957,
+ "learning_rate": 6.323458040528923e-06,
+ "loss": 3.4475464820861816,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30715532286213,
+ "grad_norm": 0.0347493551671505,
+ "learning_rate": 6.311139120451062e-06,
+ "loss": 3.4390041828155518,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32111692844677,
+ "grad_norm": 0.03485412523150444,
+ "learning_rate": 6.299059237229567e-06,
+ "loss": 3.4527699947357178,
+ "step": 3551
+ },
+ {
+ "epoch": 49.33507853403141,
+ "grad_norm": 0.03540089726448059,
+ "learning_rate": 6.287218400596961e-06,
+ "loss": 3.4362831115722656,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34904013961606,
+ "grad_norm": 0.035107024013996124,
+ "learning_rate": 6.27561662009321e-06,
+ "loss": 3.4572930335998535,
+ "step": 3553
+ },
+ {
+ "epoch": 49.3630017452007,
+ "grad_norm": 0.034272242337465286,
+ "learning_rate": 6.264253905065675e-06,
+ "loss": 3.4565186500549316,
+ "step": 3554
+ },
+ {
+ "epoch": 49.37696335078534,
+ "grad_norm": 0.0341477133333683,
+ "learning_rate": 6.2531302646690866e-06,
+ "loss": 3.4383134841918945,
+ "step": 3555
+ },
+ {
+ "epoch": 49.390924956369986,
+ "grad_norm": 0.03484411537647247,
+ "learning_rate": 6.242245707865545e-06,
+ "loss": 3.4494786262512207,
+ "step": 3556
+ },
+ {
+ "epoch": 49.404886561954626,
+ "grad_norm": 0.03493417426943779,
+ "learning_rate": 6.23160024342455e-06,
+ "loss": 3.473158836364746,
+ "step": 3557
+ },
+ {
+ "epoch": 49.41884816753927,
+ "grad_norm": 0.0343041755259037,
+ "learning_rate": 6.221193879922969e-06,
+ "loss": 3.4278626441955566,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43280977312391,
+ "grad_norm": 0.034744180738925934,
+ "learning_rate": 6.211026625745005e-06,
+ "loss": 3.4715216159820557,
+ "step": 3559
+ },
+ {
+ "epoch": 49.446771378708554,
+ "grad_norm": 0.034326206892728806,
+ "learning_rate": 6.201098489082232e-06,
+ "loss": 3.470705986022949,
+ "step": 3560
+ },
+ {
+ "epoch": 49.460732984293195,
+ "grad_norm": 0.03544093295931816,
+ "learning_rate": 6.191409477933554e-06,
+ "loss": 3.457002639770508,
+ "step": 3561
+ },
+ {
+ "epoch": 49.474694589877835,
+ "grad_norm": 0.03487639129161835,
+ "learning_rate": 6.18195960010528e-06,
+ "loss": 3.4577698707580566,
+ "step": 3562
+ },
+ {
+ "epoch": 49.488656195462475,
+ "grad_norm": 0.035099469125270844,
+ "learning_rate": 6.172748863210954e-06,
+ "loss": 3.4326319694519043,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50261780104712,
+ "grad_norm": 0.03431466966867447,
+ "learning_rate": 6.163777274671552e-06,
+ "loss": 3.4559624195098877,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51657940663176,
+ "grad_norm": 0.03481230512261391,
+ "learning_rate": 6.155044841715289e-06,
+ "loss": 3.4380736351013184,
+ "step": 3565
+ },
+ {
+ "epoch": 49.5305410122164,
+ "grad_norm": 0.03508773446083069,
+ "learning_rate": 6.146551571377715e-06,
+ "loss": 3.4580230712890625,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544502617801044,
+ "grad_norm": 0.03545770421624184,
+ "learning_rate": 6.1382974705017456e-06,
+ "loss": 3.464416980743408,
+ "step": 3567
+ },
+ {
+ "epoch": 49.55846422338569,
+ "grad_norm": 0.03599238768219948,
+ "learning_rate": 6.130282545737535e-06,
+ "loss": 3.4564952850341797,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57242582897033,
+ "grad_norm": 0.035288598388433456,
+ "learning_rate": 6.1225068035425725e-06,
+ "loss": 3.4421210289001465,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58638743455497,
+ "grad_norm": 0.035563744604587555,
+ "learning_rate": 6.114970250181616e-06,
+ "loss": 3.4366354942321777,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60034904013962,
+ "grad_norm": 0.034544121474027634,
+ "learning_rate": 6.107672891726724e-06,
+ "loss": 3.4450221061706543,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61431064572426,
+ "grad_norm": 0.0348249189555645,
+ "learning_rate": 6.1006147340572916e-06,
+ "loss": 3.4800610542297363,
+ "step": 3572
+ },
+ {
+ "epoch": 49.6282722513089,
+ "grad_norm": 0.03489942103624344,
+ "learning_rate": 6.093795782859917e-06,
+ "loss": 3.4481005668640137,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64223385689354,
+ "grad_norm": 0.03538931906223297,
+ "learning_rate": 6.087216043628502e-06,
+ "loss": 3.449239730834961,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65619546247819,
+ "grad_norm": 0.03479275852441788,
+ "learning_rate": 6.0808755216642146e-06,
+ "loss": 3.461120367050171,
+ "step": 3575
+ },
+ {
+ "epoch": 49.67015706806283,
+ "grad_norm": 0.0360809788107872,
+ "learning_rate": 6.07477422207556e-06,
+ "loss": 3.4648733139038086,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68411867364747,
+ "grad_norm": 0.03543730452656746,
+ "learning_rate": 6.06891214977818e-06,
+ "loss": 3.447707176208496,
+ "step": 3577
+ },
+ {
+ "epoch": 49.69808027923211,
+ "grad_norm": 0.03517023101449013,
+ "learning_rate": 6.063289309495051e-06,
+ "loss": 3.4664130210876465,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712041884816756,
+ "grad_norm": 0.03461436927318573,
+ "learning_rate": 6.057905705756421e-06,
+ "loss": 3.460075616836548,
+ "step": 3579
+ },
+ {
+ "epoch": 49.726003490401396,
+ "grad_norm": 0.03460308536887169,
+ "learning_rate": 6.052761342899737e-06,
+ "loss": 3.4897618293762207,
+ "step": 3580
+ },
+ {
+ "epoch": 49.73996509598604,
+ "grad_norm": 0.03601457551121712,
+ "learning_rate": 6.0478562250697185e-06,
+ "loss": 3.4659245014190674,
+ "step": 3581
+ },
+ {
+ "epoch": 49.753926701570684,
+ "grad_norm": 0.035826582461595535,
+ "learning_rate": 6.043190356218355e-06,
+ "loss": 3.4371423721313477,
+ "step": 3582
+ },
+ {
+ "epoch": 49.767888307155324,
+ "grad_norm": 0.03498508781194687,
+ "learning_rate": 6.038763740104835e-06,
+ "loss": 3.444263458251953,
+ "step": 3583
+ },
+ {
+ "epoch": 49.781849912739965,
+ "grad_norm": 0.0350525937974453,
+ "learning_rate": 6.034576380295586e-06,
+ "loss": 3.4409122467041016,
+ "step": 3584
+ },
+ {
+ "epoch": 49.795811518324605,
+ "grad_norm": 0.03666655346751213,
+ "learning_rate": 6.030628280164338e-06,
+ "loss": 3.48353910446167,
+ "step": 3585
+ },
+ {
+ "epoch": 49.80977312390925,
+ "grad_norm": 0.03535056859254837,
+ "learning_rate": 6.026919442891958e-06,
+ "loss": 3.4651553630828857,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82373472949389,
+ "grad_norm": 0.03517733886837959,
+ "learning_rate": 6.023449871466615e-06,
+ "loss": 3.4254558086395264,
+ "step": 3587
+ },
+ {
+ "epoch": 49.83769633507853,
+ "grad_norm": 0.03725713491439819,
+ "learning_rate": 6.020219568683678e-06,
+ "loss": 3.4693264961242676,
+ "step": 3588
+ },
+ {
+ "epoch": 49.85165794066317,
+ "grad_norm": 0.03621108829975128,
+ "learning_rate": 6.0172285371456925e-06,
+ "loss": 3.489861011505127,
+ "step": 3589
+ },
+ {
+ "epoch": 49.86561954624782,
+ "grad_norm": 0.035663098096847534,
+ "learning_rate": 6.014476779262568e-06,
+ "loss": 3.460881233215332,
+ "step": 3590
+ },
+ {
+ "epoch": 49.87958115183246,
+ "grad_norm": 0.03572768345475197,
+ "learning_rate": 6.011964297251285e-06,
+ "loss": 3.4626212120056152,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8935427574171,
+ "grad_norm": 0.03535124287009239,
+ "learning_rate": 6.0096910931361275e-06,
+ "loss": 3.444488525390625,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90750436300174,
+ "grad_norm": 0.035609010607004166,
+ "learning_rate": 6.007657168748551e-06,
+ "loss": 3.44132399559021,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92146596858639,
+ "grad_norm": 0.036064308136701584,
+ "learning_rate": 6.005862525727277e-06,
+ "loss": 3.4687745571136475,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93542757417103,
+ "grad_norm": 0.036159414798021317,
+ "learning_rate": 6.0043071655181986e-06,
+ "loss": 3.4495744705200195,
+ "step": 3595
+ },
+ {
+ "epoch": 49.94938917975567,
+ "grad_norm": 0.035239290446043015,
+ "learning_rate": 6.002991089374446e-06,
+ "loss": 3.456461191177368,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96335078534032,
+ "grad_norm": 0.0361039936542511,
+ "learning_rate": 6.0019142983563875e-06,
+ "loss": 3.4544308185577393,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97731239092496,
+ "grad_norm": 0.036664027720689774,
+ "learning_rate": 6.001076793331555e-06,
+ "loss": 3.4486136436462402,
+ "step": 3598
+ },
+ {
+ "epoch": 49.9912739965096,
+ "grad_norm": 0.03499891236424446,
+ "learning_rate": 6.000478574974689e-06,
+ "loss": 3.4493560791015625,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.02878538891673088,
+ "learning_rate": 6.000119643767767e-06,
+ "loss": 2.17130708694458,
+ "step": 3600
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": true
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.5366833001883238e+18,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-3600/training_args.bin b/runs/l2r90-ssl/checkpoint-3600/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..103048067ef8897d0386013ed89ce6764bad5521
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-3600/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f320fb44ea3a4a93198f1a6c2405447428148ce142cd94ee283a3c077a327c2d
+size 4792
diff --git a/runs/l2r90-ssl/checkpoint-72/chat_template.jinja b/runs/l2r90-ssl/checkpoint-72/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-72/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-ssl/checkpoint-72/config.json b/runs/l2r90-ssl/checkpoint-72/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-72/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-ssl/checkpoint-72/generation_config.json b/runs/l2r90-ssl/checkpoint-72/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-72/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-ssl/checkpoint-72/model.safetensors b/runs/l2r90-ssl/checkpoint-72/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..be5ace123e969cd7305836ed688a06b8a75374af
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-72/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:93fa7c74375e5402f705eda9aff05984d24c4d2bdf5aee62fe62f8590b53830d
+size 583356232
diff --git a/runs/l2r90-ssl/checkpoint-72/optimizer.pt b/runs/l2r90-ssl/checkpoint-72/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..7d9bb827ced5a294f134168c35837378cf33295d
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-72/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:00035386c983418a9080a5ea1dceb2194cf89c81ccdc2f5d9fa45f8d668a2b6a
+size 1166825338
diff --git a/runs/l2r90-ssl/checkpoint-72/rng_state_0.pth b/runs/l2r90-ssl/checkpoint-72/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..5631561581b3634362004f6c036145ee8e54d0f4
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-72/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b184f8fc9facd6e4a6513602952b43b9446784300c3ea2da72aefb029f93aac1
+size 15024
diff --git a/runs/l2r90-ssl/checkpoint-72/rng_state_1.pth b/runs/l2r90-ssl/checkpoint-72/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..5c9399d17854873e90933eddfa9944d5e3d0b38f
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-72/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:937f07b3f415f537fbc809c70f15f64b8410cd6a07adbe7d246f5ca7359b3b49
+size 15024
diff --git a/runs/l2r90-ssl/checkpoint-72/rng_state_2.pth b/runs/l2r90-ssl/checkpoint-72/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..8a986c37b53237a7bde101804a5602d204f737ba
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-72/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:07e3f7eacec877950b3dad99b106bf80da30203dda6159b06f954c15b7f5d367
+size 15024
diff --git a/runs/l2r90-ssl/checkpoint-72/rng_state_3.pth b/runs/l2r90-ssl/checkpoint-72/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..3ec45d94b0661de4e43586eec83b2caf40c79a8e
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-72/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2a5087a11a5f0a226f1607ec68740ea6baa2599559138de38da40f89dd666683
+size 15024
diff --git a/runs/l2r90-ssl/checkpoint-72/scheduler.pt b/runs/l2r90-ssl/checkpoint-72/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..43cf114590f1088fd438450226c42885d843da41
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-72/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:878ec53f5c2a524fc37af578b94a9025923faefea55f621be8e5643a79915d0f
+size 1064
diff --git a/runs/l2r90-ssl/checkpoint-72/tokenizer.json b/runs/l2r90-ssl/checkpoint-72/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-72/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-ssl/checkpoint-72/tokenizer_config.json b/runs/l2r90-ssl/checkpoint-72/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-72/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-72/trainer_state.json b/runs/l2r90-ssl/checkpoint-72/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..7855e945c3417118c8662ce1f06a4f8097a5b6f6
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-72/trainer_state.json
@@ -0,0 +1,538 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 1.0,
+ "eval_steps": 500,
+ "global_step": 72,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.2588357925415039,
+ "learning_rate": 0.0,
+ "loss": 3.9982247352600098,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.25455471873283386,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.998683452606201,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.2555367052555084,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.9821574687957764,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.2383691370487213,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.9531917572021484,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2023179680109024,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.9148876667022705,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.17691479623317719,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.875638961791992,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.14372293651103973,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.836796283721924,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1074419692158699,
+ "learning_rate": 4.2e-05,
+ "loss": 3.8036489486694336,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.08204353600740433,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.774808406829834,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.0632256343960762,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.7503671646118164,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.047985248267650604,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.7345473766326904,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.03675895184278488,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.721525192260742,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.02908993512392044,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.713454246520996,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.02443614974617958,
+ "learning_rate": 7.8e-05,
+ "loss": 3.7078704833984375,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02109951712191105,
+ "learning_rate": 8.4e-05,
+ "loss": 3.7034401893615723,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.018616363406181335,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.7009854316711426,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.016399752348661423,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.697768211364746,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.014256286434829235,
+ "learning_rate": 0.000102,
+ "loss": 3.6974306106567383,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.013020666316151619,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.694406509399414,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.011879238300025463,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.691807270050049,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.010290366597473621,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.6937897205352783,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.009482803754508495,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.6902384757995605,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.008425266481935978,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.686815023422241,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.007676936220377684,
+ "learning_rate": 0.000138,
+ "loss": 3.685192823410034,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.007491032127290964,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.6839442253112793,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.007428477052599192,
+ "learning_rate": 0.00015,
+ "loss": 3.6783409118652344,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.006965892389416695,
+ "learning_rate": 0.000156,
+ "loss": 3.6813433170318604,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.008106694556772709,
+ "learning_rate": 0.000162,
+ "loss": 3.6741726398468018,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.0076780058443546295,
+ "learning_rate": 0.000168,
+ "loss": 3.67673921585083,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.007904314436018467,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.6724467277526855,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.008422375656664371,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.673013925552368,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00811714306473732,
+ "learning_rate": 0.000186,
+ "loss": 3.6698389053344727,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.009063836187124252,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.668990135192871,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.01878332532942295,
+ "learning_rate": 0.000198,
+ "loss": 3.6658849716186523,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.01812782697379589,
+ "learning_rate": 0.000204,
+ "loss": 3.6697864532470703,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.020587686449289322,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.662757158279419,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.01515690516680479,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.6558732986450195,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.02510508894920349,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.6495113372802734,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.020487826317548752,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.638803243637085,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.017446046695113182,
+ "learning_rate": 0.000234,
+ "loss": 3.633307456970215,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.030667584389448166,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.6277506351470947,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.023789944127202034,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.6189804077148438,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.02660694345831871,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.610522747039795,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.023814303800463676,
+ "learning_rate": 0.000258,
+ "loss": 3.615910291671753,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.021863561123609543,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.612008571624756,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03705216199159622,
+ "learning_rate": 0.00027,
+ "loss": 3.6066887378692627,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.024143919348716736,
+ "learning_rate": 0.000276,
+ "loss": 3.5987331867218018,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.03251276910305023,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.595367431640625,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06619632989168167,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.59018611907959,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.042714253067970276,
+ "learning_rate": 0.000294,
+ "loss": 3.58981990814209,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.018893254920840263,
+ "learning_rate": 0.0003,
+ "loss": 3.578472375869751,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0749685987830162,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.5828654766082764,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.12142857164144516,
+ "learning_rate": 0.000312,
+ "loss": 3.5840892791748047,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.06944665312767029,
+ "learning_rate": 0.000318,
+ "loss": 3.580961227416992,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.04545000568032265,
+ "learning_rate": 0.000324,
+ "loss": 3.5674569606781006,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.061073049902915955,
+ "learning_rate": 0.00033,
+ "loss": 3.569991111755371,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.04133498668670654,
+ "learning_rate": 0.000336,
+ "loss": 3.5657706260681152,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.055982090532779694,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.559979200363159,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.05259208381175995,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.558039426803589,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.029976001009345055,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.5491275787353516,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.04998327046632767,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.5406670570373535,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.03230690956115723,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.5391530990600586,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.06102178618311882,
+ "learning_rate": 0.000372,
+ "loss": 3.5311684608459473,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.031715717166662216,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.518899440765381,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.06841139495372772,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.514200210571289,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.10323675721883774,
+ "learning_rate": 0.00039,
+ "loss": 3.513646125793457,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.1448228508234024,
+ "learning_rate": 0.000396,
+ "loss": 3.5110087394714355,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.12129563093185425,
+ "learning_rate": 0.000402,
+ "loss": 3.501497507095337,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.030224842950701714,
+ "learning_rate": 0.000408,
+ "loss": 3.4870352745056152,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.07492101192474365,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.482234001159668,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.03942760080099106,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.4746525287628174,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.04967200756072998,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.604069232940674,
+ "step": 72
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.0776560164274176e+16,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-72/training_args.bin b/runs/l2r90-ssl/checkpoint-72/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5924a9264314215faa31ea62b22ba59c4573b542
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-72/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2e206f0318237cba8820486751e683a8c1e9fd84a823416f09cf3e109aba53ea
+size 4792
diff --git a/runs/l2r90-ssl/checkpoint-720/chat_template.jinja b/runs/l2r90-ssl/checkpoint-720/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-720/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-ssl/checkpoint-720/config.json b/runs/l2r90-ssl/checkpoint-720/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-720/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-ssl/checkpoint-720/generation_config.json b/runs/l2r90-ssl/checkpoint-720/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-720/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-ssl/checkpoint-720/model.safetensors b/runs/l2r90-ssl/checkpoint-720/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..25cb301c083afec0c80dac68bafa58220fd25b98
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-720/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3a7c0969e46d78532ab9810a2b4c74fa390baf371eefa47a29e5c47e619ce0b5
+size 583356232
diff --git a/runs/l2r90-ssl/checkpoint-720/optimizer.pt b/runs/l2r90-ssl/checkpoint-720/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..c1ed9ac5d3436e3203cc7174978cc1de811dbec3
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-720/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ac957652bd320c3d9e57b22d37da92b04ac762d48d46b59a1085d43c66cd7186
+size 1166825338
diff --git a/runs/l2r90-ssl/checkpoint-720/rng_state_0.pth b/runs/l2r90-ssl/checkpoint-720/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..0977c0bf00c01a46d716db730a3c9bcc82905aaf
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-720/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a02d728a763af949adfec4bf1d121e1daf1ef655fd3b17a42c2399e0221768dd
+size 15024
diff --git a/runs/l2r90-ssl/checkpoint-720/rng_state_1.pth b/runs/l2r90-ssl/checkpoint-720/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..4e5cd4f14d76f3fe2c0762093e6bb01d818def44
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-720/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d60087aa7d44f5751e6a709354a90b705404a21a9437680f20377cf12743341b
+size 15024
diff --git a/runs/l2r90-ssl/checkpoint-720/rng_state_2.pth b/runs/l2r90-ssl/checkpoint-720/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..1dbfc9354146f87ee95d6d46bdc97b42ae7b6f65
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-720/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1c962bfcedafeb4fe574f63e648405b9b28a01333d49c391529b2c7a4fead463
+size 15024
diff --git a/runs/l2r90-ssl/checkpoint-720/rng_state_3.pth b/runs/l2r90-ssl/checkpoint-720/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..0a96d4dec11fa27b14ffd309db2393d72f178feb
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-720/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:478ec022631f64178ab44a8355b560b75c942f4cf348cc636a8cb37ec3bec6cc
+size 15024
diff --git a/runs/l2r90-ssl/checkpoint-720/scheduler.pt b/runs/l2r90-ssl/checkpoint-720/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..53133f8d1ca0990636972b84579aec1212febb2a
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-720/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ea248f5cdc3b9f2d6793617acdaa11d93bab34c0a87ea94ea6aaabc148225da6
+size 1064
diff --git a/runs/l2r90-ssl/checkpoint-720/tokenizer.json b/runs/l2r90-ssl/checkpoint-720/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-720/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-ssl/checkpoint-720/tokenizer_config.json b/runs/l2r90-ssl/checkpoint-720/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-720/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-720/trainer_state.json b/runs/l2r90-ssl/checkpoint-720/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..e6d461efba13fb59adbf58d5bbd7f896d65164f4
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-720/trainer_state.json
@@ -0,0 +1,5146 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 10.0,
+ "eval_steps": 500,
+ "global_step": 720,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.2588357925415039,
+ "learning_rate": 0.0,
+ "loss": 3.9982247352600098,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.25455471873283386,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.998683452606201,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.2555367052555084,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.9821574687957764,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.2383691370487213,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.9531917572021484,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2023179680109024,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.9148876667022705,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.17691479623317719,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.875638961791992,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.14372293651103973,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.836796283721924,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1074419692158699,
+ "learning_rate": 4.2e-05,
+ "loss": 3.8036489486694336,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.08204353600740433,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.774808406829834,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.0632256343960762,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.7503671646118164,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.047985248267650604,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.7345473766326904,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.03675895184278488,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.721525192260742,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.02908993512392044,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.713454246520996,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.02443614974617958,
+ "learning_rate": 7.8e-05,
+ "loss": 3.7078704833984375,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02109951712191105,
+ "learning_rate": 8.4e-05,
+ "loss": 3.7034401893615723,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.018616363406181335,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.7009854316711426,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.016399752348661423,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.697768211364746,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.014256286434829235,
+ "learning_rate": 0.000102,
+ "loss": 3.6974306106567383,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.013020666316151619,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.694406509399414,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.011879238300025463,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.691807270050049,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.010290366597473621,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.6937897205352783,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.009482803754508495,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.6902384757995605,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.008425266481935978,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.686815023422241,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.007676936220377684,
+ "learning_rate": 0.000138,
+ "loss": 3.685192823410034,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.007491032127290964,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.6839442253112793,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.007428477052599192,
+ "learning_rate": 0.00015,
+ "loss": 3.6783409118652344,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.006965892389416695,
+ "learning_rate": 0.000156,
+ "loss": 3.6813433170318604,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.008106694556772709,
+ "learning_rate": 0.000162,
+ "loss": 3.6741726398468018,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.0076780058443546295,
+ "learning_rate": 0.000168,
+ "loss": 3.67673921585083,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.007904314436018467,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.6724467277526855,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.008422375656664371,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.673013925552368,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00811714306473732,
+ "learning_rate": 0.000186,
+ "loss": 3.6698389053344727,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.009063836187124252,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.668990135192871,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.01878332532942295,
+ "learning_rate": 0.000198,
+ "loss": 3.6658849716186523,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.01812782697379589,
+ "learning_rate": 0.000204,
+ "loss": 3.6697864532470703,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.020587686449289322,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.662757158279419,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.01515690516680479,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.6558732986450195,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.02510508894920349,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.6495113372802734,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.020487826317548752,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.638803243637085,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.017446046695113182,
+ "learning_rate": 0.000234,
+ "loss": 3.633307456970215,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.030667584389448166,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.6277506351470947,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.023789944127202034,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.6189804077148438,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.02660694345831871,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.610522747039795,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.023814303800463676,
+ "learning_rate": 0.000258,
+ "loss": 3.615910291671753,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.021863561123609543,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.612008571624756,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03705216199159622,
+ "learning_rate": 0.00027,
+ "loss": 3.6066887378692627,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.024143919348716736,
+ "learning_rate": 0.000276,
+ "loss": 3.5987331867218018,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.03251276910305023,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.595367431640625,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06619632989168167,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.59018611907959,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.042714253067970276,
+ "learning_rate": 0.000294,
+ "loss": 3.58981990814209,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.018893254920840263,
+ "learning_rate": 0.0003,
+ "loss": 3.578472375869751,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0749685987830162,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.5828654766082764,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.12142857164144516,
+ "learning_rate": 0.000312,
+ "loss": 3.5840892791748047,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.06944665312767029,
+ "learning_rate": 0.000318,
+ "loss": 3.580961227416992,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.04545000568032265,
+ "learning_rate": 0.000324,
+ "loss": 3.5674569606781006,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.061073049902915955,
+ "learning_rate": 0.00033,
+ "loss": 3.569991111755371,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.04133498668670654,
+ "learning_rate": 0.000336,
+ "loss": 3.5657706260681152,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.055982090532779694,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.559979200363159,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.05259208381175995,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.558039426803589,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.029976001009345055,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.5491275787353516,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.04998327046632767,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.5406670570373535,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.03230690956115723,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.5391530990600586,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.06102178618311882,
+ "learning_rate": 0.000372,
+ "loss": 3.5311684608459473,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.031715717166662216,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.518899440765381,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.06841139495372772,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.514200210571289,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.10323675721883774,
+ "learning_rate": 0.00039,
+ "loss": 3.513646125793457,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.1448228508234024,
+ "learning_rate": 0.000396,
+ "loss": 3.5110087394714355,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.12129563093185425,
+ "learning_rate": 0.000402,
+ "loss": 3.501497507095337,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.030224842950701714,
+ "learning_rate": 0.000408,
+ "loss": 3.4870352745056152,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.07492101192474365,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.482234001159668,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.03942760080099106,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.4746525287628174,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.04967200756072998,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.604069232940674,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.8671882152557373,
+ "eval_runtime": 42.6602,
+ "eval_samples_per_second": 57.243,
+ "eval_steps_per_second": 0.469,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.05385325849056244,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.462052822113037,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.04870929941534996,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.459104537963867,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.08701689541339874,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.4533209800720215,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.12219952791929245,
+ "learning_rate": 0.00045,
+ "loss": 3.4558796882629395,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.0823654904961586,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.443049192428589,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.12765446305274963,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.4373486042022705,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.14215047657489777,
+ "learning_rate": 0.000468,
+ "loss": 3.449429988861084,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.03649887070059776,
+ "learning_rate": 0.000474,
+ "loss": 3.42877197265625,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09551949054002762,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.4352874755859375,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.030483486130833626,
+ "learning_rate": 0.000486,
+ "loss": 3.4125874042510986,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.0745147094130516,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.4153835773468018,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.06620422005653381,
+ "learning_rate": 0.000498,
+ "loss": 3.4053854942321777,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.06098725646734238,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.404346466064453,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.047781601548194885,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.3991761207580566,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.04409855231642723,
+ "learning_rate": 0.000516,
+ "loss": 3.396144390106201,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.04355994611978531,
+ "learning_rate": 0.000522,
+ "loss": 3.3928911685943604,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.0670008510351181,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.3831443786621094,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.0812622532248497,
+ "learning_rate": 0.000534,
+ "loss": 3.382413148880005,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.0632157102227211,
+ "learning_rate": 0.00054,
+ "loss": 3.3784892559051514,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.028686825186014175,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.3673207759857178,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.034818850457668304,
+ "learning_rate": 0.000552,
+ "loss": 3.3648972511291504,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.04509204626083374,
+ "learning_rate": 0.000558,
+ "loss": 3.35561203956604,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.054211195558309555,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.3612167835235596,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.03752049803733826,
+ "learning_rate": 0.00057,
+ "loss": 3.357103109359741,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.03620241954922676,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.3469700813293457,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.02454194985330105,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.3429036140441895,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.041030462831258774,
+ "learning_rate": 0.000588,
+ "loss": 3.338247299194336,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.03266424685716629,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.332688331604004,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.04730091616511345,
+ "learning_rate": 0.0006,
+ "loss": 3.335902452468872,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.065738245844841,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.3289852142333984,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.09062394499778748,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.3278567790985107,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.1482805460691452,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.332685947418213,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.20066532492637634,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.363180637359619,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.11500677466392517,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.3308753967285156,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.11754138022661209,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.3383703231811523,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.09672239422798157,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.3267624378204346,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.07190799713134766,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.317737579345703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.0815449133515358,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.3198161125183105,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09612327069044113,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.315814971923828,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.14795289933681488,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.317605495452881,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06667589396238327,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.3127357959747314,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.058537885546684265,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.319807767868042,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.06428155303001404,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.297541618347168,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.03881775587797165,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.297539710998535,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.061801470816135406,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.290297508239746,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.07642851769924164,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.284428596496582,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08123870939016342,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.2820234298706055,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07174684852361679,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.2697296142578125,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.04030010476708412,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.264328956604004,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.05662153288722038,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.2699337005615234,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.03761111572384834,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.255704879760742,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.043027132749557495,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.2579030990600586,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.044409338384866714,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.248547077178955,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.0239730142056942,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.249307632446289,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03628705441951752,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.2451741695404053,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.029564054682850838,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.238445997238159,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.032036975026130676,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.2320892810821533,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.030085794627666473,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.231468915939331,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.025232303887605667,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.232199192047119,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.027053389698266983,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.210745334625244,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.028464114293456078,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.210076332092285,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04616783559322357,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.2103490829467773,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.039081938564777374,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.202662467956543,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.03428936004638672,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.1997475624084473,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.04351917654275894,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.19844913482666,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.0840945914387703,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.2013347148895264,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.36558300256729126,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.26310396194458,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.09432770311832428,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.2881736755371094,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.061970267444849014,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.2840280532836914,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.05112256482243538,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.2632040977478027,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06797203421592712,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.254606246948242,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.051026541739702225,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.429918050765991,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.8117228746414185,
+ "eval_runtime": 42.8572,
+ "eval_samples_per_second": 56.98,
+ "eval_steps_per_second": 0.467,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.0608622208237648,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.2348179817199707,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.05581629276275635,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.22896146774292,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.09162136912345886,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.2343664169311523,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.13750384747982025,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.2507717609405518,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.0872517079114914,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.2440056800842285,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.07161762565374374,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.229494571685791,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04082738608121872,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.2183239459991455,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.04972497373819351,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.218075752258301,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.049254074692726135,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.213379383087158,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.056671254336833954,
+ "learning_rate": 0.000599663984012462,
+ "loss": 3.209568977355957,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.05541909113526344,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.207760810852051,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.03819117695093155,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 3.199732780456543,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.0529443584382534,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.1945419311523438,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.043010979890823364,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.185936212539673,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.07311078161001205,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.189894914627075,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06110258400440216,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 3.1911773681640625,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.06880149245262146,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 3.187437057495117,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.051326874643564224,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 3.1826558113098145,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.03363822400569916,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 3.175527572631836,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.10268989205360413,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 3.172367572784424,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.09723010659217834,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 3.1873984336853027,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.04129317030310631,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 3.165799617767334,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.05292229354381561,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 3.171780824661255,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.03873143345117569,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 3.1680874824523926,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.047018859535455704,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 3.161402940750122,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.03942955285310745,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 3.1530532836914062,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.039801858365535736,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 3.1465344429016113,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.03795155510306358,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 3.1364030838012695,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.0342850424349308,
+ "learning_rate": 0.000599379982511273,
+ "loss": 3.1391525268554688,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.03243280574679375,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 3.1331253051757812,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.029679352417588234,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 3.1346120834350586,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.030230293050408363,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 3.132615804672241,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.02845083735883236,
+ "learning_rate": 0.000599309205504528,
+ "loss": 3.1242547035217285,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.03208369016647339,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 3.119110584259033,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.040268607437610626,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 3.113833427429199,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.06951889395713806,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 3.1186914443969727,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.08879823237657547,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 3.1140987873077393,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07374681532382965,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 3.1147828102111816,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.038190409541130066,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 3.1161253452301025,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.03297388181090355,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 3.1006083488464355,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.05905177444219589,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 3.106564521789551,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.037567321211099625,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 3.0884456634521484,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.02725658379495144,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 3.09000825881958,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.04272424429655075,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 3.090528964996338,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.027587950229644775,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 3.085131883621216,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.03282274678349495,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 3.0880112648010254,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.04064026474952698,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 3.079259157180786,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.023374097421765327,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 3.0747299194335938,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.025003870949149132,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 3.0721259117126465,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.030445197597146034,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 3.0733323097229004,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.025935636833310127,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 3.070709228515625,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.021337421610951424,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 3.0627551078796387,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.040481697767972946,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 3.0529394149780273,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.047608256340026855,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 3.041020631790161,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.03356955572962761,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 3.0532307624816895,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.02101682312786579,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 3.0492985248565674,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.030098581686615944,
+ "learning_rate": 0.000598804365316586,
+ "loss": 3.0527801513671875,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.040166422724723816,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 3.0505762100219727,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.051579318940639496,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 3.0422110557556152,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.06485334783792496,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 3.0435101985931396,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.055590901523828506,
+ "learning_rate": 0.000598706872384039,
+ "loss": 3.0448739528656006,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.03183475881814957,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 3.0416858196258545,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.03995629772543907,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 3.0255980491638184,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.03907827287912369,
+ "learning_rate": 0.000598631251036868,
+ "loss": 3.032339096069336,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.027681387960910797,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 3.025439739227295,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.03470184653997421,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 3.02883243560791,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.031407181173563004,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 3.0289883613586426,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.0371522456407547,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 3.013608932495117,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.03651345893740654,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 3.0120253562927246,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.06004595756530762,
+ "learning_rate": 0.000598473577921366,
+ "loss": 3.017958402633667,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.09979518502950668,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 3.023041248321533,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.10138462483882904,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.275729179382324,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.7571707367897034,
+ "eval_runtime": 42.8283,
+ "eval_samples_per_second": 57.018,
+ "eval_steps_per_second": 0.467,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.04982660710811615,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 3.016209602355957,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.06532539427280426,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 3.015043020248413,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.04369623214006424,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 3.0096778869628906,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.05180002748966217,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 3.0121543407440186,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.0440233014523983,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.9937305450439453,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.05645671859383583,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 3.0023865699768066,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.03990716487169266,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 3.0003232955932617,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.03818608075380325,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.994137763977051,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.035388972610235214,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.993170738220215,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.03162284567952156,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.9922614097595215,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.029497336596250534,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.9852986335754395,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.02586926706135273,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.9798672199249268,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.027355050668120384,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.985158920288086,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.0341978520154953,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.982581615447998,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.030727429315447807,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.9701411724090576,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.04429870471358299,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.9789254665374756,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.05775132030248642,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.9758143424987793,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.05635962635278702,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.981480121612549,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.03639192506670952,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.968752861022949,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.028748342767357826,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.9584972858428955,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.035653006285429,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.967545747756958,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.03697885945439339,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.9578099250793457,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.03518654406070709,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.9680328369140625,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.034762222319841385,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.9670262336730957,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03813016042113304,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.957522392272949,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.04372904822230339,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.9540514945983887,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.06646446138620377,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.951028823852539,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.07831092923879623,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.95898699760437,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.05779588967561722,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.965884208679199,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.04414296895265579,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.9510560035705566,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.04797815904021263,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.955803394317627,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.04482145980000496,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.954059362411499,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.037104446440935135,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.94792103767395,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.03616701066493988,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.9500932693481445,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.03882797807455063,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.9558305740356445,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.04356204345822334,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.9237029552459717,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.04535257816314697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.925853729248047,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.04519587755203247,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.924984931945801,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.055922407656908035,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.932561159133911,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.06330762803554535,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.9388322830200195,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.0517115443944931,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.9272942543029785,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.057369694113731384,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.9201431274414062,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.05687880516052246,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.922886371612549,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.055525410920381546,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.9211201667785645,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04503652825951576,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.9164583683013916,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.03385302051901817,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.9145395755767822,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03564384952187538,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.916003704071045,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.04188712686300278,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.915839433670044,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03657370060682297,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.903120279312134,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.03126626834273338,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.895824432373047,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.04196290299296379,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.9093422889709473,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.1706087589263916,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.9183316230773926,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.05810421332716942,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.911245107650757,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.061330314725637436,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.9159231185913086,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.06062281131744385,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.9211535453796387,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.05590800568461418,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.9140496253967285,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.04496239125728607,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.9040920734405518,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.03835335746407509,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.9026548862457275,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.05137286335229874,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.889634132385254,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.04091101884841919,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.886838674545288,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.04076292738318443,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.892918109893799,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.045437734574079514,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.893305778503418,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.041114673018455505,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.88517165184021,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.03600773215293884,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.8779795169830322,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.04311220720410347,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.87247896194458,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.04818756878376007,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.889103412628174,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.053460877388715744,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.8899683952331543,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.06763825565576553,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.8823537826538086,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.07304801046848297,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.877340316772461,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.061337269842624664,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.8820111751556396,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.04660911113023758,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.868886709213257,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.053929202258586884,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.159337043762207,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.721762478351593,
+ "eval_runtime": 42.7977,
+ "eval_samples_per_second": 57.059,
+ "eval_steps_per_second": 0.467,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.04259369149804115,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.859579086303711,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.044425517320632935,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.8681139945983887,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.04301869496703148,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.853911876678467,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.04574296995997429,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.8724427223205566,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.050247322767972946,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.8528172969818115,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.046683840453624725,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.853405714035034,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.05966169387102127,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.8605127334594727,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.08935574442148209,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.862339973449707,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.08019179105758667,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.8655967712402344,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.04753892868757248,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.8657939434051514,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06868117302656174,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.861976146697998,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.07040470093488693,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.8562135696411133,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.05683179199695587,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.8576629161834717,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.04918001592159271,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.861943244934082,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.04921915754675865,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.8495378494262695,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.037651967257261276,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.8428618907928467,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.03522057458758354,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.8372769355773926,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.03111186809837818,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.8377151489257812,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.03906773030757904,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.846196174621582,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.02546556107699871,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.8288016319274902,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.03119821846485138,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.8334617614746094,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.030583377927541733,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.8209362030029297,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.029693076387047768,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.8267486095428467,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.028978517279028893,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.819244861602783,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.027871523052453995,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.8289647102355957,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.029287064447999,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.8235747814178467,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.024990517646074295,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.819016218185425,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.030980689451098442,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.828155040740967,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03721801936626434,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.813418388366699,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03738148882985115,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.808328866958618,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.03580411896109581,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.812070846557617,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.039036594331264496,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.807159423828125,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.044666968286037445,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.799992561340332,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.055589817464351654,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.804142951965332,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.06139020994305611,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.8182730674743652,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.07769318670034409,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.8166518211364746,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.06521332263946533,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.8142249584198,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.09083619713783264,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.8166933059692383,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.09737467020750046,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.8111555576324463,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.07148188352584839,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.8168413639068604,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.055516090244054794,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.8183553218841553,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.05338391661643982,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.8127028942108154,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.040785472840070724,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.8124775886535645,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.037135496735572815,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.8065357208251953,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.032455578446388245,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.798459053039551,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.03465662896633148,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.798649549484253,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.032026033848524094,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.7906997203826904,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.026019956916570663,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.791969060897827,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.030298635363578796,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.784593105316162,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.028147678822278976,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.782454013824463,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.023933131247758865,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.787851333618164,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.026789527386426926,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.788067579269409,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.024554884061217308,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.777033567428589,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.028679484501481056,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.786342144012451,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.026410218328237534,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.7790610790252686,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.027413051575422287,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.778944969177246,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.02908242493867874,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.763850212097168,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.029632994905114174,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.7672982215881348,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.027458716183900833,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.7702178955078125,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.03263958916068077,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.7581803798675537,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04427146911621094,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.768141508102417,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.07142394036054611,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.75819730758667,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.08833584934473038,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.782716751098633,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.08168833702802658,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.7763240337371826,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07253874838352203,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.7717690467834473,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.08186902105808258,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.767303466796875,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.05854064226150513,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.761544704437256,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.04602917283773422,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.7669177055358887,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.03920970857143402,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.766085624694824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.04551195353269577,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.7623450756073,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.040861159563064575,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.7471840381622314,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.026907291263341904,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 2.0710244178771973,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6934371590614319,
+ "eval_runtime": 42.6015,
+ "eval_samples_per_second": 57.322,
+ "eval_steps_per_second": 0.469,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.03801899403333664,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.739086627960205,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.03707238286733627,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.741638422012329,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.0387551411986351,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.740777015686035,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.028010359033942223,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.7353010177612305,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.026474518701434135,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.73842453956604,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.02920856513082981,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.736295461654663,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.026074763387441635,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.7295751571655273,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.026862047612667084,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.723040819168091,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.026300370693206787,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.7199485301971436,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.026030950248241425,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.7276217937469482,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.028342435136437416,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.7193045616149902,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.032135091722011566,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.7143118381500244,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.03068603202700615,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.7191050052642822,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.035828713327646255,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.718085765838623,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.03343002125620842,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.707131862640381,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.035051796585321426,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.7091879844665527,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.03395913913846016,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6980862617492676,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.034198079258203506,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.699704647064209,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.0322839580476284,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.702791213989258,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.041531339287757874,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.7018051147460938,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.040806882083415985,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.7096598148345947,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.04156940430402756,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6956775188446045,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.05608797073364258,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.70127272605896,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0729481428861618,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.708817720413208,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.0905836969614029,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.718085289001465,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.08345075696706772,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.7235910892486572,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.08599896728992462,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.7178783416748047,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.07301691174507141,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.714029312133789,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.08231015503406525,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.7196826934814453,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.08286412060260773,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.7176082134246826,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.08590883016586304,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.7242164611816406,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.07531397044658661,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.7225024700164795,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.04711190238595009,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.695887327194214,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.05076327919960022,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.7029781341552734,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.04495669901371002,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.7057900428771973,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.04312257468700409,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.697537899017334,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.043580975383520126,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.67315673828125,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.03360886126756668,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.6883413791656494,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.030360296368598938,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.700927257537842,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.0310810673981905,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.6917803287506104,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.02703232690691948,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.6667580604553223,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.02756045199930668,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.6702821254730225,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.02611083723604679,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.669790744781494,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.02615324780344963,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.6672821044921875,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.026000119745731354,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.6671881675720215,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.02882794849574566,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.67586088180542,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03327213600277901,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.6618690490722656,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.032920945435762405,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.665444850921631,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.02812962979078293,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.6680707931518555,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.025886602699756622,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.652898073196411,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.025222789496183395,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.6560399532318115,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.023131893947720528,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.66127347946167,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.02513662725687027,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.6463732719421387,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.027243830263614655,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.6507630348205566,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.025437965989112854,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.6485800743103027,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.02770340070128441,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.657241106033325,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.03594067320227623,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.6281137466430664,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04486451297998428,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.6340909004211426,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.05102410912513733,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.6463394165039062,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04514491558074951,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.636554002761841,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04469453915953636,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.6453680992126465,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.05421523004770279,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.646010398864746,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04779462888836861,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.6315622329711914,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.06697724014520645,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.6416730880737305,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.07520416378974915,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.641632318496704,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.055692676454782486,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.6289877891540527,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.04539895057678223,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.6363251209259033,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.04829065501689911,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.641608476638794,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.041314542293548584,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.6269643306732178,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.04100198298692703,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.6243174076080322,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.04105962812900543,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.626585006713867,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.03172122687101364,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9550899267196655,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6614077091217041,
+ "eval_runtime": 42.9009,
+ "eval_samples_per_second": 56.922,
+ "eval_steps_per_second": 0.466,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.03258531913161278,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.612126588821411,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.040250636637210846,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.6025328636169434,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.034234337508678436,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.6085972785949707,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.03635786101222038,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.6126222610473633,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.039404451847076416,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.6035327911376953,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.038504838943481445,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5985798835754395,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.03744236007332802,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5873115062713623,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.034184787422418594,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.5899314880371094,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.03766429424285889,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.602734327316284,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.03741588443517685,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.592207193374634,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.03187853842973709,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5850372314453125,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.038447409868240356,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.592654228210449,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.04046263173222542,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.5881710052490234,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.03872602432966232,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5881471633911133,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.03960775211453438,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5826945304870605,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.033132150769233704,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.584038734436035,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.029504230245947838,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.584357261657715,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.03665253147482872,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5677168369293213,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.037208545953035355,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.560795783996582,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.038244519382715225,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.591458559036255,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.043307431042194366,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.5642588138580322,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.055119726806879044,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.57377290725708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.053323689848184586,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5767312049865723,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.0502864345908165,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.562190055847168,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.05568842589855194,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.565556049346924,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.061149828135967255,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5883288383483887,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.059933699667453766,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.5710184574127197,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.06041933596134186,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.572603225708008,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.05482695251703262,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.5667271614074707,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.04712968319654465,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.554041624069214,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.04842432215809822,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.5660672187805176,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.04273047298192978,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.554839849472046,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.05356749892234802,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.5521483421325684,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.06062676012516022,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.5551886558532715,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.05431462824344635,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5654542446136475,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.0505850687623024,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5648036003112793,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.04696856811642647,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.5559377670288086,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.042585279792547226,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.5660319328308105,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.04026484116911888,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.551206588745117,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03382421284914017,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.556361198425293,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.031177924945950508,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.5413894653320312,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.028217196464538574,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.5600497722625732,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.027062954381108284,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5447568893432617,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.03001212701201439,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.540614604949951,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.028869854286313057,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.544583797454834,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.02869918756186962,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.5471596717834473,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03359716013073921,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5416982173919678,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.04000182822346687,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.534921407699585,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.043030042201280594,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5368564128875732,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.04329848289489746,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.539764881134033,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.03740967437624931,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.5246665477752686,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.03733935207128525,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.529362916946411,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.039609868079423904,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.542544364929199,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.03937261551618576,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5366649627685547,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.03261927142739296,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.5250535011291504,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.03904294595122337,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.520721435546875,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.03913239762187004,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.5122013092041016,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.03610851988196373,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.5212767124176025,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.03944548964500427,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.537186622619629,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04762452840805054,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.5204620361328125,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04427091404795647,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5286149978637695,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04366641491651535,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.5398361682891846,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.04644155502319336,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.519331932067871,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.04372382536530495,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.531022787094116,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.04404324293136597,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.521758794784546,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.042115990072488785,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.5146074295043945,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.03796013817191124,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.5040054321289062,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.04327977076172829,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.5209312438964844,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04150385782122612,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.515871286392212,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.04465346783399582,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.5247764587402344,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.05690346285700798,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.501911163330078,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04907930642366409,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8968371152877808,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6358620524406433,
+ "eval_runtime": 42.9991,
+ "eval_samples_per_second": 56.792,
+ "eval_steps_per_second": 0.465,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.03755255416035652,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.486576795578003,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.040234558284282684,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.501087188720703,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.041416723281145096,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4942233562469482,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.039655014872550964,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.503453254699707,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.041707683354616165,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.490978717803955,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.04545537009835243,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4757118225097656,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.04328389838337898,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.504516839981079,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.043312299996614456,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.4920215606689453,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.046466976404190063,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.494760513305664,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.041762981563806534,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4875552654266357,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.03713781014084816,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.495652675628662,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.040603458881378174,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4821407794952393,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.037969645112752914,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.484981060028076,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.03652074560523033,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4736409187316895,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.03783264383673668,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.483346700668335,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.03404324874281883,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.48856782913208,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.03294725716114044,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.483706474304199,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.0347001850605011,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.474888324737549,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.0359107106924057,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.478990077972412,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.03738103806972504,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.4709420204162598,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.03722188249230385,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.465726852416992,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.034907322376966476,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.482231855392456,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.035695310682058334,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4785828590393066,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.04009179025888443,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.4795875549316406,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.038996778428554535,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.478696823120117,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04082407429814339,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.465200901031494,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04856136813759804,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4605681896209717,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.05671004578471184,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.47277569770813,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.0481606163084507,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.476151466369629,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.037689510732889175,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.466142177581787,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.05005146935582161,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.4798035621643066,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.04745829105377197,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.46163272857666,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.04638158157467842,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4728071689605713,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.04530330374836922,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.4621877670288086,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.03971488028764725,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.474224090576172,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.04521876201033592,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.46122670173645,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03923020884394646,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.4603679180145264,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03986494988203049,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.4589664936065674,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.04280020296573639,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.4624319076538086,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.03930043429136276,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.47080135345459,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.0314876064658165,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.458728790283203,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.03189660236239433,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4467148780822754,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03264275938272476,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4613699913024902,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03021484799683094,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.4535279273986816,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.030740290880203247,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.4621875286102295,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03096810355782509,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.458500385284424,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03105205111205578,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.461958646774292,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.03330068290233612,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.449777364730835,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.032831475138664246,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.4616191387176514,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.0387776717543602,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.445256471633911,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.041803423315286636,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.4545960426330566,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.04148292914032936,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.433408737182617,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.04648579657077789,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.449134349822998,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.04186409339308739,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.4449565410614014,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.038569554686546326,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.443675994873047,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03908300772309303,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4626526832580566,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.038897737860679626,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.438591957092285,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.0360480472445488,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.436077117919922,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0429830439388752,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4384970664978027,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.04465422406792641,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.458029270172119,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.037096526473760605,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.436816930770874,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.047823466360569,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.44439959526062,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.05929310992360115,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.4599215984344482,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.053512319922447205,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4387903213500977,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.042625587433576584,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4359588623046875,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.056042976677417755,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.4441447257995605,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05813189223408699,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4399144649505615,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.06313478201627731,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.453447103500366,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.05281798169016838,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.4428963661193848,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.03991910442709923,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4359219074249268,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.047981832176446915,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4506607055664062,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.03646523132920265,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8225739002227783,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6196661591529846,
+ "eval_runtime": 43.8106,
+ "eval_samples_per_second": 55.74,
+ "eval_steps_per_second": 0.457,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.03673036769032478,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.4227209091186523,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.044547371566295624,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.4073169231414795,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.04089285805821419,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4279446601867676,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.03538106009364128,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.4234392642974854,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.03152978792786598,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4214606285095215,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.03324202075600624,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.4085447788238525,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.03352442383766174,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.417776107788086,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.03582802042365074,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.411144256591797,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.0351114459335804,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.405872344970703,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.038898173719644547,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4029338359832764,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.03926656022667885,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4100089073181152,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.04285634309053421,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.410360336303711,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.04179581627249718,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4093637466430664,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.03356701880693436,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4118056297302246,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.038140635937452316,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.4154677391052246,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.04468780383467674,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416846513748169,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.062421441078186035,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4157638549804688,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06317313015460968,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.418593406677246,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.049498945474624634,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4204845428466797,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.046322405338287354,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.403489589691162,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.053081002086400986,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.4148740768432617,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05222945287823677,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.416757583618164,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04893574118614197,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4241795539855957,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04951956123113632,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.401386260986328,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04019852355122566,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.4168622493743896,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.03625848889350891,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4026846885681152,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.03464345261454582,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4021215438842773,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.03750818222761154,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4113571643829346,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.0360100194811821,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.3977622985839844,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.03464623540639877,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.413273334503174,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.03193575143814087,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.423659324645996,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.02710454724729061,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.396085262298584,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.02830432914197445,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.420255422592163,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.029717663303017616,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.3958120346069336,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.030005933716893196,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3908543586730957,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.029054716229438782,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.3873138427734375,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.031556472182273865,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3892760276794434,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03518872708082199,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4113478660583496,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.03471249341964722,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.384857654571533,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03223155066370964,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3840765953063965,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.030012644827365875,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.388247489929199,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.03162285313010216,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.3987512588500977,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.03700442612171173,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.3993430137634277,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.03710822016000748,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.387359380722046,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.028411181643605232,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3925249576568604,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031246855854988098,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.38845157623291,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.034664880484342575,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.392436981201172,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.03354701027274132,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.390890121459961,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.03527764603495598,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.3954997062683105,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03325049206614494,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3847177028656006,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.028021875768899918,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.385910987854004,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.02964954823255539,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3787484169006348,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.03068520873785019,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.400784969329834,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03416873514652252,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.391033411026001,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.03534875437617302,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.3796474933624268,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.03732474148273468,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3831934928894043,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.03648678958415985,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3785600662231445,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03801294043660164,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.387587070465088,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.04186025261878967,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.3893089294433594,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.04416101053357124,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3852195739746094,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.04137178510427475,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.391308307647705,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.05010323226451874,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.3906192779541016,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.049587685614824295,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.3846192359924316,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.04125971347093582,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.384507179260254,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.039676666259765625,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.38167142868042,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.038775283843278885,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.3789167404174805,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.03634221479296684,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.383060932159424,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03464275971055031,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.388552665710449,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.031778786331415176,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.3903298377990723,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.027904309332370758,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.377493381500244,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.028964119032025337,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.381959915161133,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.0233962032943964,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7810559272766113,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6070130467414856,
+ "eval_runtime": 50.871,
+ "eval_samples_per_second": 48.004,
+ "eval_steps_per_second": 0.393,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.030209315940737724,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.3569841384887695,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.04112662002444267,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.376680374145508,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.055858466774225235,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.3621644973754883,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.06746986508369446,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3607563972473145,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.06736797094345093,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3548707962036133,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.0562627948820591,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.364654541015625,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.051833927631378174,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.3748884201049805,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.04824110120534897,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.3502087593078613,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.05518793687224388,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3687515258789062,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.06564056873321533,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.367131233215332,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.0605909489095211,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.3589868545532227,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.05744732916355133,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.370117664337158,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.06430432945489883,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.3703484535217285,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.06186915934085846,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3720293045043945,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.05977998673915863,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.3512699604034424,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.06413097679615021,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.377257823944092,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.053935080766677856,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.350717544555664,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.049110863357782364,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.3688902854919434,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.04313219338655472,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3374459743499756,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.039556123316287994,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3482556343078613,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.03796547278761864,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3547937870025635,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.03646104037761688,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.344686269760132,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.03355357423424721,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.3588008880615234,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.03004707582294941,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.352043628692627,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.030186176300048828,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3500709533691406,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.029748762026429176,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.3523058891296387,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.03238696977496147,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3635129928588867,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.032599564641714096,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.346269130706787,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.028626957908272743,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.357151746749878,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.02909751422703266,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3557677268981934,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.029951099306344986,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.34753680229187,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.029768310487270355,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3408985137939453,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.03005973994731903,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.362542152404785,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03360462561249733,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.352529287338257,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.034002047032117844,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.3334460258483887,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.03292316198348999,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3446991443634033,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.0337117575109005,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.350605010986328,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03428212180733681,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.357408046722412,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.03169601410627365,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.349088191986084,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.03389411047101021,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.361576795578003,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.036497049033641815,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.35420560836792,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03574785590171814,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3678526878356934,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.030757993459701538,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3520517349243164,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.0289121363312006,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.337864637374878,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.028012607246637344,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.353999614715576,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.027558492496609688,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.342132568359375,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.02871510572731495,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3425655364990234,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03070886805653572,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3401665687561035,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.03056657873094082,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.345277786254883,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.030907969921827316,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.346494674682617,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0318334735929966,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.3390543460845947,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.029700733721256256,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.3398728370666504,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.029687359929084778,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.3474364280700684,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.027234600856900215,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3457484245300293,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.02823043428361416,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.337700366973877,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03002646192908287,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.335954189300537,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.027246717363595963,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.330207347869873,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.029703574255108833,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.328280448913574,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.028796236962080002,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3448386192321777,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.027944380417466164,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.320234537124634,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.030626650899648666,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.334698438644409,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.029164662584662437,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3450326919555664,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03034045360982418,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.3420588970184326,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.02856968529522419,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3396072387695312,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.02892562374472618,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.340027332305908,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.02938242256641388,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.33550763130188,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.029261089861392975,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3246421813964844,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.030340535566210747,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.3465514183044434,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.033335570245981216,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.3467206954956055,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.03452218323945999,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.343702793121338,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.03326255828142166,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.337631940841675,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025433633476495743,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7402067184448242,
+ "step": 720
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.0776560164274176e+17,
+ "train_batch_size": 32,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-ssl/checkpoint-720/training_args.bin b/runs/l2r90-ssl/checkpoint-720/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5924a9264314215faa31ea62b22ba59c4573b542
--- /dev/null
+++ b/runs/l2r90-ssl/checkpoint-720/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2e206f0318237cba8820486751e683a8c1e9fd84a823416f09cf3e109aba53ea
+size 4792