ardanila commited on
Commit
12b352f
1 Parent(s): 71638c9

Upload 19 files

Browse files
adapter_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model_name_or_path": "sharpbai/Llama-2-13b-hf",
3
+ "bias": "none",
4
+ "enable_lora": null,
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "lora_alpha": 16,
9
+ "lora_dropout": 0.05,
10
+ "merge_weights": false,
11
+ "modules_to_save": null,
12
+ "peft_type": "LORA",
13
+ "r": 8,
14
+ "target_modules": [
15
+ "q_proj",
16
+ "v_proj",
17
+ "k_proj",
18
+ "o_proj"
19
+ ],
20
+ "task_type": "CAUSAL_LM"
21
+ }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29c63d6be77bdc2d95f5a1b4d74a93c2ae25e4763f5d675146ffa31d950f130f
3
+ size 26329549
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "<pad>": 32000
3
+ }
finetune_args.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "micro_batch_size": 2,
3
+ "gradient_accumulation_steps": 2,
4
+ "num_train_epochs": 33,
5
+ "learning_rate": 0.0003,
6
+ "cutoff_len": 2500,
7
+ "val_set_size": 0,
8
+ "lora_r": 8,
9
+ "lora_alpha": 16,
10
+ "lora_dropout": 0.05,
11
+ "lora_target_modules": [
12
+ "q_proj",
13
+ "v_proj",
14
+ "k_proj",
15
+ "o_proj"
16
+ ],
17
+ "lora_modules_to_save": [],
18
+ "train_on_inputs": true,
19
+ "group_by_length": false,
20
+ "load_in_8bit": true,
21
+ "fp16": false,
22
+ "bf16": true,
23
+ "gradient_checkpointing": false,
24
+ "save_steps": 100,
25
+ "save_total_limit": 100,
26
+ "logging_steps": 10,
27
+ "additional_training_arguments": null,
28
+ "additional_lora_config": null,
29
+ "resume_from_checkpoint": "/content/drive/My Drive/Colab Data/LLaMA-LoRA Tuner/lora_models/13b5/checkpoint-2550"
30
+ }
info.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model": "sharpbai/Llama-2-13b-hf",
3
+ "prompt_template": "user_and_ai",
4
+ "dataset_name": "simple.json",
5
+ "dataset_rows": 335,
6
+ "trained_on_machine": "9c8fd9be693a",
7
+ "timestamp": 1695898930.9098132,
8
+ "continued_from_model": "13b5",
9
+ "continued_from_checkpoint": "checkpoint-2550",
10
+ "tuner_version": "d83a6747"
11
+ }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cea48a6f23e84f176da840bbb25cbab27712d268619e61c7b8f7a1905f082d8
3
+ size 52696837
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37eca9cb327161d0664faba4648161c3348b5f9b43dce2494a387655adc83ad3
3
+ size 26329549
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9041dbbb98d1e355e367d325aa060833e77c32c1413391c8f2c58c3c6b8d089d
3
+ size 14575
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4765951f05a8e198a09bb197b1b516d0bf93dbc9707332af6d7544b6ef97360a
3
+ size 627
special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<unk>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "<s>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "clean_up_tokenization_spaces": false,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "</s>",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "legacy": false,
22
+ "model_max_length": 1000000000000000019884624838656,
23
+ "pad_token": null,
24
+ "sp_model_kwargs": {},
25
+ "tokenizer_class": "LlamaTokenizer",
26
+ "unk_token": {
27
+ "__type": "AddedToken",
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false
33
+ }
34
+ }
train_data_samples.json ADDED
The diff for this file is too large to render. See raw diff
 
train_output.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ 2772,
3
+ 0.007915083082422378,
4
+ {
5
+ "train_runtime": 2005.6988,
6
+ "train_samples_per_second": 5.512,
7
+ "train_steps_per_second": 1.382,
8
+ "total_flos": 1.6099719938337178e+18,
9
+ "train_loss": 0.007915083082422378,
10
+ "epoch": 33.0
11
+ }
12
+ ]
trainer_args.json ADDED
@@ -0,0 +1,115 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_dir": "/content/drive/My Drive/Colab Data/LLaMA-LoRA Tuner/lora_models/13b9",
3
+ "overwrite_output_dir": false,
4
+ "do_train": false,
5
+ "do_eval": false,
6
+ "do_predict": false,
7
+ "evaluation_strategy": "no",
8
+ "prediction_loss_only": false,
9
+ "per_device_train_batch_size": 2,
10
+ "per_device_eval_batch_size": 8,
11
+ "per_gpu_train_batch_size": null,
12
+ "per_gpu_eval_batch_size": null,
13
+ "gradient_accumulation_steps": 2,
14
+ "eval_accumulation_steps": null,
15
+ "eval_delay": 0,
16
+ "learning_rate": 0.0003,
17
+ "weight_decay": 0.0,
18
+ "adam_beta1": 0.9,
19
+ "adam_beta2": 0.999,
20
+ "adam_epsilon": 1e-08,
21
+ "max_grad_norm": 1.0,
22
+ "num_train_epochs": 33,
23
+ "max_steps": -1,
24
+ "lr_scheduler_type": "linear",
25
+ "warmup_ratio": 0.0,
26
+ "warmup_steps": 100,
27
+ "log_level": "passive",
28
+ "log_level_replica": "warning",
29
+ "log_on_each_node": true,
30
+ "logging_dir": "/content/drive/My Drive/Colab Data/LLaMA-LoRA Tuner/lora_models/13b9/runs/Sep28_11-02-45_9c8fd9be693a",
31
+ "logging_strategy": "steps",
32
+ "logging_first_step": false,
33
+ "logging_steps": 10,
34
+ "logging_nan_inf_filter": true,
35
+ "save_strategy": "steps",
36
+ "save_steps": 100,
37
+ "save_total_limit": 100,
38
+ "save_safetensors": false,
39
+ "save_on_each_node": false,
40
+ "no_cuda": false,
41
+ "use_mps_device": false,
42
+ "seed": 42,
43
+ "data_seed": null,
44
+ "jit_mode_eval": false,
45
+ "use_ipex": false,
46
+ "bf16": true,
47
+ "fp16": false,
48
+ "fp16_opt_level": "O1",
49
+ "half_precision_backend": "cuda_amp",
50
+ "bf16_full_eval": false,
51
+ "fp16_full_eval": false,
52
+ "tf32": null,
53
+ "local_rank": -1,
54
+ "xpu_backend": null,
55
+ "tpu_num_cores": null,
56
+ "tpu_metrics_debug": false,
57
+ "debug": [],
58
+ "dataloader_drop_last": false,
59
+ "eval_steps": null,
60
+ "dataloader_num_workers": 0,
61
+ "past_index": -1,
62
+ "run_name": "/content/drive/My Drive/Colab Data/LLaMA-LoRA Tuner/lora_models/13b9",
63
+ "disable_tqdm": false,
64
+ "remove_unused_columns": true,
65
+ "label_names": null,
66
+ "load_best_model_at_end": false,
67
+ "metric_for_best_model": null,
68
+ "greater_is_better": null,
69
+ "ignore_data_skip": false,
70
+ "sharded_ddp": [],
71
+ "fsdp": [],
72
+ "fsdp_min_num_params": 0,
73
+ "fsdp_config": {
74
+ "fsdp_min_num_params": 0,
75
+ "xla": false,
76
+ "xla_fsdp_grad_ckpt": false
77
+ },
78
+ "fsdp_transformer_layer_cls_to_wrap": null,
79
+ "deepspeed": null,
80
+ "label_smoothing_factor": 0.0,
81
+ "optim": "adamw_torch",
82
+ "optim_args": null,
83
+ "adafactor": false,
84
+ "group_by_length": false,
85
+ "length_column_name": "length",
86
+ "report_to": [
87
+ "tensorboard"
88
+ ],
89
+ "ddp_find_unused_parameters": null,
90
+ "ddp_bucket_cap_mb": null,
91
+ "dataloader_pin_memory": true,
92
+ "skip_memory_metrics": true,
93
+ "use_legacy_prediction_loop": false,
94
+ "push_to_hub": false,
95
+ "resume_from_checkpoint": null,
96
+ "hub_model_id": null,
97
+ "hub_strategy": "every_save",
98
+ "hub_token": "<HUB_TOKEN>",
99
+ "hub_private_repo": false,
100
+ "gradient_checkpointing": false,
101
+ "include_inputs_for_metrics": false,
102
+ "fp16_backend": "auto",
103
+ "push_to_hub_model_id": null,
104
+ "push_to_hub_organization": null,
105
+ "push_to_hub_token": "<PUSH_TO_HUB_TOKEN>",
106
+ "mp_parameters": "",
107
+ "auto_find_batch_size": false,
108
+ "full_determinism": false,
109
+ "torchdynamo": null,
110
+ "ray_scope": "last",
111
+ "ddp_timeout": 1800,
112
+ "torch_compile": false,
113
+ "torch_compile_backend": null,
114
+ "torch_compile_mode": null
115
+ }
trainer_log_history.jsonl ADDED
@@ -0,0 +1,278 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"epoch": 0.12, "learning_rate": 2.9999999999999997e-05, "loss": 1.3126, "step": 10}
2
+ {"epoch": 0.24, "learning_rate": 5.9999999999999995e-05, "loss": 1.3439, "step": 20}
3
+ {"epoch": 0.36, "learning_rate": 8.999999999999999e-05, "loss": 1.3104, "step": 30}
4
+ {"epoch": 0.48, "learning_rate": 0.00011999999999999999, "loss": 1.2288, "step": 40}
5
+ {"epoch": 0.6, "learning_rate": 0.00015, "loss": 1.1731, "step": 50}
6
+ {"epoch": 0.71, "learning_rate": 0.00017999999999999998, "loss": 1.1577, "step": 60}
7
+ {"epoch": 0.83, "learning_rate": 0.00020999999999999998, "loss": 1.1174, "step": 70}
8
+ {"epoch": 0.95, "learning_rate": 0.00023999999999999998, "loss": 1.0971, "step": 80}
9
+ {"epoch": 1.07, "learning_rate": 0.00027, "loss": 1.0506, "step": 90}
10
+ {"epoch": 1.19, "learning_rate": 0.0003, "loss": 1.0976, "step": 100}
11
+ {"epoch": 1.31, "learning_rate": 0.0002996385542168674, "loss": 1.125, "step": 110}
12
+ {"epoch": 1.43, "learning_rate": 0.00029927710843373495, "loss": 1.0538, "step": 120}
13
+ {"epoch": 1.55, "learning_rate": 0.0002989156626506024, "loss": 1.1258, "step": 130}
14
+ {"epoch": 1.67, "learning_rate": 0.00029855421686746987, "loss": 1.0822, "step": 140}
15
+ {"epoch": 1.79, "learning_rate": 0.00029819277108433735, "loss": 1.1109, "step": 150}
16
+ {"epoch": 1.9, "learning_rate": 0.0002978313253012048, "loss": 1.0948, "step": 160}
17
+ {"epoch": 2.02, "learning_rate": 0.00029746987951807227, "loss": 1.0414, "step": 170}
18
+ {"epoch": 2.14, "learning_rate": 0.00029710843373493976, "loss": 1.0601, "step": 180}
19
+ {"epoch": 2.26, "learning_rate": 0.0002967469879518072, "loss": 1.0577, "step": 190}
20
+ {"epoch": 2.38, "learning_rate": 0.0002963855421686747, "loss": 1.0807, "step": 200}
21
+ {"epoch": 2.5, "learning_rate": 0.00029602409638554216, "loss": 1.1074, "step": 210}
22
+ {"epoch": 2.62, "learning_rate": 0.0002956626506024096, "loss": 1.069, "step": 220}
23
+ {"epoch": 2.74, "learning_rate": 0.0002953012048192771, "loss": 1.0588, "step": 230}
24
+ {"epoch": 2.86, "learning_rate": 0.00029493975903614457, "loss": 1.1084, "step": 240}
25
+ {"epoch": 2.98, "learning_rate": 0.000294578313253012, "loss": 1.0767, "step": 250}
26
+ {"epoch": 3.1, "learning_rate": 0.0002942168674698795, "loss": 1.0462, "step": 260}
27
+ {"epoch": 3.21, "learning_rate": 0.000293855421686747, "loss": 1.0666, "step": 270}
28
+ {"epoch": 3.33, "learning_rate": 0.0002934939759036144, "loss": 1.0284, "step": 280}
29
+ {"epoch": 3.45, "learning_rate": 0.0002931325301204819, "loss": 1.0777, "step": 290}
30
+ {"epoch": 3.57, "learning_rate": 0.0002927710843373494, "loss": 1.0611, "step": 300}
31
+ {"epoch": 3.69, "learning_rate": 0.00029240963855421687, "loss": 1.0645, "step": 310}
32
+ {"epoch": 3.81, "learning_rate": 0.0002920481927710843, "loss": 1.0233, "step": 320}
33
+ {"epoch": 3.93, "learning_rate": 0.0002916867469879518, "loss": 1.0475, "step": 330}
34
+ {"epoch": 4.05, "learning_rate": 0.00029132530120481927, "loss": 1.0703, "step": 340}
35
+ {"epoch": 4.17, "learning_rate": 0.0002909638554216867, "loss": 1.0487, "step": 350}
36
+ {"epoch": 4.29, "learning_rate": 0.0002906024096385542, "loss": 1.0293, "step": 360}
37
+ {"epoch": 4.4, "learning_rate": 0.0002902409638554217, "loss": 1.037, "step": 370}
38
+ {"epoch": 4.52, "learning_rate": 0.0002898795180722891, "loss": 1.0678, "step": 380}
39
+ {"epoch": 4.64, "learning_rate": 0.0002895180722891566, "loss": 1.0603, "step": 390}
40
+ {"epoch": 4.76, "learning_rate": 0.0002891566265060241, "loss": 1.0428, "step": 400}
41
+ {"epoch": 4.88, "learning_rate": 0.0002887951807228915, "loss": 1.035, "step": 410}
42
+ {"epoch": 5.0, "learning_rate": 0.000288433734939759, "loss": 1.0401, "step": 420}
43
+ {"epoch": 5.12, "learning_rate": 0.0002880722891566265, "loss": 1.0475, "step": 430}
44
+ {"epoch": 5.24, "learning_rate": 0.0002877108433734939, "loss": 1.0064, "step": 440}
45
+ {"epoch": 5.36, "learning_rate": 0.0002873493975903614, "loss": 1.0432, "step": 450}
46
+ {"epoch": 5.48, "learning_rate": 0.0002869879518072289, "loss": 1.0251, "step": 460}
47
+ {"epoch": 5.6, "learning_rate": 0.0002866265060240963, "loss": 1.0731, "step": 470}
48
+ {"epoch": 5.71, "learning_rate": 0.0002862650602409638, "loss": 1.0108, "step": 480}
49
+ {"epoch": 5.83, "learning_rate": 0.0002859036144578313, "loss": 1.0032, "step": 490}
50
+ {"epoch": 5.95, "learning_rate": 0.00028554216867469873, "loss": 1.0084, "step": 500}
51
+ {"epoch": 6.07, "learning_rate": 0.00028518072289156627, "loss": 1.011, "step": 510}
52
+ {"epoch": 6.19, "learning_rate": 0.0002848192771084337, "loss": 1.0214, "step": 520}
53
+ {"epoch": 6.31, "learning_rate": 0.0002844578313253012, "loss": 1.0117, "step": 530}
54
+ {"epoch": 6.43, "learning_rate": 0.0002840963855421687, "loss": 1.0065, "step": 540}
55
+ {"epoch": 6.55, "learning_rate": 0.0002837349397590361, "loss": 1.0052, "step": 550}
56
+ {"epoch": 6.67, "learning_rate": 0.0002833734939759036, "loss": 1.0286, "step": 560}
57
+ {"epoch": 6.79, "learning_rate": 0.0002830120481927711, "loss": 0.9932, "step": 570}
58
+ {"epoch": 6.9, "learning_rate": 0.0002826506024096385, "loss": 1.0117, "step": 580}
59
+ {"epoch": 7.02, "learning_rate": 0.000282289156626506, "loss": 0.9999, "step": 590}
60
+ {"epoch": 7.14, "learning_rate": 0.0002819277108433735, "loss": 0.992, "step": 600}
61
+ {"epoch": 7.26, "learning_rate": 0.0002815662650602409, "loss": 1.0053, "step": 610}
62
+ {"epoch": 7.38, "learning_rate": 0.0002812048192771084, "loss": 0.9967, "step": 620}
63
+ {"epoch": 7.5, "learning_rate": 0.0002808433734939759, "loss": 0.9603, "step": 630}
64
+ {"epoch": 7.62, "learning_rate": 0.0002804819277108433, "loss": 0.9778, "step": 640}
65
+ {"epoch": 7.74, "learning_rate": 0.0002801204819277108, "loss": 1.0112, "step": 650}
66
+ {"epoch": 7.86, "learning_rate": 0.0002797590361445783, "loss": 0.9785, "step": 660}
67
+ {"epoch": 7.98, "learning_rate": 0.00027939759036144573, "loss": 0.9976, "step": 670}
68
+ {"epoch": 8.1, "learning_rate": 0.0002790361445783132, "loss": 0.9728, "step": 680}
69
+ {"epoch": 8.21, "learning_rate": 0.0002786746987951807, "loss": 0.9541, "step": 690}
70
+ {"epoch": 8.33, "learning_rate": 0.0002783132530120482, "loss": 0.9576, "step": 700}
71
+ {"epoch": 8.45, "learning_rate": 0.0002779518072289156, "loss": 0.9821, "step": 710}
72
+ {"epoch": 8.57, "learning_rate": 0.0002775903614457831, "loss": 0.9833, "step": 720}
73
+ {"epoch": 8.69, "learning_rate": 0.0002772289156626506, "loss": 0.9733, "step": 730}
74
+ {"epoch": 8.81, "learning_rate": 0.00027686746987951803, "loss": 0.9662, "step": 740}
75
+ {"epoch": 8.93, "learning_rate": 0.0002765060240963855, "loss": 0.9954, "step": 750}
76
+ {"epoch": 9.05, "learning_rate": 0.000276144578313253, "loss": 0.9656, "step": 760}
77
+ {"epoch": 9.17, "learning_rate": 0.00027578313253012044, "loss": 0.9462, "step": 770}
78
+ {"epoch": 9.29, "learning_rate": 0.0002754216867469879, "loss": 0.8893, "step": 780}
79
+ {"epoch": 9.4, "learning_rate": 0.0002750602409638554, "loss": 0.967, "step": 790}
80
+ {"epoch": 9.52, "learning_rate": 0.00027469879518072284, "loss": 0.8942, "step": 800}
81
+ {"epoch": 9.64, "learning_rate": 0.00027433734939759033, "loss": 0.9385, "step": 810}
82
+ {"epoch": 9.76, "learning_rate": 0.0002739759036144578, "loss": 0.9863, "step": 820}
83
+ {"epoch": 9.88, "learning_rate": 0.00027361445783132525, "loss": 0.945, "step": 830}
84
+ {"epoch": 10.0, "learning_rate": 0.00027325301204819273, "loss": 0.9697, "step": 840}
85
+ {"epoch": 10.12, "learning_rate": 0.0002728915662650602, "loss": 0.9053, "step": 850}
86
+ {"epoch": 10.24, "learning_rate": 0.00027253012048192765, "loss": 0.916, "step": 860}
87
+ {"epoch": 10.36, "learning_rate": 0.00027216867469879514, "loss": 0.9316, "step": 870}
88
+ {"epoch": 10.48, "learning_rate": 0.0002718072289156626, "loss": 0.9116, "step": 880}
89
+ {"epoch": 10.6, "learning_rate": 0.00027144578313253006, "loss": 0.8975, "step": 890}
90
+ {"epoch": 10.71, "learning_rate": 0.0002710843373493976, "loss": 0.9206, "step": 900}
91
+ {"epoch": 10.83, "learning_rate": 0.00027072289156626503, "loss": 0.9543, "step": 910}
92
+ {"epoch": 10.95, "learning_rate": 0.0002703614457831325, "loss": 0.8909, "step": 920}
93
+ {"epoch": 11.07, "learning_rate": 0.00027, "loss": 0.8982, "step": 930}
94
+ {"epoch": 11.19, "learning_rate": 0.00026963855421686744, "loss": 0.9051, "step": 940}
95
+ {"epoch": 11.31, "learning_rate": 0.0002692771084337349, "loss": 0.8927, "step": 950}
96
+ {"epoch": 11.43, "learning_rate": 0.0002689156626506024, "loss": 0.8364, "step": 960}
97
+ {"epoch": 11.55, "learning_rate": 0.00026855421686746984, "loss": 0.9083, "step": 970}
98
+ {"epoch": 11.67, "learning_rate": 0.00026819277108433733, "loss": 0.9085, "step": 980}
99
+ {"epoch": 11.79, "learning_rate": 0.0002678313253012048, "loss": 0.8402, "step": 990}
100
+ {"epoch": 11.9, "learning_rate": 0.00026746987951807225, "loss": 0.8964, "step": 1000}
101
+ {"epoch": 12.02, "learning_rate": 0.00026710843373493973, "loss": 0.8608, "step": 1010}
102
+ {"epoch": 12.14, "learning_rate": 0.0002667469879518072, "loss": 0.8442, "step": 1020}
103
+ {"epoch": 12.26, "learning_rate": 0.00026638554216867465, "loss": 0.8686, "step": 1030}
104
+ {"epoch": 12.38, "learning_rate": 0.00026602409638554214, "loss": 0.8222, "step": 1040}
105
+ {"epoch": 12.5, "learning_rate": 0.0002656626506024096, "loss": 0.8634, "step": 1050}
106
+ {"epoch": 12.62, "learning_rate": 0.00026530120481927706, "loss": 0.8106, "step": 1060}
107
+ {"epoch": 12.74, "learning_rate": 0.00026493975903614454, "loss": 0.8377, "step": 1070}
108
+ {"epoch": 12.86, "learning_rate": 0.00026457831325301203, "loss": 0.8477, "step": 1080}
109
+ {"epoch": 12.98, "learning_rate": 0.00026421686746987946, "loss": 0.8738, "step": 1090}
110
+ {"epoch": 13.1, "learning_rate": 0.00026385542168674695, "loss": 0.8378, "step": 1100}
111
+ {"epoch": 13.21, "learning_rate": 0.00026349397590361444, "loss": 0.7851, "step": 1110}
112
+ {"epoch": 13.33, "learning_rate": 0.0002631325301204819, "loss": 0.8583, "step": 1120}
113
+ {"epoch": 13.45, "learning_rate": 0.00026277108433734936, "loss": 0.7784, "step": 1130}
114
+ {"epoch": 13.57, "learning_rate": 0.00026240963855421684, "loss": 0.8323, "step": 1140}
115
+ {"epoch": 13.69, "learning_rate": 0.00026204819277108433, "loss": 0.7988, "step": 1150}
116
+ {"epoch": 13.81, "learning_rate": 0.0002616867469879518, "loss": 0.7671, "step": 1160}
117
+ {"epoch": 13.93, "learning_rate": 0.00026132530120481925, "loss": 0.8336, "step": 1170}
118
+ {"epoch": 14.05, "learning_rate": 0.00026096385542168673, "loss": 0.7438, "step": 1180}
119
+ {"epoch": 14.17, "learning_rate": 0.0002606024096385542, "loss": 0.7234, "step": 1190}
120
+ {"epoch": 14.29, "learning_rate": 0.00026024096385542165, "loss": 0.7257, "step": 1200}
121
+ {"epoch": 14.4, "learning_rate": 0.00025987951807228914, "loss": 0.7833, "step": 1210}
122
+ {"epoch": 14.52, "learning_rate": 0.0002595180722891566, "loss": 0.77, "step": 1220}
123
+ {"epoch": 14.64, "learning_rate": 0.00025915662650602406, "loss": 0.7756, "step": 1230}
124
+ {"epoch": 14.76, "learning_rate": 0.00025879518072289154, "loss": 0.7867, "step": 1240}
125
+ {"epoch": 14.88, "learning_rate": 0.00025843373493975903, "loss": 0.7665, "step": 1250}
126
+ {"epoch": 15.0, "learning_rate": 0.00025807228915662646, "loss": 0.7284, "step": 1260}
127
+ {"epoch": 15.12, "learning_rate": 0.00025771084337349395, "loss": 0.6533, "step": 1270}
128
+ {"epoch": 15.24, "learning_rate": 0.00025734939759036144, "loss": 0.712, "step": 1280}
129
+ {"epoch": 15.36, "learning_rate": 0.0002569879518072289, "loss": 0.7408, "step": 1290}
130
+ {"epoch": 15.48, "learning_rate": 0.00025662650602409636, "loss": 0.6888, "step": 1300}
131
+ {"epoch": 15.6, "learning_rate": 0.00025626506024096384, "loss": 0.7257, "step": 1310}
132
+ {"epoch": 15.71, "learning_rate": 0.00025590361445783133, "loss": 0.687, "step": 1320}
133
+ {"epoch": 15.83, "learning_rate": 0.00025554216867469876, "loss": 0.7182, "step": 1330}
134
+ {"epoch": 15.95, "learning_rate": 0.00025518072289156625, "loss": 0.7337, "step": 1340}
135
+ {"epoch": 16.07, "learning_rate": 0.00025481927710843373, "loss": 0.6581, "step": 1350}
136
+ {"epoch": 16.19, "learning_rate": 0.00025445783132530117, "loss": 0.6839, "step": 1360}
137
+ {"epoch": 16.31, "learning_rate": 0.00025409638554216865, "loss": 0.6336, "step": 1370}
138
+ {"epoch": 16.43, "learning_rate": 0.00025373493975903614, "loss": 0.6679, "step": 1380}
139
+ {"epoch": 16.55, "learning_rate": 0.00025337349397590357, "loss": 0.632, "step": 1390}
140
+ {"epoch": 16.67, "learning_rate": 0.00025301204819277106, "loss": 0.6286, "step": 1400}
141
+ {"epoch": 16.79, "learning_rate": 0.00025265060240963855, "loss": 0.7263, "step": 1410}
142
+ {"epoch": 16.9, "learning_rate": 0.000252289156626506, "loss": 0.6465, "step": 1420}
143
+ {"epoch": 17.02, "learning_rate": 0.00025192771084337346, "loss": 0.617, "step": 1430}
144
+ {"epoch": 17.14, "learning_rate": 0.00025156626506024095, "loss": 0.599, "step": 1440}
145
+ {"epoch": 17.26, "learning_rate": 0.0002512048192771084, "loss": 0.6203, "step": 1450}
146
+ {"epoch": 17.38, "learning_rate": 0.00025084337349397587, "loss": 0.5545, "step": 1460}
147
+ {"epoch": 17.5, "learning_rate": 0.00025048192771084336, "loss": 0.6239, "step": 1470}
148
+ {"epoch": 17.62, "learning_rate": 0.0002501204819277108, "loss": 0.5739, "step": 1480}
149
+ {"epoch": 17.74, "learning_rate": 0.00024975903614457833, "loss": 0.6289, "step": 1490}
150
+ {"epoch": 17.86, "learning_rate": 0.00024939759036144576, "loss": 0.5728, "step": 1500}
151
+ {"epoch": 17.98, "learning_rate": 0.00024903614457831325, "loss": 0.6469, "step": 1510}
152
+ {"epoch": 18.1, "learning_rate": 0.00024867469879518074, "loss": 0.5318, "step": 1520}
153
+ {"epoch": 18.21, "learning_rate": 0.00024831325301204817, "loss": 0.4846, "step": 1530}
154
+ {"epoch": 18.33, "learning_rate": 0.00024795180722891565, "loss": 0.5352, "step": 1540}
155
+ {"epoch": 18.45, "learning_rate": 0.00024759036144578314, "loss": 0.5341, "step": 1550}
156
+ {"epoch": 18.57, "learning_rate": 0.0002472289156626506, "loss": 0.57, "step": 1560}
157
+ {"epoch": 18.69, "learning_rate": 0.00024686746987951806, "loss": 0.5228, "step": 1570}
158
+ {"epoch": 18.81, "learning_rate": 0.00024650602409638555, "loss": 0.5857, "step": 1580}
159
+ {"epoch": 18.93, "learning_rate": 0.000246144578313253, "loss": 0.5219, "step": 1590}
160
+ {"epoch": 19.05, "learning_rate": 0.00024578313253012046, "loss": 0.5233, "step": 1600}
161
+ {"epoch": 19.17, "learning_rate": 0.00024542168674698795, "loss": 0.4589, "step": 1610}
162
+ {"epoch": 19.29, "learning_rate": 0.0002450602409638554, "loss": 0.5354, "step": 1620}
163
+ {"epoch": 19.4, "learning_rate": 0.00024469879518072287, "loss": 0.4662, "step": 1630}
164
+ {"epoch": 19.52, "learning_rate": 0.00024433734939759036, "loss": 0.465, "step": 1640}
165
+ {"epoch": 19.64, "learning_rate": 0.00024397590361445782, "loss": 0.4602, "step": 1650}
166
+ {"epoch": 19.76, "learning_rate": 0.00024361445783132528, "loss": 0.4851, "step": 1660}
167
+ {"epoch": 19.88, "learning_rate": 0.00024325301204819276, "loss": 0.495, "step": 1670}
168
+ {"epoch": 20.0, "learning_rate": 0.00024289156626506022, "loss": 0.4556, "step": 1680}
169
+ {"epoch": 20.12, "learning_rate": 0.00024253012048192768, "loss": 0.3657, "step": 1690}
170
+ {"epoch": 20.24, "learning_rate": 0.00024216867469879517, "loss": 0.4285, "step": 1700}
171
+ {"epoch": 20.36, "learning_rate": 0.00024180722891566263, "loss": 0.3793, "step": 1710}
172
+ {"epoch": 20.48, "learning_rate": 0.0002414457831325301, "loss": 0.4846, "step": 1720}
173
+ {"epoch": 20.6, "learning_rate": 0.00024108433734939757, "loss": 0.4543, "step": 1730}
174
+ {"epoch": 20.71, "learning_rate": 0.00024072289156626503, "loss": 0.4694, "step": 1740}
175
+ {"epoch": 20.83, "learning_rate": 0.0002403614457831325, "loss": 0.3779, "step": 1750}
176
+ {"epoch": 20.95, "learning_rate": 0.00023999999999999998, "loss": 0.431, "step": 1760}
177
+ {"epoch": 21.07, "learning_rate": 0.00023963855421686744, "loss": 0.367, "step": 1770}
178
+ {"epoch": 21.19, "learning_rate": 0.0002392771084337349, "loss": 0.4387, "step": 1780}
179
+ {"epoch": 21.31, "learning_rate": 0.0002389156626506024, "loss": 0.3326, "step": 1790}
180
+ {"epoch": 21.43, "learning_rate": 0.00023855421686746987, "loss": 0.3899, "step": 1800}
181
+ {"epoch": 21.55, "learning_rate": 0.00023819277108433733, "loss": 0.3869, "step": 1810}
182
+ {"epoch": 21.67, "learning_rate": 0.00023783132530120482, "loss": 0.3685, "step": 1820}
183
+ {"epoch": 21.79, "learning_rate": 0.00023746987951807228, "loss": 0.3711, "step": 1830}
184
+ {"epoch": 21.9, "learning_rate": 0.00023710843373493974, "loss": 0.3748, "step": 1840}
185
+ {"epoch": 22.02, "learning_rate": 0.00023674698795180722, "loss": 0.3345, "step": 1850}
186
+ {"epoch": 22.14, "learning_rate": 0.00023638554216867468, "loss": 0.3128, "step": 1860}
187
+ {"epoch": 22.26, "learning_rate": 0.00023602409638554214, "loss": 0.348, "step": 1870}
188
+ {"epoch": 22.38, "learning_rate": 0.00023566265060240963, "loss": 0.3506, "step": 1880}
189
+ {"epoch": 22.5, "learning_rate": 0.0002353012048192771, "loss": 0.289, "step": 1890}
190
+ {"epoch": 22.62, "learning_rate": 0.00023493975903614455, "loss": 0.3008, "step": 1900}
191
+ {"epoch": 22.74, "learning_rate": 0.00023457831325301203, "loss": 0.3372, "step": 1910}
192
+ {"epoch": 22.86, "learning_rate": 0.0002342168674698795, "loss": 0.3642, "step": 1920}
193
+ {"epoch": 22.98, "learning_rate": 0.00023385542168674695, "loss": 0.3117, "step": 1930}
194
+ {"epoch": 23.1, "learning_rate": 0.00023349397590361444, "loss": 0.3003, "step": 1940}
195
+ {"epoch": 23.21, "learning_rate": 0.0002331325301204819, "loss": 0.2342, "step": 1950}
196
+ {"epoch": 23.33, "learning_rate": 0.00023277108433734936, "loss": 0.3119, "step": 1960}
197
+ {"epoch": 23.45, "learning_rate": 0.00023240963855421687, "loss": 0.2618, "step": 1970}
198
+ {"epoch": 23.57, "learning_rate": 0.00023204819277108433, "loss": 0.3138, "step": 1980}
199
+ {"epoch": 23.69, "learning_rate": 0.00023168674698795176, "loss": 0.2934, "step": 1990}
200
+ {"epoch": 23.81, "learning_rate": 0.00023132530120481928, "loss": 0.2713, "step": 2000}
201
+ {"epoch": 23.93, "learning_rate": 0.00023096385542168674, "loss": 0.3095, "step": 2010}
202
+ {"epoch": 24.05, "learning_rate": 0.0002306024096385542, "loss": 0.2688, "step": 2020}
203
+ {"epoch": 24.17, "learning_rate": 0.00023024096385542168, "loss": 0.2309, "step": 2030}
204
+ {"epoch": 24.29, "learning_rate": 0.00022987951807228914, "loss": 0.2614, "step": 2040}
205
+ {"epoch": 24.4, "learning_rate": 0.0002295180722891566, "loss": 0.2264, "step": 2050}
206
+ {"epoch": 24.52, "learning_rate": 0.0002291566265060241, "loss": 0.2506, "step": 2060}
207
+ {"epoch": 24.64, "learning_rate": 0.00022879518072289155, "loss": 0.2241, "step": 2070}
208
+ {"epoch": 24.76, "learning_rate": 0.000228433734939759, "loss": 0.2882, "step": 2080}
209
+ {"epoch": 24.88, "learning_rate": 0.0002280722891566265, "loss": 0.2868, "step": 2090}
210
+ {"epoch": 25.0, "learning_rate": 0.00022771084337349395, "loss": 0.284, "step": 2100}
211
+ {"epoch": 25.12, "learning_rate": 0.0002273493975903614, "loss": 0.2088, "step": 2110}
212
+ {"epoch": 25.24, "learning_rate": 0.0002269879518072289, "loss": 0.1954, "step": 2120}
213
+ {"epoch": 25.36, "learning_rate": 0.00022662650602409636, "loss": 0.2121, "step": 2130}
214
+ {"epoch": 25.48, "learning_rate": 0.00022626506024096382, "loss": 0.2345, "step": 2140}
215
+ {"epoch": 25.6, "learning_rate": 0.0002259036144578313, "loss": 0.221, "step": 2150}
216
+ {"epoch": 25.71, "learning_rate": 0.00022554216867469876, "loss": 0.2258, "step": 2160}
217
+ {"epoch": 25.83, "learning_rate": 0.00022518072289156622, "loss": 0.2467, "step": 2170}
218
+ {"epoch": 25.95, "learning_rate": 0.00022481927710843374, "loss": 0.2216, "step": 2180}
219
+ {"epoch": 26.07, "learning_rate": 0.0002244578313253012, "loss": 0.2106, "step": 2190}
220
+ {"epoch": 26.19, "learning_rate": 0.00022409638554216866, "loss": 0.2242, "step": 2200}
221
+ {"epoch": 26.31, "learning_rate": 0.00022373493975903614, "loss": 0.1702, "step": 2210}
222
+ {"epoch": 26.43, "learning_rate": 0.0002233734939759036, "loss": 0.2117, "step": 2220}
223
+ {"epoch": 26.55, "learning_rate": 0.00022301204819277106, "loss": 0.1841, "step": 2230}
224
+ {"epoch": 26.67, "learning_rate": 0.00022265060240963855, "loss": 0.1786, "step": 2240}
225
+ {"epoch": 26.79, "learning_rate": 0.000222289156626506, "loss": 0.201, "step": 2250}
226
+ {"epoch": 26.9, "learning_rate": 0.00022192771084337347, "loss": 0.1915, "step": 2260}
227
+ {"epoch": 27.02, "learning_rate": 0.00022156626506024095, "loss": 0.1973, "step": 2270}
228
+ {"epoch": 27.14, "learning_rate": 0.0002212048192771084, "loss": 0.1649, "step": 2280}
229
+ {"epoch": 27.26, "learning_rate": 0.00022084337349397587, "loss": 0.1665, "step": 2290}
230
+ {"epoch": 27.38, "learning_rate": 0.00022048192771084336, "loss": 0.2129, "step": 2300}
231
+ {"epoch": 27.5, "learning_rate": 0.00022012048192771082, "loss": 0.194, "step": 2310}
232
+ {"epoch": 27.62, "learning_rate": 0.00021975903614457828, "loss": 0.1684, "step": 2320}
233
+ {"epoch": 27.74, "learning_rate": 0.00021939759036144576, "loss": 0.1312, "step": 2330}
234
+ {"epoch": 27.86, "learning_rate": 0.00021903614457831322, "loss": 0.1711, "step": 2340}
235
+ {"epoch": 27.98, "learning_rate": 0.00021867469879518068, "loss": 0.1863, "step": 2350}
236
+ {"epoch": 28.1, "learning_rate": 0.0002183132530120482, "loss": 0.1425, "step": 2360}
237
+ {"epoch": 28.21, "learning_rate": 0.00021795180722891563, "loss": 0.1293, "step": 2370}
238
+ {"epoch": 28.33, "learning_rate": 0.0002175903614457831, "loss": 0.1406, "step": 2380}
239
+ {"epoch": 28.45, "learning_rate": 0.0002172289156626506, "loss": 0.1583, "step": 2390}
240
+ {"epoch": 28.57, "learning_rate": 0.00021686746987951806, "loss": 0.1706, "step": 2400}
241
+ {"epoch": 28.69, "learning_rate": 0.00021650602409638552, "loss": 0.1629, "step": 2410}
242
+ {"epoch": 28.81, "learning_rate": 0.000216144578313253, "loss": 0.1788, "step": 2420}
243
+ {"epoch": 28.93, "learning_rate": 0.00021578313253012047, "loss": 0.1618, "step": 2430}
244
+ {"epoch": 29.05, "learning_rate": 0.00021542168674698793, "loss": 0.1391, "step": 2440}
245
+ {"epoch": 29.17, "learning_rate": 0.00021506024096385541, "loss": 0.1271, "step": 2450}
246
+ {"epoch": 29.29, "learning_rate": 0.00021469879518072287, "loss": 0.1312, "step": 2460}
247
+ {"epoch": 29.4, "learning_rate": 0.00021433734939759033, "loss": 0.1401, "step": 2470}
248
+ {"epoch": 29.52, "learning_rate": 0.00021397590361445782, "loss": 0.1222, "step": 2480}
249
+ {"epoch": 29.64, "learning_rate": 0.00021361445783132528, "loss": 0.1124, "step": 2490}
250
+ {"epoch": 29.76, "learning_rate": 0.00021325301204819274, "loss": 0.1796, "step": 2500}
251
+ {"epoch": 29.88, "learning_rate": 0.00021289156626506022, "loss": 0.1502, "step": 2510}
252
+ {"epoch": 30.0, "learning_rate": 0.00021253012048192768, "loss": 0.137, "step": 2520}
253
+ {"epoch": 30.12, "learning_rate": 0.00021216867469879514, "loss": 0.1021, "step": 2530}
254
+ {"epoch": 30.24, "learning_rate": 0.00021180722891566263, "loss": 0.129, "step": 2540}
255
+ {"epoch": 30.36, "learning_rate": 0.0002114457831325301, "loss": 0.1119, "step": 2550}
256
+ {"loss": 0.1135, "learning_rate": 2.3802395209580838e-05, "epoch": 30.48, "step": 2560}
257
+ {"loss": 0.1224, "learning_rate": 2.2679640718562872e-05, "epoch": 30.6, "step": 2570}
258
+ {"loss": 0.1004, "learning_rate": 2.1556886227544907e-05, "epoch": 30.71, "step": 2580}
259
+ {"loss": 0.1048, "learning_rate": 2.043413173652694e-05, "epoch": 30.83, "step": 2590}
260
+ {"loss": 0.0978, "learning_rate": 1.931137724550898e-05, "epoch": 30.95, "step": 2600}
261
+ {"loss": 0.0981, "learning_rate": 1.8188622754491017e-05, "epoch": 31.07, "step": 2610}
262
+ {"loss": 0.0979, "learning_rate": 1.7065868263473052e-05, "epoch": 31.19, "step": 2620}
263
+ {"loss": 0.103, "learning_rate": 1.594311377245509e-05, "epoch": 31.31, "step": 2630}
264
+ {"loss": 0.1065, "learning_rate": 1.4820359281437124e-05, "epoch": 31.43, "step": 2640}
265
+ {"loss": 0.0912, "learning_rate": 1.369760479041916e-05, "epoch": 31.55, "step": 2650}
266
+ {"loss": 0.0927, "learning_rate": 1.2574850299401195e-05, "epoch": 31.67, "step": 2660}
267
+ {"loss": 0.0942, "learning_rate": 1.1452095808383233e-05, "epoch": 31.79, "step": 2670}
268
+ {"loss": 0.0896, "learning_rate": 1.032934131736527e-05, "epoch": 31.9, "step": 2680}
269
+ {"loss": 0.0965, "learning_rate": 9.206586826347304e-06, "epoch": 32.02, "step": 2690}
270
+ {"loss": 0.0943, "learning_rate": 8.08383233532934e-06, "epoch": 32.14, "step": 2700}
271
+ {"loss": 0.0866, "learning_rate": 6.961077844311377e-06, "epoch": 32.26, "step": 2710}
272
+ {"loss": 0.1012, "learning_rate": 5.838323353293413e-06, "epoch": 32.38, "step": 2720}
273
+ {"loss": 0.1023, "learning_rate": 4.7155688622754485e-06, "epoch": 32.5, "step": 2730}
274
+ {"loss": 0.0856, "learning_rate": 3.5928143712574848e-06, "epoch": 32.62, "step": 2740}
275
+ {"loss": 0.1008, "learning_rate": 2.470059880239521e-06, "epoch": 32.74, "step": 2750}
276
+ {"loss": 0.0932, "learning_rate": 1.3473053892215567e-06, "epoch": 32.86, "step": 2760}
277
+ {"loss": 0.1046, "learning_rate": 2.245508982035928e-07, "epoch": 32.98, "step": 2770}
278
+ {"train_runtime": 2005.6988, "train_samples_per_second": 5.512, "train_steps_per_second": 1.382, "total_flos": 1.6099719938337178e+18, "train_loss": 0.007915083082422378, "epoch": 33.0, "step": 2772}
trainer_state.json ADDED
@@ -0,0 +1,1636 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 32.142857142857146,
5
+ "global_step": 2700,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.12,
12
+ "learning_rate": 2.9999999999999997e-05,
13
+ "loss": 1.3126,
14
+ "step": 10
15
+ },
16
+ {
17
+ "epoch": 0.24,
18
+ "learning_rate": 5.9999999999999995e-05,
19
+ "loss": 1.3439,
20
+ "step": 20
21
+ },
22
+ {
23
+ "epoch": 0.36,
24
+ "learning_rate": 8.999999999999999e-05,
25
+ "loss": 1.3104,
26
+ "step": 30
27
+ },
28
+ {
29
+ "epoch": 0.48,
30
+ "learning_rate": 0.00011999999999999999,
31
+ "loss": 1.2288,
32
+ "step": 40
33
+ },
34
+ {
35
+ "epoch": 0.6,
36
+ "learning_rate": 0.00015,
37
+ "loss": 1.1731,
38
+ "step": 50
39
+ },
40
+ {
41
+ "epoch": 0.71,
42
+ "learning_rate": 0.00017999999999999998,
43
+ "loss": 1.1577,
44
+ "step": 60
45
+ },
46
+ {
47
+ "epoch": 0.83,
48
+ "learning_rate": 0.00020999999999999998,
49
+ "loss": 1.1174,
50
+ "step": 70
51
+ },
52
+ {
53
+ "epoch": 0.95,
54
+ "learning_rate": 0.00023999999999999998,
55
+ "loss": 1.0971,
56
+ "step": 80
57
+ },
58
+ {
59
+ "epoch": 1.07,
60
+ "learning_rate": 0.00027,
61
+ "loss": 1.0506,
62
+ "step": 90
63
+ },
64
+ {
65
+ "epoch": 1.19,
66
+ "learning_rate": 0.0003,
67
+ "loss": 1.0976,
68
+ "step": 100
69
+ },
70
+ {
71
+ "epoch": 1.31,
72
+ "learning_rate": 0.0002996385542168674,
73
+ "loss": 1.125,
74
+ "step": 110
75
+ },
76
+ {
77
+ "epoch": 1.43,
78
+ "learning_rate": 0.00029927710843373495,
79
+ "loss": 1.0538,
80
+ "step": 120
81
+ },
82
+ {
83
+ "epoch": 1.55,
84
+ "learning_rate": 0.0002989156626506024,
85
+ "loss": 1.1258,
86
+ "step": 130
87
+ },
88
+ {
89
+ "epoch": 1.67,
90
+ "learning_rate": 0.00029855421686746987,
91
+ "loss": 1.0822,
92
+ "step": 140
93
+ },
94
+ {
95
+ "epoch": 1.79,
96
+ "learning_rate": 0.00029819277108433735,
97
+ "loss": 1.1109,
98
+ "step": 150
99
+ },
100
+ {
101
+ "epoch": 1.9,
102
+ "learning_rate": 0.0002978313253012048,
103
+ "loss": 1.0948,
104
+ "step": 160
105
+ },
106
+ {
107
+ "epoch": 2.02,
108
+ "learning_rate": 0.00029746987951807227,
109
+ "loss": 1.0414,
110
+ "step": 170
111
+ },
112
+ {
113
+ "epoch": 2.14,
114
+ "learning_rate": 0.00029710843373493976,
115
+ "loss": 1.0601,
116
+ "step": 180
117
+ },
118
+ {
119
+ "epoch": 2.26,
120
+ "learning_rate": 0.0002967469879518072,
121
+ "loss": 1.0577,
122
+ "step": 190
123
+ },
124
+ {
125
+ "epoch": 2.38,
126
+ "learning_rate": 0.0002963855421686747,
127
+ "loss": 1.0807,
128
+ "step": 200
129
+ },
130
+ {
131
+ "epoch": 2.5,
132
+ "learning_rate": 0.00029602409638554216,
133
+ "loss": 1.1074,
134
+ "step": 210
135
+ },
136
+ {
137
+ "epoch": 2.62,
138
+ "learning_rate": 0.0002956626506024096,
139
+ "loss": 1.069,
140
+ "step": 220
141
+ },
142
+ {
143
+ "epoch": 2.74,
144
+ "learning_rate": 0.0002953012048192771,
145
+ "loss": 1.0588,
146
+ "step": 230
147
+ },
148
+ {
149
+ "epoch": 2.86,
150
+ "learning_rate": 0.00029493975903614457,
151
+ "loss": 1.1084,
152
+ "step": 240
153
+ },
154
+ {
155
+ "epoch": 2.98,
156
+ "learning_rate": 0.000294578313253012,
157
+ "loss": 1.0767,
158
+ "step": 250
159
+ },
160
+ {
161
+ "epoch": 3.1,
162
+ "learning_rate": 0.0002942168674698795,
163
+ "loss": 1.0462,
164
+ "step": 260
165
+ },
166
+ {
167
+ "epoch": 3.21,
168
+ "learning_rate": 0.000293855421686747,
169
+ "loss": 1.0666,
170
+ "step": 270
171
+ },
172
+ {
173
+ "epoch": 3.33,
174
+ "learning_rate": 0.0002934939759036144,
175
+ "loss": 1.0284,
176
+ "step": 280
177
+ },
178
+ {
179
+ "epoch": 3.45,
180
+ "learning_rate": 0.0002931325301204819,
181
+ "loss": 1.0777,
182
+ "step": 290
183
+ },
184
+ {
185
+ "epoch": 3.57,
186
+ "learning_rate": 0.0002927710843373494,
187
+ "loss": 1.0611,
188
+ "step": 300
189
+ },
190
+ {
191
+ "epoch": 3.69,
192
+ "learning_rate": 0.00029240963855421687,
193
+ "loss": 1.0645,
194
+ "step": 310
195
+ },
196
+ {
197
+ "epoch": 3.81,
198
+ "learning_rate": 0.0002920481927710843,
199
+ "loss": 1.0233,
200
+ "step": 320
201
+ },
202
+ {
203
+ "epoch": 3.93,
204
+ "learning_rate": 0.0002916867469879518,
205
+ "loss": 1.0475,
206
+ "step": 330
207
+ },
208
+ {
209
+ "epoch": 4.05,
210
+ "learning_rate": 0.00029132530120481927,
211
+ "loss": 1.0703,
212
+ "step": 340
213
+ },
214
+ {
215
+ "epoch": 4.17,
216
+ "learning_rate": 0.0002909638554216867,
217
+ "loss": 1.0487,
218
+ "step": 350
219
+ },
220
+ {
221
+ "epoch": 4.29,
222
+ "learning_rate": 0.0002906024096385542,
223
+ "loss": 1.0293,
224
+ "step": 360
225
+ },
226
+ {
227
+ "epoch": 4.4,
228
+ "learning_rate": 0.0002902409638554217,
229
+ "loss": 1.037,
230
+ "step": 370
231
+ },
232
+ {
233
+ "epoch": 4.52,
234
+ "learning_rate": 0.0002898795180722891,
235
+ "loss": 1.0678,
236
+ "step": 380
237
+ },
238
+ {
239
+ "epoch": 4.64,
240
+ "learning_rate": 0.0002895180722891566,
241
+ "loss": 1.0603,
242
+ "step": 390
243
+ },
244
+ {
245
+ "epoch": 4.76,
246
+ "learning_rate": 0.0002891566265060241,
247
+ "loss": 1.0428,
248
+ "step": 400
249
+ },
250
+ {
251
+ "epoch": 4.88,
252
+ "learning_rate": 0.0002887951807228915,
253
+ "loss": 1.035,
254
+ "step": 410
255
+ },
256
+ {
257
+ "epoch": 5.0,
258
+ "learning_rate": 0.000288433734939759,
259
+ "loss": 1.0401,
260
+ "step": 420
261
+ },
262
+ {
263
+ "epoch": 5.12,
264
+ "learning_rate": 0.0002880722891566265,
265
+ "loss": 1.0475,
266
+ "step": 430
267
+ },
268
+ {
269
+ "epoch": 5.24,
270
+ "learning_rate": 0.0002877108433734939,
271
+ "loss": 1.0064,
272
+ "step": 440
273
+ },
274
+ {
275
+ "epoch": 5.36,
276
+ "learning_rate": 0.0002873493975903614,
277
+ "loss": 1.0432,
278
+ "step": 450
279
+ },
280
+ {
281
+ "epoch": 5.48,
282
+ "learning_rate": 0.0002869879518072289,
283
+ "loss": 1.0251,
284
+ "step": 460
285
+ },
286
+ {
287
+ "epoch": 5.6,
288
+ "learning_rate": 0.0002866265060240963,
289
+ "loss": 1.0731,
290
+ "step": 470
291
+ },
292
+ {
293
+ "epoch": 5.71,
294
+ "learning_rate": 0.0002862650602409638,
295
+ "loss": 1.0108,
296
+ "step": 480
297
+ },
298
+ {
299
+ "epoch": 5.83,
300
+ "learning_rate": 0.0002859036144578313,
301
+ "loss": 1.0032,
302
+ "step": 490
303
+ },
304
+ {
305
+ "epoch": 5.95,
306
+ "learning_rate": 0.00028554216867469873,
307
+ "loss": 1.0084,
308
+ "step": 500
309
+ },
310
+ {
311
+ "epoch": 6.07,
312
+ "learning_rate": 0.00028518072289156627,
313
+ "loss": 1.011,
314
+ "step": 510
315
+ },
316
+ {
317
+ "epoch": 6.19,
318
+ "learning_rate": 0.0002848192771084337,
319
+ "loss": 1.0214,
320
+ "step": 520
321
+ },
322
+ {
323
+ "epoch": 6.31,
324
+ "learning_rate": 0.0002844578313253012,
325
+ "loss": 1.0117,
326
+ "step": 530
327
+ },
328
+ {
329
+ "epoch": 6.43,
330
+ "learning_rate": 0.0002840963855421687,
331
+ "loss": 1.0065,
332
+ "step": 540
333
+ },
334
+ {
335
+ "epoch": 6.55,
336
+ "learning_rate": 0.0002837349397590361,
337
+ "loss": 1.0052,
338
+ "step": 550
339
+ },
340
+ {
341
+ "epoch": 6.67,
342
+ "learning_rate": 0.0002833734939759036,
343
+ "loss": 1.0286,
344
+ "step": 560
345
+ },
346
+ {
347
+ "epoch": 6.79,
348
+ "learning_rate": 0.0002830120481927711,
349
+ "loss": 0.9932,
350
+ "step": 570
351
+ },
352
+ {
353
+ "epoch": 6.9,
354
+ "learning_rate": 0.0002826506024096385,
355
+ "loss": 1.0117,
356
+ "step": 580
357
+ },
358
+ {
359
+ "epoch": 7.02,
360
+ "learning_rate": 0.000282289156626506,
361
+ "loss": 0.9999,
362
+ "step": 590
363
+ },
364
+ {
365
+ "epoch": 7.14,
366
+ "learning_rate": 0.0002819277108433735,
367
+ "loss": 0.992,
368
+ "step": 600
369
+ },
370
+ {
371
+ "epoch": 7.26,
372
+ "learning_rate": 0.0002815662650602409,
373
+ "loss": 1.0053,
374
+ "step": 610
375
+ },
376
+ {
377
+ "epoch": 7.38,
378
+ "learning_rate": 0.0002812048192771084,
379
+ "loss": 0.9967,
380
+ "step": 620
381
+ },
382
+ {
383
+ "epoch": 7.5,
384
+ "learning_rate": 0.0002808433734939759,
385
+ "loss": 0.9603,
386
+ "step": 630
387
+ },
388
+ {
389
+ "epoch": 7.62,
390
+ "learning_rate": 0.0002804819277108433,
391
+ "loss": 0.9778,
392
+ "step": 640
393
+ },
394
+ {
395
+ "epoch": 7.74,
396
+ "learning_rate": 0.0002801204819277108,
397
+ "loss": 1.0112,
398
+ "step": 650
399
+ },
400
+ {
401
+ "epoch": 7.86,
402
+ "learning_rate": 0.0002797590361445783,
403
+ "loss": 0.9785,
404
+ "step": 660
405
+ },
406
+ {
407
+ "epoch": 7.98,
408
+ "learning_rate": 0.00027939759036144573,
409
+ "loss": 0.9976,
410
+ "step": 670
411
+ },
412
+ {
413
+ "epoch": 8.1,
414
+ "learning_rate": 0.0002790361445783132,
415
+ "loss": 0.9728,
416
+ "step": 680
417
+ },
418
+ {
419
+ "epoch": 8.21,
420
+ "learning_rate": 0.0002786746987951807,
421
+ "loss": 0.9541,
422
+ "step": 690
423
+ },
424
+ {
425
+ "epoch": 8.33,
426
+ "learning_rate": 0.0002783132530120482,
427
+ "loss": 0.9576,
428
+ "step": 700
429
+ },
430
+ {
431
+ "epoch": 8.45,
432
+ "learning_rate": 0.0002779518072289156,
433
+ "loss": 0.9821,
434
+ "step": 710
435
+ },
436
+ {
437
+ "epoch": 8.57,
438
+ "learning_rate": 0.0002775903614457831,
439
+ "loss": 0.9833,
440
+ "step": 720
441
+ },
442
+ {
443
+ "epoch": 8.69,
444
+ "learning_rate": 0.0002772289156626506,
445
+ "loss": 0.9733,
446
+ "step": 730
447
+ },
448
+ {
449
+ "epoch": 8.81,
450
+ "learning_rate": 0.00027686746987951803,
451
+ "loss": 0.9662,
452
+ "step": 740
453
+ },
454
+ {
455
+ "epoch": 8.93,
456
+ "learning_rate": 0.0002765060240963855,
457
+ "loss": 0.9954,
458
+ "step": 750
459
+ },
460
+ {
461
+ "epoch": 9.05,
462
+ "learning_rate": 0.000276144578313253,
463
+ "loss": 0.9656,
464
+ "step": 760
465
+ },
466
+ {
467
+ "epoch": 9.17,
468
+ "learning_rate": 0.00027578313253012044,
469
+ "loss": 0.9462,
470
+ "step": 770
471
+ },
472
+ {
473
+ "epoch": 9.29,
474
+ "learning_rate": 0.0002754216867469879,
475
+ "loss": 0.8893,
476
+ "step": 780
477
+ },
478
+ {
479
+ "epoch": 9.4,
480
+ "learning_rate": 0.0002750602409638554,
481
+ "loss": 0.967,
482
+ "step": 790
483
+ },
484
+ {
485
+ "epoch": 9.52,
486
+ "learning_rate": 0.00027469879518072284,
487
+ "loss": 0.8942,
488
+ "step": 800
489
+ },
490
+ {
491
+ "epoch": 9.64,
492
+ "learning_rate": 0.00027433734939759033,
493
+ "loss": 0.9385,
494
+ "step": 810
495
+ },
496
+ {
497
+ "epoch": 9.76,
498
+ "learning_rate": 0.0002739759036144578,
499
+ "loss": 0.9863,
500
+ "step": 820
501
+ },
502
+ {
503
+ "epoch": 9.88,
504
+ "learning_rate": 0.00027361445783132525,
505
+ "loss": 0.945,
506
+ "step": 830
507
+ },
508
+ {
509
+ "epoch": 10.0,
510
+ "learning_rate": 0.00027325301204819273,
511
+ "loss": 0.9697,
512
+ "step": 840
513
+ },
514
+ {
515
+ "epoch": 10.12,
516
+ "learning_rate": 0.0002728915662650602,
517
+ "loss": 0.9053,
518
+ "step": 850
519
+ },
520
+ {
521
+ "epoch": 10.24,
522
+ "learning_rate": 0.00027253012048192765,
523
+ "loss": 0.916,
524
+ "step": 860
525
+ },
526
+ {
527
+ "epoch": 10.36,
528
+ "learning_rate": 0.00027216867469879514,
529
+ "loss": 0.9316,
530
+ "step": 870
531
+ },
532
+ {
533
+ "epoch": 10.48,
534
+ "learning_rate": 0.0002718072289156626,
535
+ "loss": 0.9116,
536
+ "step": 880
537
+ },
538
+ {
539
+ "epoch": 10.6,
540
+ "learning_rate": 0.00027144578313253006,
541
+ "loss": 0.8975,
542
+ "step": 890
543
+ },
544
+ {
545
+ "epoch": 10.71,
546
+ "learning_rate": 0.0002710843373493976,
547
+ "loss": 0.9206,
548
+ "step": 900
549
+ },
550
+ {
551
+ "epoch": 10.83,
552
+ "learning_rate": 0.00027072289156626503,
553
+ "loss": 0.9543,
554
+ "step": 910
555
+ },
556
+ {
557
+ "epoch": 10.95,
558
+ "learning_rate": 0.0002703614457831325,
559
+ "loss": 0.8909,
560
+ "step": 920
561
+ },
562
+ {
563
+ "epoch": 11.07,
564
+ "learning_rate": 0.00027,
565
+ "loss": 0.8982,
566
+ "step": 930
567
+ },
568
+ {
569
+ "epoch": 11.19,
570
+ "learning_rate": 0.00026963855421686744,
571
+ "loss": 0.9051,
572
+ "step": 940
573
+ },
574
+ {
575
+ "epoch": 11.31,
576
+ "learning_rate": 0.0002692771084337349,
577
+ "loss": 0.8927,
578
+ "step": 950
579
+ },
580
+ {
581
+ "epoch": 11.43,
582
+ "learning_rate": 0.0002689156626506024,
583
+ "loss": 0.8364,
584
+ "step": 960
585
+ },
586
+ {
587
+ "epoch": 11.55,
588
+ "learning_rate": 0.00026855421686746984,
589
+ "loss": 0.9083,
590
+ "step": 970
591
+ },
592
+ {
593
+ "epoch": 11.67,
594
+ "learning_rate": 0.00026819277108433733,
595
+ "loss": 0.9085,
596
+ "step": 980
597
+ },
598
+ {
599
+ "epoch": 11.79,
600
+ "learning_rate": 0.0002678313253012048,
601
+ "loss": 0.8402,
602
+ "step": 990
603
+ },
604
+ {
605
+ "epoch": 11.9,
606
+ "learning_rate": 0.00026746987951807225,
607
+ "loss": 0.8964,
608
+ "step": 1000
609
+ },
610
+ {
611
+ "epoch": 12.02,
612
+ "learning_rate": 0.00026710843373493973,
613
+ "loss": 0.8608,
614
+ "step": 1010
615
+ },
616
+ {
617
+ "epoch": 12.14,
618
+ "learning_rate": 0.0002667469879518072,
619
+ "loss": 0.8442,
620
+ "step": 1020
621
+ },
622
+ {
623
+ "epoch": 12.26,
624
+ "learning_rate": 0.00026638554216867465,
625
+ "loss": 0.8686,
626
+ "step": 1030
627
+ },
628
+ {
629
+ "epoch": 12.38,
630
+ "learning_rate": 0.00026602409638554214,
631
+ "loss": 0.8222,
632
+ "step": 1040
633
+ },
634
+ {
635
+ "epoch": 12.5,
636
+ "learning_rate": 0.0002656626506024096,
637
+ "loss": 0.8634,
638
+ "step": 1050
639
+ },
640
+ {
641
+ "epoch": 12.62,
642
+ "learning_rate": 0.00026530120481927706,
643
+ "loss": 0.8106,
644
+ "step": 1060
645
+ },
646
+ {
647
+ "epoch": 12.74,
648
+ "learning_rate": 0.00026493975903614454,
649
+ "loss": 0.8377,
650
+ "step": 1070
651
+ },
652
+ {
653
+ "epoch": 12.86,
654
+ "learning_rate": 0.00026457831325301203,
655
+ "loss": 0.8477,
656
+ "step": 1080
657
+ },
658
+ {
659
+ "epoch": 12.98,
660
+ "learning_rate": 0.00026421686746987946,
661
+ "loss": 0.8738,
662
+ "step": 1090
663
+ },
664
+ {
665
+ "epoch": 13.1,
666
+ "learning_rate": 0.00026385542168674695,
667
+ "loss": 0.8378,
668
+ "step": 1100
669
+ },
670
+ {
671
+ "epoch": 13.21,
672
+ "learning_rate": 0.00026349397590361444,
673
+ "loss": 0.7851,
674
+ "step": 1110
675
+ },
676
+ {
677
+ "epoch": 13.33,
678
+ "learning_rate": 0.0002631325301204819,
679
+ "loss": 0.8583,
680
+ "step": 1120
681
+ },
682
+ {
683
+ "epoch": 13.45,
684
+ "learning_rate": 0.00026277108433734936,
685
+ "loss": 0.7784,
686
+ "step": 1130
687
+ },
688
+ {
689
+ "epoch": 13.57,
690
+ "learning_rate": 0.00026240963855421684,
691
+ "loss": 0.8323,
692
+ "step": 1140
693
+ },
694
+ {
695
+ "epoch": 13.69,
696
+ "learning_rate": 0.00026204819277108433,
697
+ "loss": 0.7988,
698
+ "step": 1150
699
+ },
700
+ {
701
+ "epoch": 13.81,
702
+ "learning_rate": 0.0002616867469879518,
703
+ "loss": 0.7671,
704
+ "step": 1160
705
+ },
706
+ {
707
+ "epoch": 13.93,
708
+ "learning_rate": 0.00026132530120481925,
709
+ "loss": 0.8336,
710
+ "step": 1170
711
+ },
712
+ {
713
+ "epoch": 14.05,
714
+ "learning_rate": 0.00026096385542168673,
715
+ "loss": 0.7438,
716
+ "step": 1180
717
+ },
718
+ {
719
+ "epoch": 14.17,
720
+ "learning_rate": 0.0002606024096385542,
721
+ "loss": 0.7234,
722
+ "step": 1190
723
+ },
724
+ {
725
+ "epoch": 14.29,
726
+ "learning_rate": 0.00026024096385542165,
727
+ "loss": 0.7257,
728
+ "step": 1200
729
+ },
730
+ {
731
+ "epoch": 14.4,
732
+ "learning_rate": 0.00025987951807228914,
733
+ "loss": 0.7833,
734
+ "step": 1210
735
+ },
736
+ {
737
+ "epoch": 14.52,
738
+ "learning_rate": 0.0002595180722891566,
739
+ "loss": 0.77,
740
+ "step": 1220
741
+ },
742
+ {
743
+ "epoch": 14.64,
744
+ "learning_rate": 0.00025915662650602406,
745
+ "loss": 0.7756,
746
+ "step": 1230
747
+ },
748
+ {
749
+ "epoch": 14.76,
750
+ "learning_rate": 0.00025879518072289154,
751
+ "loss": 0.7867,
752
+ "step": 1240
753
+ },
754
+ {
755
+ "epoch": 14.88,
756
+ "learning_rate": 0.00025843373493975903,
757
+ "loss": 0.7665,
758
+ "step": 1250
759
+ },
760
+ {
761
+ "epoch": 15.0,
762
+ "learning_rate": 0.00025807228915662646,
763
+ "loss": 0.7284,
764
+ "step": 1260
765
+ },
766
+ {
767
+ "epoch": 15.12,
768
+ "learning_rate": 0.00025771084337349395,
769
+ "loss": 0.6533,
770
+ "step": 1270
771
+ },
772
+ {
773
+ "epoch": 15.24,
774
+ "learning_rate": 0.00025734939759036144,
775
+ "loss": 0.712,
776
+ "step": 1280
777
+ },
778
+ {
779
+ "epoch": 15.36,
780
+ "learning_rate": 0.0002569879518072289,
781
+ "loss": 0.7408,
782
+ "step": 1290
783
+ },
784
+ {
785
+ "epoch": 15.48,
786
+ "learning_rate": 0.00025662650602409636,
787
+ "loss": 0.6888,
788
+ "step": 1300
789
+ },
790
+ {
791
+ "epoch": 15.6,
792
+ "learning_rate": 0.00025626506024096384,
793
+ "loss": 0.7257,
794
+ "step": 1310
795
+ },
796
+ {
797
+ "epoch": 15.71,
798
+ "learning_rate": 0.00025590361445783133,
799
+ "loss": 0.687,
800
+ "step": 1320
801
+ },
802
+ {
803
+ "epoch": 15.83,
804
+ "learning_rate": 0.00025554216867469876,
805
+ "loss": 0.7182,
806
+ "step": 1330
807
+ },
808
+ {
809
+ "epoch": 15.95,
810
+ "learning_rate": 0.00025518072289156625,
811
+ "loss": 0.7337,
812
+ "step": 1340
813
+ },
814
+ {
815
+ "epoch": 16.07,
816
+ "learning_rate": 0.00025481927710843373,
817
+ "loss": 0.6581,
818
+ "step": 1350
819
+ },
820
+ {
821
+ "epoch": 16.19,
822
+ "learning_rate": 0.00025445783132530117,
823
+ "loss": 0.6839,
824
+ "step": 1360
825
+ },
826
+ {
827
+ "epoch": 16.31,
828
+ "learning_rate": 0.00025409638554216865,
829
+ "loss": 0.6336,
830
+ "step": 1370
831
+ },
832
+ {
833
+ "epoch": 16.43,
834
+ "learning_rate": 0.00025373493975903614,
835
+ "loss": 0.6679,
836
+ "step": 1380
837
+ },
838
+ {
839
+ "epoch": 16.55,
840
+ "learning_rate": 0.00025337349397590357,
841
+ "loss": 0.632,
842
+ "step": 1390
843
+ },
844
+ {
845
+ "epoch": 16.67,
846
+ "learning_rate": 0.00025301204819277106,
847
+ "loss": 0.6286,
848
+ "step": 1400
849
+ },
850
+ {
851
+ "epoch": 16.79,
852
+ "learning_rate": 0.00025265060240963855,
853
+ "loss": 0.7263,
854
+ "step": 1410
855
+ },
856
+ {
857
+ "epoch": 16.9,
858
+ "learning_rate": 0.000252289156626506,
859
+ "loss": 0.6465,
860
+ "step": 1420
861
+ },
862
+ {
863
+ "epoch": 17.02,
864
+ "learning_rate": 0.00025192771084337346,
865
+ "loss": 0.617,
866
+ "step": 1430
867
+ },
868
+ {
869
+ "epoch": 17.14,
870
+ "learning_rate": 0.00025156626506024095,
871
+ "loss": 0.599,
872
+ "step": 1440
873
+ },
874
+ {
875
+ "epoch": 17.26,
876
+ "learning_rate": 0.0002512048192771084,
877
+ "loss": 0.6203,
878
+ "step": 1450
879
+ },
880
+ {
881
+ "epoch": 17.38,
882
+ "learning_rate": 0.00025084337349397587,
883
+ "loss": 0.5545,
884
+ "step": 1460
885
+ },
886
+ {
887
+ "epoch": 17.5,
888
+ "learning_rate": 0.00025048192771084336,
889
+ "loss": 0.6239,
890
+ "step": 1470
891
+ },
892
+ {
893
+ "epoch": 17.62,
894
+ "learning_rate": 0.0002501204819277108,
895
+ "loss": 0.5739,
896
+ "step": 1480
897
+ },
898
+ {
899
+ "epoch": 17.74,
900
+ "learning_rate": 0.00024975903614457833,
901
+ "loss": 0.6289,
902
+ "step": 1490
903
+ },
904
+ {
905
+ "epoch": 17.86,
906
+ "learning_rate": 0.00024939759036144576,
907
+ "loss": 0.5728,
908
+ "step": 1500
909
+ },
910
+ {
911
+ "epoch": 17.98,
912
+ "learning_rate": 0.00024903614457831325,
913
+ "loss": 0.6469,
914
+ "step": 1510
915
+ },
916
+ {
917
+ "epoch": 18.1,
918
+ "learning_rate": 0.00024867469879518074,
919
+ "loss": 0.5318,
920
+ "step": 1520
921
+ },
922
+ {
923
+ "epoch": 18.21,
924
+ "learning_rate": 0.00024831325301204817,
925
+ "loss": 0.4846,
926
+ "step": 1530
927
+ },
928
+ {
929
+ "epoch": 18.33,
930
+ "learning_rate": 0.00024795180722891565,
931
+ "loss": 0.5352,
932
+ "step": 1540
933
+ },
934
+ {
935
+ "epoch": 18.45,
936
+ "learning_rate": 0.00024759036144578314,
937
+ "loss": 0.5341,
938
+ "step": 1550
939
+ },
940
+ {
941
+ "epoch": 18.57,
942
+ "learning_rate": 0.0002472289156626506,
943
+ "loss": 0.57,
944
+ "step": 1560
945
+ },
946
+ {
947
+ "epoch": 18.69,
948
+ "learning_rate": 0.00024686746987951806,
949
+ "loss": 0.5228,
950
+ "step": 1570
951
+ },
952
+ {
953
+ "epoch": 18.81,
954
+ "learning_rate": 0.00024650602409638555,
955
+ "loss": 0.5857,
956
+ "step": 1580
957
+ },
958
+ {
959
+ "epoch": 18.93,
960
+ "learning_rate": 0.000246144578313253,
961
+ "loss": 0.5219,
962
+ "step": 1590
963
+ },
964
+ {
965
+ "epoch": 19.05,
966
+ "learning_rate": 0.00024578313253012046,
967
+ "loss": 0.5233,
968
+ "step": 1600
969
+ },
970
+ {
971
+ "epoch": 19.17,
972
+ "learning_rate": 0.00024542168674698795,
973
+ "loss": 0.4589,
974
+ "step": 1610
975
+ },
976
+ {
977
+ "epoch": 19.29,
978
+ "learning_rate": 0.0002450602409638554,
979
+ "loss": 0.5354,
980
+ "step": 1620
981
+ },
982
+ {
983
+ "epoch": 19.4,
984
+ "learning_rate": 0.00024469879518072287,
985
+ "loss": 0.4662,
986
+ "step": 1630
987
+ },
988
+ {
989
+ "epoch": 19.52,
990
+ "learning_rate": 0.00024433734939759036,
991
+ "loss": 0.465,
992
+ "step": 1640
993
+ },
994
+ {
995
+ "epoch": 19.64,
996
+ "learning_rate": 0.00024397590361445782,
997
+ "loss": 0.4602,
998
+ "step": 1650
999
+ },
1000
+ {
1001
+ "epoch": 19.76,
1002
+ "learning_rate": 0.00024361445783132528,
1003
+ "loss": 0.4851,
1004
+ "step": 1660
1005
+ },
1006
+ {
1007
+ "epoch": 19.88,
1008
+ "learning_rate": 0.00024325301204819276,
1009
+ "loss": 0.495,
1010
+ "step": 1670
1011
+ },
1012
+ {
1013
+ "epoch": 20.0,
1014
+ "learning_rate": 0.00024289156626506022,
1015
+ "loss": 0.4556,
1016
+ "step": 1680
1017
+ },
1018
+ {
1019
+ "epoch": 20.12,
1020
+ "learning_rate": 0.00024253012048192768,
1021
+ "loss": 0.3657,
1022
+ "step": 1690
1023
+ },
1024
+ {
1025
+ "epoch": 20.24,
1026
+ "learning_rate": 0.00024216867469879517,
1027
+ "loss": 0.4285,
1028
+ "step": 1700
1029
+ },
1030
+ {
1031
+ "epoch": 20.36,
1032
+ "learning_rate": 0.00024180722891566263,
1033
+ "loss": 0.3793,
1034
+ "step": 1710
1035
+ },
1036
+ {
1037
+ "epoch": 20.48,
1038
+ "learning_rate": 0.0002414457831325301,
1039
+ "loss": 0.4846,
1040
+ "step": 1720
1041
+ },
1042
+ {
1043
+ "epoch": 20.6,
1044
+ "learning_rate": 0.00024108433734939757,
1045
+ "loss": 0.4543,
1046
+ "step": 1730
1047
+ },
1048
+ {
1049
+ "epoch": 20.71,
1050
+ "learning_rate": 0.00024072289156626503,
1051
+ "loss": 0.4694,
1052
+ "step": 1740
1053
+ },
1054
+ {
1055
+ "epoch": 20.83,
1056
+ "learning_rate": 0.0002403614457831325,
1057
+ "loss": 0.3779,
1058
+ "step": 1750
1059
+ },
1060
+ {
1061
+ "epoch": 20.95,
1062
+ "learning_rate": 0.00023999999999999998,
1063
+ "loss": 0.431,
1064
+ "step": 1760
1065
+ },
1066
+ {
1067
+ "epoch": 21.07,
1068
+ "learning_rate": 0.00023963855421686744,
1069
+ "loss": 0.367,
1070
+ "step": 1770
1071
+ },
1072
+ {
1073
+ "epoch": 21.19,
1074
+ "learning_rate": 0.0002392771084337349,
1075
+ "loss": 0.4387,
1076
+ "step": 1780
1077
+ },
1078
+ {
1079
+ "epoch": 21.31,
1080
+ "learning_rate": 0.0002389156626506024,
1081
+ "loss": 0.3326,
1082
+ "step": 1790
1083
+ },
1084
+ {
1085
+ "epoch": 21.43,
1086
+ "learning_rate": 0.00023855421686746987,
1087
+ "loss": 0.3899,
1088
+ "step": 1800
1089
+ },
1090
+ {
1091
+ "epoch": 21.55,
1092
+ "learning_rate": 0.00023819277108433733,
1093
+ "loss": 0.3869,
1094
+ "step": 1810
1095
+ },
1096
+ {
1097
+ "epoch": 21.67,
1098
+ "learning_rate": 0.00023783132530120482,
1099
+ "loss": 0.3685,
1100
+ "step": 1820
1101
+ },
1102
+ {
1103
+ "epoch": 21.79,
1104
+ "learning_rate": 0.00023746987951807228,
1105
+ "loss": 0.3711,
1106
+ "step": 1830
1107
+ },
1108
+ {
1109
+ "epoch": 21.9,
1110
+ "learning_rate": 0.00023710843373493974,
1111
+ "loss": 0.3748,
1112
+ "step": 1840
1113
+ },
1114
+ {
1115
+ "epoch": 22.02,
1116
+ "learning_rate": 0.00023674698795180722,
1117
+ "loss": 0.3345,
1118
+ "step": 1850
1119
+ },
1120
+ {
1121
+ "epoch": 22.14,
1122
+ "learning_rate": 0.00023638554216867468,
1123
+ "loss": 0.3128,
1124
+ "step": 1860
1125
+ },
1126
+ {
1127
+ "epoch": 22.26,
1128
+ "learning_rate": 0.00023602409638554214,
1129
+ "loss": 0.348,
1130
+ "step": 1870
1131
+ },
1132
+ {
1133
+ "epoch": 22.38,
1134
+ "learning_rate": 0.00023566265060240963,
1135
+ "loss": 0.3506,
1136
+ "step": 1880
1137
+ },
1138
+ {
1139
+ "epoch": 22.5,
1140
+ "learning_rate": 0.0002353012048192771,
1141
+ "loss": 0.289,
1142
+ "step": 1890
1143
+ },
1144
+ {
1145
+ "epoch": 22.62,
1146
+ "learning_rate": 0.00023493975903614455,
1147
+ "loss": 0.3008,
1148
+ "step": 1900
1149
+ },
1150
+ {
1151
+ "epoch": 22.74,
1152
+ "learning_rate": 0.00023457831325301203,
1153
+ "loss": 0.3372,
1154
+ "step": 1910
1155
+ },
1156
+ {
1157
+ "epoch": 22.86,
1158
+ "learning_rate": 0.0002342168674698795,
1159
+ "loss": 0.3642,
1160
+ "step": 1920
1161
+ },
1162
+ {
1163
+ "epoch": 22.98,
1164
+ "learning_rate": 0.00023385542168674695,
1165
+ "loss": 0.3117,
1166
+ "step": 1930
1167
+ },
1168
+ {
1169
+ "epoch": 23.1,
1170
+ "learning_rate": 0.00023349397590361444,
1171
+ "loss": 0.3003,
1172
+ "step": 1940
1173
+ },
1174
+ {
1175
+ "epoch": 23.21,
1176
+ "learning_rate": 0.0002331325301204819,
1177
+ "loss": 0.2342,
1178
+ "step": 1950
1179
+ },
1180
+ {
1181
+ "epoch": 23.33,
1182
+ "learning_rate": 0.00023277108433734936,
1183
+ "loss": 0.3119,
1184
+ "step": 1960
1185
+ },
1186
+ {
1187
+ "epoch": 23.45,
1188
+ "learning_rate": 0.00023240963855421687,
1189
+ "loss": 0.2618,
1190
+ "step": 1970
1191
+ },
1192
+ {
1193
+ "epoch": 23.57,
1194
+ "learning_rate": 0.00023204819277108433,
1195
+ "loss": 0.3138,
1196
+ "step": 1980
1197
+ },
1198
+ {
1199
+ "epoch": 23.69,
1200
+ "learning_rate": 0.00023168674698795176,
1201
+ "loss": 0.2934,
1202
+ "step": 1990
1203
+ },
1204
+ {
1205
+ "epoch": 23.81,
1206
+ "learning_rate": 0.00023132530120481928,
1207
+ "loss": 0.2713,
1208
+ "step": 2000
1209
+ },
1210
+ {
1211
+ "epoch": 23.93,
1212
+ "learning_rate": 0.00023096385542168674,
1213
+ "loss": 0.3095,
1214
+ "step": 2010
1215
+ },
1216
+ {
1217
+ "epoch": 24.05,
1218
+ "learning_rate": 0.0002306024096385542,
1219
+ "loss": 0.2688,
1220
+ "step": 2020
1221
+ },
1222
+ {
1223
+ "epoch": 24.17,
1224
+ "learning_rate": 0.00023024096385542168,
1225
+ "loss": 0.2309,
1226
+ "step": 2030
1227
+ },
1228
+ {
1229
+ "epoch": 24.29,
1230
+ "learning_rate": 0.00022987951807228914,
1231
+ "loss": 0.2614,
1232
+ "step": 2040
1233
+ },
1234
+ {
1235
+ "epoch": 24.4,
1236
+ "learning_rate": 0.0002295180722891566,
1237
+ "loss": 0.2264,
1238
+ "step": 2050
1239
+ },
1240
+ {
1241
+ "epoch": 24.52,
1242
+ "learning_rate": 0.0002291566265060241,
1243
+ "loss": 0.2506,
1244
+ "step": 2060
1245
+ },
1246
+ {
1247
+ "epoch": 24.64,
1248
+ "learning_rate": 0.00022879518072289155,
1249
+ "loss": 0.2241,
1250
+ "step": 2070
1251
+ },
1252
+ {
1253
+ "epoch": 24.76,
1254
+ "learning_rate": 0.000228433734939759,
1255
+ "loss": 0.2882,
1256
+ "step": 2080
1257
+ },
1258
+ {
1259
+ "epoch": 24.88,
1260
+ "learning_rate": 0.0002280722891566265,
1261
+ "loss": 0.2868,
1262
+ "step": 2090
1263
+ },
1264
+ {
1265
+ "epoch": 25.0,
1266
+ "learning_rate": 0.00022771084337349395,
1267
+ "loss": 0.284,
1268
+ "step": 2100
1269
+ },
1270
+ {
1271
+ "epoch": 25.12,
1272
+ "learning_rate": 0.0002273493975903614,
1273
+ "loss": 0.2088,
1274
+ "step": 2110
1275
+ },
1276
+ {
1277
+ "epoch": 25.24,
1278
+ "learning_rate": 0.0002269879518072289,
1279
+ "loss": 0.1954,
1280
+ "step": 2120
1281
+ },
1282
+ {
1283
+ "epoch": 25.36,
1284
+ "learning_rate": 0.00022662650602409636,
1285
+ "loss": 0.2121,
1286
+ "step": 2130
1287
+ },
1288
+ {
1289
+ "epoch": 25.48,
1290
+ "learning_rate": 0.00022626506024096382,
1291
+ "loss": 0.2345,
1292
+ "step": 2140
1293
+ },
1294
+ {
1295
+ "epoch": 25.6,
1296
+ "learning_rate": 0.0002259036144578313,
1297
+ "loss": 0.221,
1298
+ "step": 2150
1299
+ },
1300
+ {
1301
+ "epoch": 25.71,
1302
+ "learning_rate": 0.00022554216867469876,
1303
+ "loss": 0.2258,
1304
+ "step": 2160
1305
+ },
1306
+ {
1307
+ "epoch": 25.83,
1308
+ "learning_rate": 0.00022518072289156622,
1309
+ "loss": 0.2467,
1310
+ "step": 2170
1311
+ },
1312
+ {
1313
+ "epoch": 25.95,
1314
+ "learning_rate": 0.00022481927710843374,
1315
+ "loss": 0.2216,
1316
+ "step": 2180
1317
+ },
1318
+ {
1319
+ "epoch": 26.07,
1320
+ "learning_rate": 0.0002244578313253012,
1321
+ "loss": 0.2106,
1322
+ "step": 2190
1323
+ },
1324
+ {
1325
+ "epoch": 26.19,
1326
+ "learning_rate": 0.00022409638554216866,
1327
+ "loss": 0.2242,
1328
+ "step": 2200
1329
+ },
1330
+ {
1331
+ "epoch": 26.31,
1332
+ "learning_rate": 0.00022373493975903614,
1333
+ "loss": 0.1702,
1334
+ "step": 2210
1335
+ },
1336
+ {
1337
+ "epoch": 26.43,
1338
+ "learning_rate": 0.0002233734939759036,
1339
+ "loss": 0.2117,
1340
+ "step": 2220
1341
+ },
1342
+ {
1343
+ "epoch": 26.55,
1344
+ "learning_rate": 0.00022301204819277106,
1345
+ "loss": 0.1841,
1346
+ "step": 2230
1347
+ },
1348
+ {
1349
+ "epoch": 26.67,
1350
+ "learning_rate": 0.00022265060240963855,
1351
+ "loss": 0.1786,
1352
+ "step": 2240
1353
+ },
1354
+ {
1355
+ "epoch": 26.79,
1356
+ "learning_rate": 0.000222289156626506,
1357
+ "loss": 0.201,
1358
+ "step": 2250
1359
+ },
1360
+ {
1361
+ "epoch": 26.9,
1362
+ "learning_rate": 0.00022192771084337347,
1363
+ "loss": 0.1915,
1364
+ "step": 2260
1365
+ },
1366
+ {
1367
+ "epoch": 27.02,
1368
+ "learning_rate": 0.00022156626506024095,
1369
+ "loss": 0.1973,
1370
+ "step": 2270
1371
+ },
1372
+ {
1373
+ "epoch": 27.14,
1374
+ "learning_rate": 0.0002212048192771084,
1375
+ "loss": 0.1649,
1376
+ "step": 2280
1377
+ },
1378
+ {
1379
+ "epoch": 27.26,
1380
+ "learning_rate": 0.00022084337349397587,
1381
+ "loss": 0.1665,
1382
+ "step": 2290
1383
+ },
1384
+ {
1385
+ "epoch": 27.38,
1386
+ "learning_rate": 0.00022048192771084336,
1387
+ "loss": 0.2129,
1388
+ "step": 2300
1389
+ },
1390
+ {
1391
+ "epoch": 27.5,
1392
+ "learning_rate": 0.00022012048192771082,
1393
+ "loss": 0.194,
1394
+ "step": 2310
1395
+ },
1396
+ {
1397
+ "epoch": 27.62,
1398
+ "learning_rate": 0.00021975903614457828,
1399
+ "loss": 0.1684,
1400
+ "step": 2320
1401
+ },
1402
+ {
1403
+ "epoch": 27.74,
1404
+ "learning_rate": 0.00021939759036144576,
1405
+ "loss": 0.1312,
1406
+ "step": 2330
1407
+ },
1408
+ {
1409
+ "epoch": 27.86,
1410
+ "learning_rate": 0.00021903614457831322,
1411
+ "loss": 0.1711,
1412
+ "step": 2340
1413
+ },
1414
+ {
1415
+ "epoch": 27.98,
1416
+ "learning_rate": 0.00021867469879518068,
1417
+ "loss": 0.1863,
1418
+ "step": 2350
1419
+ },
1420
+ {
1421
+ "epoch": 28.1,
1422
+ "learning_rate": 0.0002183132530120482,
1423
+ "loss": 0.1425,
1424
+ "step": 2360
1425
+ },
1426
+ {
1427
+ "epoch": 28.21,
1428
+ "learning_rate": 0.00021795180722891563,
1429
+ "loss": 0.1293,
1430
+ "step": 2370
1431
+ },
1432
+ {
1433
+ "epoch": 28.33,
1434
+ "learning_rate": 0.0002175903614457831,
1435
+ "loss": 0.1406,
1436
+ "step": 2380
1437
+ },
1438
+ {
1439
+ "epoch": 28.45,
1440
+ "learning_rate": 0.0002172289156626506,
1441
+ "loss": 0.1583,
1442
+ "step": 2390
1443
+ },
1444
+ {
1445
+ "epoch": 28.57,
1446
+ "learning_rate": 0.00021686746987951806,
1447
+ "loss": 0.1706,
1448
+ "step": 2400
1449
+ },
1450
+ {
1451
+ "epoch": 28.69,
1452
+ "learning_rate": 0.00021650602409638552,
1453
+ "loss": 0.1629,
1454
+ "step": 2410
1455
+ },
1456
+ {
1457
+ "epoch": 28.81,
1458
+ "learning_rate": 0.000216144578313253,
1459
+ "loss": 0.1788,
1460
+ "step": 2420
1461
+ },
1462
+ {
1463
+ "epoch": 28.93,
1464
+ "learning_rate": 0.00021578313253012047,
1465
+ "loss": 0.1618,
1466
+ "step": 2430
1467
+ },
1468
+ {
1469
+ "epoch": 29.05,
1470
+ "learning_rate": 0.00021542168674698793,
1471
+ "loss": 0.1391,
1472
+ "step": 2440
1473
+ },
1474
+ {
1475
+ "epoch": 29.17,
1476
+ "learning_rate": 0.00021506024096385541,
1477
+ "loss": 0.1271,
1478
+ "step": 2450
1479
+ },
1480
+ {
1481
+ "epoch": 29.29,
1482
+ "learning_rate": 0.00021469879518072287,
1483
+ "loss": 0.1312,
1484
+ "step": 2460
1485
+ },
1486
+ {
1487
+ "epoch": 29.4,
1488
+ "learning_rate": 0.00021433734939759033,
1489
+ "loss": 0.1401,
1490
+ "step": 2470
1491
+ },
1492
+ {
1493
+ "epoch": 29.52,
1494
+ "learning_rate": 0.00021397590361445782,
1495
+ "loss": 0.1222,
1496
+ "step": 2480
1497
+ },
1498
+ {
1499
+ "epoch": 29.64,
1500
+ "learning_rate": 0.00021361445783132528,
1501
+ "loss": 0.1124,
1502
+ "step": 2490
1503
+ },
1504
+ {
1505
+ "epoch": 29.76,
1506
+ "learning_rate": 0.00021325301204819274,
1507
+ "loss": 0.1796,
1508
+ "step": 2500
1509
+ },
1510
+ {
1511
+ "epoch": 29.88,
1512
+ "learning_rate": 0.00021289156626506022,
1513
+ "loss": 0.1502,
1514
+ "step": 2510
1515
+ },
1516
+ {
1517
+ "epoch": 30.0,
1518
+ "learning_rate": 0.00021253012048192768,
1519
+ "loss": 0.137,
1520
+ "step": 2520
1521
+ },
1522
+ {
1523
+ "epoch": 30.12,
1524
+ "learning_rate": 0.00021216867469879514,
1525
+ "loss": 0.1021,
1526
+ "step": 2530
1527
+ },
1528
+ {
1529
+ "epoch": 30.24,
1530
+ "learning_rate": 0.00021180722891566263,
1531
+ "loss": 0.129,
1532
+ "step": 2540
1533
+ },
1534
+ {
1535
+ "epoch": 30.36,
1536
+ "learning_rate": 0.0002114457831325301,
1537
+ "loss": 0.1119,
1538
+ "step": 2550
1539
+ },
1540
+ {
1541
+ "epoch": 30.48,
1542
+ "learning_rate": 2.3802395209580838e-05,
1543
+ "loss": 0.1135,
1544
+ "step": 2560
1545
+ },
1546
+ {
1547
+ "epoch": 30.6,
1548
+ "learning_rate": 2.2679640718562872e-05,
1549
+ "loss": 0.1224,
1550
+ "step": 2570
1551
+ },
1552
+ {
1553
+ "epoch": 30.71,
1554
+ "learning_rate": 2.1556886227544907e-05,
1555
+ "loss": 0.1004,
1556
+ "step": 2580
1557
+ },
1558
+ {
1559
+ "epoch": 30.83,
1560
+ "learning_rate": 2.043413173652694e-05,
1561
+ "loss": 0.1048,
1562
+ "step": 2590
1563
+ },
1564
+ {
1565
+ "epoch": 30.95,
1566
+ "learning_rate": 1.931137724550898e-05,
1567
+ "loss": 0.0978,
1568
+ "step": 2600
1569
+ },
1570
+ {
1571
+ "epoch": 31.07,
1572
+ "learning_rate": 1.8188622754491017e-05,
1573
+ "loss": 0.0981,
1574
+ "step": 2610
1575
+ },
1576
+ {
1577
+ "epoch": 31.19,
1578
+ "learning_rate": 1.7065868263473052e-05,
1579
+ "loss": 0.0979,
1580
+ "step": 2620
1581
+ },
1582
+ {
1583
+ "epoch": 31.31,
1584
+ "learning_rate": 1.594311377245509e-05,
1585
+ "loss": 0.103,
1586
+ "step": 2630
1587
+ },
1588
+ {
1589
+ "epoch": 31.43,
1590
+ "learning_rate": 1.4820359281437124e-05,
1591
+ "loss": 0.1065,
1592
+ "step": 2640
1593
+ },
1594
+ {
1595
+ "epoch": 31.55,
1596
+ "learning_rate": 1.369760479041916e-05,
1597
+ "loss": 0.0912,
1598
+ "step": 2650
1599
+ },
1600
+ {
1601
+ "epoch": 31.67,
1602
+ "learning_rate": 1.2574850299401195e-05,
1603
+ "loss": 0.0927,
1604
+ "step": 2660
1605
+ },
1606
+ {
1607
+ "epoch": 31.79,
1608
+ "learning_rate": 1.1452095808383233e-05,
1609
+ "loss": 0.0942,
1610
+ "step": 2670
1611
+ },
1612
+ {
1613
+ "epoch": 31.9,
1614
+ "learning_rate": 1.032934131736527e-05,
1615
+ "loss": 0.0896,
1616
+ "step": 2680
1617
+ },
1618
+ {
1619
+ "epoch": 32.02,
1620
+ "learning_rate": 9.206586826347304e-06,
1621
+ "loss": 0.0965,
1622
+ "step": 2690
1623
+ },
1624
+ {
1625
+ "epoch": 32.14,
1626
+ "learning_rate": 8.08383233532934e-06,
1627
+ "loss": 0.0943,
1628
+ "step": 2700
1629
+ }
1630
+ ],
1631
+ "max_steps": 2772,
1632
+ "num_train_epochs": 33,
1633
+ "total_flos": 1.5684921338658816e+18,
1634
+ "trial_name": null,
1635
+ "trial_params": null
1636
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccebab85e4dcd23b16c1544739b7d1206238f9ebf5151885432bf020c9346adf
3
+ size 3707