alvarobartt commited on
Commit
8d0cf53
1 Parent(s): 2839446

End of training

Browse files
README.md ADDED
@@ -0,0 +1,78 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: mistralai/Mistral-7B-v0.1
4
+ tags:
5
+ - trl
6
+ - orpo
7
+ - generated_from_trainer
8
+ model-index:
9
+ - name: mistral-orpo-mix-beta
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # mistral-orpo-mix-beta
17
+
18
+ This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the None dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 1.2152
21
+ - Rewards/chosen: -0.1113
22
+ - Rewards/rejected: -0.1461
23
+ - Rewards/accuracies: 0.6562
24
+ - Rewards/margins: 0.0348
25
+ - Logps/rejected: -1.4607
26
+ - Logps/chosen: -1.1130
27
+ - Logits/rejected: -2.4201
28
+ - Logits/chosen: -2.4178
29
+ - Nll Loss: 1.1711
30
+ - Log Odds Ratio: -0.6108
31
+ - Log Odds Chosen: 0.5395
32
+
33
+ ## Model description
34
+
35
+ More information needed
36
+
37
+ ## Intended uses & limitations
38
+
39
+ More information needed
40
+
41
+ ## Training and evaluation data
42
+
43
+ More information needed
44
+
45
+ ## Training procedure
46
+
47
+ ### Training hyperparameters
48
+
49
+ The following hyperparameters were used during training:
50
+ - learning_rate: 5e-06
51
+ - train_batch_size: 8
52
+ - eval_batch_size: 8
53
+ - seed: 42
54
+ - distributed_type: multi-GPU
55
+ - num_devices: 4
56
+ - total_train_batch_size: 32
57
+ - total_eval_batch_size: 32
58
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
59
+ - lr_scheduler_type: inverse_sqrt
60
+ - lr_scheduler_warmup_ratio: 0.1
61
+ - lr_scheduler_warmup_steps: 100
62
+ - num_epochs: 3
63
+
64
+ ### Training results
65
+
66
+ | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen | Nll Loss | Log Odds Ratio | Log Odds Chosen |
67
+ |:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|:--------:|:--------------:|:---------------:|
68
+ | 0.9961 | 1.0 | 211 | 0.9199 | -0.0831 | -0.1055 | 0.6302 | 0.0224 | -1.0547 | -0.8311 | -2.3542 | -2.3590 | 0.8774 | -0.6294 | 0.3896 |
69
+ | 0.4999 | 2.0 | 422 | 0.9813 | -0.0894 | -0.1184 | 0.6615 | 0.0290 | -1.1836 | -0.8937 | -2.5184 | -2.5177 | 0.9404 | -0.6012 | 0.4994 |
70
+ | 0.1899 | 3.0 | 633 | 1.2152 | -0.1113 | -0.1461 | 0.6562 | 0.0348 | -1.4607 | -1.1130 | -2.4201 | -2.4178 | 1.1711 | -0.6108 | 0.5395 |
71
+
72
+
73
+ ### Framework versions
74
+
75
+ - Transformers 4.39.0
76
+ - Pytorch 2.1.1+cu121
77
+ - Datasets 2.16.1
78
+ - Tokenizers 0.15.2
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<|im_end|>": 32001,
3
+ "<|im_start|>": 32000
4
+ }
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "mistralai/Mistral-7B-v0.1",
3
+ "architectures": [
4
+ "MistralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 4096,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 14336,
13
+ "max_position_embeddings": 32768,
14
+ "model_type": "mistral",
15
+ "num_attention_heads": 32,
16
+ "num_hidden_layers": 32,
17
+ "num_key_value_heads": 8,
18
+ "rms_norm_eps": 1e-05,
19
+ "rope_theta": 10000.0,
20
+ "sliding_window": 4096,
21
+ "tie_word_embeddings": false,
22
+ "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.39.0",
24
+ "use_cache": true,
25
+ "vocab_size": 32002
26
+ }
final_checkpoint/added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<|im_end|>": 32001,
3
+ "<|im_start|>": 32000
4
+ }
final_checkpoint/config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "mistralai/Mistral-7B-v0.1",
3
+ "architectures": [
4
+ "MistralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 4096,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 14336,
13
+ "max_position_embeddings": 32768,
14
+ "model_type": "mistral",
15
+ "num_attention_heads": 32,
16
+ "num_hidden_layers": 32,
17
+ "num_key_value_heads": 8,
18
+ "rms_norm_eps": 1e-05,
19
+ "rope_theta": 10000.0,
20
+ "sliding_window": 4096,
21
+ "tie_word_embeddings": false,
22
+ "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.39.0",
24
+ "use_cache": true,
25
+ "vocab_size": 32002
26
+ }
final_checkpoint/special_tokens_map.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<|im_start|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "<|im_end|>",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ }
17
+ ],
18
+ "bos_token": "<|im_start|>",
19
+ "eos_token": "<|im_end|>",
20
+ "pad_token": "<|im_end|>",
21
+ "unk_token": {
22
+ "content": "<unk>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false
27
+ }
28
+ }
final_checkpoint/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
final_checkpoint/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
final_checkpoint/tokenizer_config.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "32000": {
30
+ "content": "<|im_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "32001": {
38
+ "content": "<|im_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ }
45
+ },
46
+ "additional_special_tokens": [
47
+ "<|im_start|>",
48
+ "<|im_end|>"
49
+ ],
50
+ "bos_token": "<|im_start|>",
51
+ "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
52
+ "clean_up_tokenization_spaces": false,
53
+ "eos_token": "<|im_end|>",
54
+ "legacy": true,
55
+ "model_max_length": 2048,
56
+ "pad_token": "<|im_end|>",
57
+ "sp_model_kwargs": {},
58
+ "spaces_between_special_tokens": false,
59
+ "tokenizer_class": "LlamaTokenizer",
60
+ "unk_token": "<unk>",
61
+ "use_default_system_prompt": false
62
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 32000,
4
+ "eos_token_id": 32001,
5
+ "pad_token_id": 32001,
6
+ "transformers_version": "4.39.0"
7
+ }
model-00001-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba9dfcf688bfdd55a8f711a646014d65ed31c1855ea4b247eea69677f63ab26e
3
+ size 4943178720
model-00002-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0d8c0fc0a340374dfddd590f5070f60dfa480394a08b1c6e94378a339f1c3d9
3
+ size 4999819336
model-00003-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:734f7e2a0baefd53bac6d6d88304db9295b9a9f892879ebcd3900d3990d059ea
3
+ size 4540532728
model.safetensors.index.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 14483496960
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00003-of-00003.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00003.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00003.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
13
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
14
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
15
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
16
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
17
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00003.safetensors",
18
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
19
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
20
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
21
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
22
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
23
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
24
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
25
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
26
+ "model.layers.10.input_layernorm.weight": "model-00002-of-00003.safetensors",
27
+ "model.layers.10.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
28
+ "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
29
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
30
+ "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
31
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
32
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
33
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
34
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
35
+ "model.layers.11.input_layernorm.weight": "model-00002-of-00003.safetensors",
36
+ "model.layers.11.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
37
+ "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
38
+ "model.layers.11.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
39
+ "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
40
+ "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
41
+ "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
42
+ "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
43
+ "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
44
+ "model.layers.12.input_layernorm.weight": "model-00002-of-00003.safetensors",
45
+ "model.layers.12.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
46
+ "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
47
+ "model.layers.12.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
48
+ "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
49
+ "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
50
+ "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
51
+ "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
52
+ "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
53
+ "model.layers.13.input_layernorm.weight": "model-00002-of-00003.safetensors",
54
+ "model.layers.13.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
55
+ "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
56
+ "model.layers.13.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
57
+ "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
58
+ "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
59
+ "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
60
+ "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
61
+ "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
62
+ "model.layers.14.input_layernorm.weight": "model-00002-of-00003.safetensors",
63
+ "model.layers.14.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
64
+ "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
65
+ "model.layers.14.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
66
+ "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
67
+ "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
68
+ "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
69
+ "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
70
+ "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
71
+ "model.layers.15.input_layernorm.weight": "model-00002-of-00003.safetensors",
72
+ "model.layers.15.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
73
+ "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
74
+ "model.layers.15.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
75
+ "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
76
+ "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
77
+ "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
78
+ "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
79
+ "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
80
+ "model.layers.16.input_layernorm.weight": "model-00002-of-00003.safetensors",
81
+ "model.layers.16.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
82
+ "model.layers.16.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
83
+ "model.layers.16.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
84
+ "model.layers.16.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
85
+ "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
86
+ "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
87
+ "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
88
+ "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
89
+ "model.layers.17.input_layernorm.weight": "model-00002-of-00003.safetensors",
90
+ "model.layers.17.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
91
+ "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
92
+ "model.layers.17.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
93
+ "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
94
+ "model.layers.17.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
95
+ "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
96
+ "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
97
+ "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
98
+ "model.layers.18.input_layernorm.weight": "model-00002-of-00003.safetensors",
99
+ "model.layers.18.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
100
+ "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
101
+ "model.layers.18.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
102
+ "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
103
+ "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
104
+ "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
105
+ "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
106
+ "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
107
+ "model.layers.19.input_layernorm.weight": "model-00002-of-00003.safetensors",
108
+ "model.layers.19.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
109
+ "model.layers.19.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
110
+ "model.layers.19.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
111
+ "model.layers.19.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
112
+ "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
113
+ "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
114
+ "model.layers.19.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
115
+ "model.layers.19.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
116
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00003.safetensors",
117
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
118
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
119
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
120
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
121
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
122
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
123
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
124
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
125
+ "model.layers.20.input_layernorm.weight": "model-00002-of-00003.safetensors",
126
+ "model.layers.20.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
127
+ "model.layers.20.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
128
+ "model.layers.20.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
129
+ "model.layers.20.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
130
+ "model.layers.20.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
131
+ "model.layers.20.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
132
+ "model.layers.20.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
133
+ "model.layers.20.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
134
+ "model.layers.21.input_layernorm.weight": "model-00002-of-00003.safetensors",
135
+ "model.layers.21.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
136
+ "model.layers.21.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
137
+ "model.layers.21.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
138
+ "model.layers.21.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
139
+ "model.layers.21.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
140
+ "model.layers.21.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
141
+ "model.layers.21.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
142
+ "model.layers.21.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
143
+ "model.layers.22.input_layernorm.weight": "model-00003-of-00003.safetensors",
144
+ "model.layers.22.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
145
+ "model.layers.22.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
146
+ "model.layers.22.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
147
+ "model.layers.22.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
148
+ "model.layers.22.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
149
+ "model.layers.22.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
150
+ "model.layers.22.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
151
+ "model.layers.22.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
152
+ "model.layers.23.input_layernorm.weight": "model-00003-of-00003.safetensors",
153
+ "model.layers.23.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
154
+ "model.layers.23.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
155
+ "model.layers.23.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
156
+ "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
157
+ "model.layers.23.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
158
+ "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
159
+ "model.layers.23.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
160
+ "model.layers.23.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
161
+ "model.layers.24.input_layernorm.weight": "model-00003-of-00003.safetensors",
162
+ "model.layers.24.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
163
+ "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
164
+ "model.layers.24.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
165
+ "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
166
+ "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
167
+ "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
168
+ "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
169
+ "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
170
+ "model.layers.25.input_layernorm.weight": "model-00003-of-00003.safetensors",
171
+ "model.layers.25.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
172
+ "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
173
+ "model.layers.25.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
174
+ "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
175
+ "model.layers.25.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
176
+ "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
177
+ "model.layers.25.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
178
+ "model.layers.25.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
179
+ "model.layers.26.input_layernorm.weight": "model-00003-of-00003.safetensors",
180
+ "model.layers.26.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
181
+ "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
182
+ "model.layers.26.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
183
+ "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
184
+ "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
185
+ "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
186
+ "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
187
+ "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
188
+ "model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
189
+ "model.layers.27.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
190
+ "model.layers.27.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
191
+ "model.layers.27.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
192
+ "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
193
+ "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
194
+ "model.layers.27.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
195
+ "model.layers.27.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
196
+ "model.layers.27.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
197
+ "model.layers.28.input_layernorm.weight": "model-00003-of-00003.safetensors",
198
+ "model.layers.28.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
199
+ "model.layers.28.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
200
+ "model.layers.28.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
201
+ "model.layers.28.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
202
+ "model.layers.28.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
203
+ "model.layers.28.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
204
+ "model.layers.28.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
205
+ "model.layers.28.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
206
+ "model.layers.29.input_layernorm.weight": "model-00003-of-00003.safetensors",
207
+ "model.layers.29.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
208
+ "model.layers.29.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
209
+ "model.layers.29.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
210
+ "model.layers.29.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
211
+ "model.layers.29.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
212
+ "model.layers.29.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
213
+ "model.layers.29.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
214
+ "model.layers.29.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
215
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
216
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
217
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
218
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
219
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
220
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
221
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
222
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
223
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
224
+ "model.layers.30.input_layernorm.weight": "model-00003-of-00003.safetensors",
225
+ "model.layers.30.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
226
+ "model.layers.30.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
227
+ "model.layers.30.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
228
+ "model.layers.30.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
229
+ "model.layers.30.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
230
+ "model.layers.30.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
231
+ "model.layers.30.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
232
+ "model.layers.30.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
233
+ "model.layers.31.input_layernorm.weight": "model-00003-of-00003.safetensors",
234
+ "model.layers.31.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
235
+ "model.layers.31.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
236
+ "model.layers.31.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
237
+ "model.layers.31.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
238
+ "model.layers.31.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
239
+ "model.layers.31.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
240
+ "model.layers.31.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
241
+ "model.layers.31.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
242
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00003.safetensors",
243
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
244
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
245
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
246
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
247
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
248
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
249
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
250
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
251
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00003.safetensors",
252
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
253
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
254
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
255
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
256
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
257
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
258
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
259
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
260
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00003.safetensors",
261
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
262
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
263
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
264
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
265
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
266
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
267
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
268
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
269
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00003.safetensors",
270
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
271
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
272
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
273
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
274
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
275
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
276
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
277
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
278
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00003.safetensors",
279
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
280
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
281
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
282
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
283
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
284
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
285
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
286
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
287
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00003.safetensors",
288
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
289
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
290
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
291
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
292
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
293
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
294
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
295
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
296
+ "model.norm.weight": "model-00003-of-00003.safetensors"
297
+ }
298
+ }
runs/Mar26_03-51-15_alvarobartt-kind-frog-1-0/events.out.tfevents.1711439538.alvarobartt-kind-frog-1-0.445.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9026aed9b1566f975b01ce90693209832a3d4d347d152972bd92f768994bb7ef
3
+ size 62089
special_tokens_map.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<|im_start|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "<|im_end|>",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ }
17
+ ],
18
+ "bos_token": "<|im_start|>",
19
+ "eos_token": "<|im_end|>",
20
+ "pad_token": "<|im_end|>",
21
+ "unk_token": {
22
+ "content": "<unk>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false
27
+ }
28
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
tokenizer_config.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "32000": {
30
+ "content": "<|im_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "32001": {
38
+ "content": "<|im_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ }
45
+ },
46
+ "additional_special_tokens": [
47
+ "<|im_start|>",
48
+ "<|im_end|>"
49
+ ],
50
+ "bos_token": "<|im_start|>",
51
+ "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
52
+ "clean_up_tokenization_spaces": false,
53
+ "eos_token": "<|im_end|>",
54
+ "legacy": true,
55
+ "model_max_length": 2048,
56
+ "pad_token": "<|im_end|>",
57
+ "sp_model_kwargs": {},
58
+ "spaces_between_special_tokens": false,
59
+ "tokenizer_class": "LlamaTokenizer",
60
+ "unk_token": "<unk>",
61
+ "use_default_system_prompt": false
62
+ }
trainer_state.json ADDED
@@ -0,0 +1,1221 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 633,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.05,
13
+ "grad_norm": 45.76888063592087,
14
+ "learning_rate": 5.000000000000001e-07,
15
+ "log_odds_chosen": 0.2575085163116455,
16
+ "log_odds_ratio": -0.704267144203186,
17
+ "logits/chosen": -2.8495709896087646,
18
+ "logits/rejected": -2.8519797325134277,
19
+ "logps/chosen": -1.2749807834625244,
20
+ "logps/rejected": -1.5134317874908447,
21
+ "loss": 1.2045,
22
+ "nll_loss": 1.1489894390106201,
23
+ "rewards/accuracies": 0.5625,
24
+ "rewards/chosen": -0.127498060464859,
25
+ "rewards/margins": 0.023845110088586807,
26
+ "rewards/rejected": -0.1513431817293167,
27
+ "step": 10
28
+ },
29
+ {
30
+ "epoch": 0.09,
31
+ "grad_norm": 17.550572571834728,
32
+ "learning_rate": 1.0000000000000002e-06,
33
+ "log_odds_chosen": 0.18295660614967346,
34
+ "log_odds_ratio": -0.7083484530448914,
35
+ "logits/chosen": -2.9011380672454834,
36
+ "logits/rejected": -2.903749465942383,
37
+ "logps/chosen": -0.9543957710266113,
38
+ "logps/rejected": -1.0926854610443115,
39
+ "loss": 1.0092,
40
+ "nll_loss": 0.9102001190185547,
41
+ "rewards/accuracies": 0.5,
42
+ "rewards/chosen": -0.09543957561254501,
43
+ "rewards/margins": 0.013828963041305542,
44
+ "rewards/rejected": -0.10926854610443115,
45
+ "step": 20
46
+ },
47
+ {
48
+ "epoch": 0.14,
49
+ "grad_norm": 10.772595379798407,
50
+ "learning_rate": 1.5e-06,
51
+ "log_odds_chosen": 0.22791412472724915,
52
+ "log_odds_ratio": -0.6785577535629272,
53
+ "logits/chosen": -2.970428943634033,
54
+ "logits/rejected": -2.9696388244628906,
55
+ "logps/chosen": -0.9358822107315063,
56
+ "logps/rejected": -1.0665825605392456,
57
+ "loss": 0.968,
58
+ "nll_loss": 0.8899785280227661,
59
+ "rewards/accuracies": 0.6000000238418579,
60
+ "rewards/chosen": -0.0935882106423378,
61
+ "rewards/margins": 0.013070052489638329,
62
+ "rewards/rejected": -0.10665826499462128,
63
+ "step": 30
64
+ },
65
+ {
66
+ "epoch": 0.19,
67
+ "grad_norm": 11.974913883612869,
68
+ "learning_rate": 2.0000000000000003e-06,
69
+ "log_odds_chosen": 0.12087659537792206,
70
+ "log_odds_ratio": -0.7188054323196411,
71
+ "logits/chosen": -2.883781671524048,
72
+ "logits/rejected": -2.8842344284057617,
73
+ "logps/chosen": -0.9183737635612488,
74
+ "logps/rejected": -0.9950372576713562,
75
+ "loss": 0.9799,
76
+ "nll_loss": 0.8861400485038757,
77
+ "rewards/accuracies": 0.4749999940395355,
78
+ "rewards/chosen": -0.09183738380670547,
79
+ "rewards/margins": 0.007666352204978466,
80
+ "rewards/rejected": -0.09950373321771622,
81
+ "step": 40
82
+ },
83
+ {
84
+ "epoch": 0.24,
85
+ "grad_norm": 13.063526867338696,
86
+ "learning_rate": 2.5e-06,
87
+ "log_odds_chosen": 0.4301052689552307,
88
+ "log_odds_ratio": -0.6214904189109802,
89
+ "logits/chosen": -2.832704782485962,
90
+ "logits/rejected": -2.8430252075195312,
91
+ "logps/chosen": -0.9002718925476074,
92
+ "logps/rejected": -1.2055391073226929,
93
+ "loss": 0.9584,
94
+ "nll_loss": 0.9324554204940796,
95
+ "rewards/accuracies": 0.637499988079071,
96
+ "rewards/chosen": -0.09002719074487686,
97
+ "rewards/margins": 0.03052671253681183,
98
+ "rewards/rejected": -0.12055391073226929,
99
+ "step": 50
100
+ },
101
+ {
102
+ "epoch": 0.28,
103
+ "grad_norm": 10.235387913194325,
104
+ "learning_rate": 3e-06,
105
+ "log_odds_chosen": 0.3391026258468628,
106
+ "log_odds_ratio": -0.6046633720397949,
107
+ "logits/chosen": -2.597712993621826,
108
+ "logits/rejected": -2.598874568939209,
109
+ "logps/chosen": -0.7375704050064087,
110
+ "logps/rejected": -0.9383118748664856,
111
+ "loss": 0.8856,
112
+ "nll_loss": 0.7253891229629517,
113
+ "rewards/accuracies": 0.6625000238418579,
114
+ "rewards/chosen": -0.07375704497098923,
115
+ "rewards/margins": 0.020074138417840004,
116
+ "rewards/rejected": -0.09383118897676468,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.33,
121
+ "grad_norm": 10.615024222922383,
122
+ "learning_rate": 3.5e-06,
123
+ "log_odds_chosen": 0.34223267436027527,
124
+ "log_odds_ratio": -0.6422346830368042,
125
+ "logits/chosen": -2.540403366088867,
126
+ "logits/rejected": -2.5442099571228027,
127
+ "logps/chosen": -0.8444018363952637,
128
+ "logps/rejected": -1.083935022354126,
129
+ "loss": 0.9384,
130
+ "nll_loss": 0.8686600923538208,
131
+ "rewards/accuracies": 0.6000000238418579,
132
+ "rewards/chosen": -0.08444017916917801,
133
+ "rewards/margins": 0.02395332232117653,
134
+ "rewards/rejected": -0.10839351266622543,
135
+ "step": 70
136
+ },
137
+ {
138
+ "epoch": 0.38,
139
+ "grad_norm": 10.120353883295946,
140
+ "learning_rate": 4.000000000000001e-06,
141
+ "log_odds_chosen": 0.026019399985671043,
142
+ "log_odds_ratio": -0.7628353834152222,
143
+ "logits/chosen": -2.6254281997680664,
144
+ "logits/rejected": -2.6229329109191895,
145
+ "logps/chosen": -0.8879353404045105,
146
+ "logps/rejected": -0.9107568860054016,
147
+ "loss": 0.921,
148
+ "nll_loss": 0.9201810956001282,
149
+ "rewards/accuracies": 0.4749999940395355,
150
+ "rewards/chosen": -0.08879353106021881,
151
+ "rewards/margins": 0.0022821640595793724,
152
+ "rewards/rejected": -0.09107569605112076,
153
+ "step": 80
154
+ },
155
+ {
156
+ "epoch": 0.43,
157
+ "grad_norm": 9.558778662613774,
158
+ "learning_rate": 4.5e-06,
159
+ "log_odds_chosen": 0.29862236976623535,
160
+ "log_odds_ratio": -0.6326144933700562,
161
+ "logits/chosen": -2.6428685188293457,
162
+ "logits/rejected": -2.621931552886963,
163
+ "logps/chosen": -0.8801389932632446,
164
+ "logps/rejected": -1.0846771001815796,
165
+ "loss": 0.9325,
166
+ "nll_loss": 0.8730384707450867,
167
+ "rewards/accuracies": 0.574999988079071,
168
+ "rewards/chosen": -0.0880139023065567,
169
+ "rewards/margins": 0.020453810691833496,
170
+ "rewards/rejected": -0.1084677204489708,
171
+ "step": 90
172
+ },
173
+ {
174
+ "epoch": 0.47,
175
+ "grad_norm": 11.437968315222852,
176
+ "learning_rate": 5e-06,
177
+ "log_odds_chosen": 0.3877044916152954,
178
+ "log_odds_ratio": -0.6339091658592224,
179
+ "logits/chosen": -2.583739757537842,
180
+ "logits/rejected": -2.6161398887634277,
181
+ "logps/chosen": -0.8836472630500793,
182
+ "logps/rejected": -1.1088001728057861,
183
+ "loss": 0.9742,
184
+ "nll_loss": 0.9505327939987183,
185
+ "rewards/accuracies": 0.6875,
186
+ "rewards/chosen": -0.08836471289396286,
187
+ "rewards/margins": 0.02251528576016426,
188
+ "rewards/rejected": -0.11088001728057861,
189
+ "step": 100
190
+ },
191
+ {
192
+ "epoch": 0.52,
193
+ "grad_norm": 9.765547500332794,
194
+ "learning_rate": 4.767312946227961e-06,
195
+ "log_odds_chosen": 0.33769896626472473,
196
+ "log_odds_ratio": -0.6257166266441345,
197
+ "logits/chosen": -2.6281538009643555,
198
+ "logits/rejected": -2.6198818683624268,
199
+ "logps/chosen": -0.7726272344589233,
200
+ "logps/rejected": -0.9787896871566772,
201
+ "loss": 0.9403,
202
+ "nll_loss": 0.8244824409484863,
203
+ "rewards/accuracies": 0.675000011920929,
204
+ "rewards/chosen": -0.07726272195577621,
205
+ "rewards/margins": 0.020616240799427032,
206
+ "rewards/rejected": -0.09787896275520325,
207
+ "step": 110
208
+ },
209
+ {
210
+ "epoch": 0.57,
211
+ "grad_norm": 10.192809233857067,
212
+ "learning_rate": 4.564354645876385e-06,
213
+ "log_odds_chosen": 0.31468966603279114,
214
+ "log_odds_ratio": -0.6638727784156799,
215
+ "logits/chosen": -2.6107370853424072,
216
+ "logits/rejected": -2.601783275604248,
217
+ "logps/chosen": -0.8392475247383118,
218
+ "logps/rejected": -1.036677598953247,
219
+ "loss": 0.9841,
220
+ "nll_loss": 0.8654748797416687,
221
+ "rewards/accuracies": 0.5874999761581421,
222
+ "rewards/chosen": -0.08392475545406342,
223
+ "rewards/margins": 0.019743015989661217,
224
+ "rewards/rejected": -0.10366777330636978,
225
+ "step": 120
226
+ },
227
+ {
228
+ "epoch": 0.62,
229
+ "grad_norm": 11.038589594109835,
230
+ "learning_rate": 4.385290096535147e-06,
231
+ "log_odds_chosen": 0.2760984003543854,
232
+ "log_odds_ratio": -0.6513851881027222,
233
+ "logits/chosen": -2.566074848175049,
234
+ "logits/rejected": -2.565674304962158,
235
+ "logps/chosen": -0.833086371421814,
236
+ "logps/rejected": -0.9857648611068726,
237
+ "loss": 0.9998,
238
+ "nll_loss": 0.9077309370040894,
239
+ "rewards/accuracies": 0.6000000238418579,
240
+ "rewards/chosen": -0.0833086371421814,
241
+ "rewards/margins": 0.015267851762473583,
242
+ "rewards/rejected": -0.09857650101184845,
243
+ "step": 130
244
+ },
245
+ {
246
+ "epoch": 0.66,
247
+ "grad_norm": 9.191362766920426,
248
+ "learning_rate": 4.2257712736425835e-06,
249
+ "log_odds_chosen": 0.37576600909233093,
250
+ "log_odds_ratio": -0.6204187273979187,
251
+ "logits/chosen": -2.577667474746704,
252
+ "logits/rejected": -2.584357976913452,
253
+ "logps/chosen": -0.8501531481742859,
254
+ "logps/rejected": -1.1260815858840942,
255
+ "loss": 0.9545,
256
+ "nll_loss": 0.8476670980453491,
257
+ "rewards/accuracies": 0.699999988079071,
258
+ "rewards/chosen": -0.08501531183719635,
259
+ "rewards/margins": 0.027592848986387253,
260
+ "rewards/rejected": -0.1126081570982933,
261
+ "step": 140
262
+ },
263
+ {
264
+ "epoch": 0.71,
265
+ "grad_norm": 10.020928126499054,
266
+ "learning_rate": 4.082482904638631e-06,
267
+ "log_odds_chosen": 0.40555381774902344,
268
+ "log_odds_ratio": -0.6359796524047852,
269
+ "logits/chosen": -2.624826669692993,
270
+ "logits/rejected": -2.6218724250793457,
271
+ "logps/chosen": -0.8737274408340454,
272
+ "logps/rejected": -1.1572520732879639,
273
+ "loss": 0.9391,
274
+ "nll_loss": 0.8726488947868347,
275
+ "rewards/accuracies": 0.612500011920929,
276
+ "rewards/chosen": -0.08737273514270782,
277
+ "rewards/margins": 0.028352469205856323,
278
+ "rewards/rejected": -0.11572520434856415,
279
+ "step": 150
280
+ },
281
+ {
282
+ "epoch": 0.76,
283
+ "grad_norm": 11.288206284049263,
284
+ "learning_rate": 3.952847075210474e-06,
285
+ "log_odds_chosen": 0.35032859444618225,
286
+ "log_odds_ratio": -0.6347723007202148,
287
+ "logits/chosen": -2.7082316875457764,
288
+ "logits/rejected": -2.711594820022583,
289
+ "logps/chosen": -0.8456757664680481,
290
+ "logps/rejected": -1.0942612886428833,
291
+ "loss": 0.9052,
292
+ "nll_loss": 0.8740180730819702,
293
+ "rewards/accuracies": 0.6000000238418579,
294
+ "rewards/chosen": -0.0845675840973854,
295
+ "rewards/margins": 0.024858549237251282,
296
+ "rewards/rejected": -0.10942612588405609,
297
+ "step": 160
298
+ },
299
+ {
300
+ "epoch": 0.81,
301
+ "grad_norm": 9.222481367340817,
302
+ "learning_rate": 3.834824944236852e-06,
303
+ "log_odds_chosen": 0.14469774067401886,
304
+ "log_odds_ratio": -0.6979460120201111,
305
+ "logits/chosen": -2.575446844100952,
306
+ "logits/rejected": -2.572465419769287,
307
+ "logps/chosen": -0.8792887926101685,
308
+ "logps/rejected": -0.9753986597061157,
309
+ "loss": 0.9229,
310
+ "nll_loss": 0.8886219263076782,
311
+ "rewards/accuracies": 0.5375000238418579,
312
+ "rewards/chosen": -0.0879288762807846,
313
+ "rewards/margins": 0.009610971435904503,
314
+ "rewards/rejected": -0.09753986448049545,
315
+ "step": 170
316
+ },
317
+ {
318
+ "epoch": 0.85,
319
+ "grad_norm": 10.622839554119508,
320
+ "learning_rate": 3.72677996249965e-06,
321
+ "log_odds_chosen": 0.18417394161224365,
322
+ "log_odds_ratio": -0.7312262058258057,
323
+ "logits/chosen": -2.558830976486206,
324
+ "logits/rejected": -2.5850634574890137,
325
+ "logps/chosen": -0.9030648469924927,
326
+ "logps/rejected": -1.0534106492996216,
327
+ "loss": 0.9502,
328
+ "nll_loss": 0.9085375070571899,
329
+ "rewards/accuracies": 0.5,
330
+ "rewards/chosen": -0.09030647575855255,
331
+ "rewards/margins": 0.015034586191177368,
332
+ "rewards/rejected": -0.10534106194972992,
333
+ "step": 180
334
+ },
335
+ {
336
+ "epoch": 0.9,
337
+ "grad_norm": 10.63166001842217,
338
+ "learning_rate": 3.6273812505500587e-06,
339
+ "log_odds_chosen": 0.26296746730804443,
340
+ "log_odds_ratio": -0.6528322100639343,
341
+ "logits/chosen": -2.5235934257507324,
342
+ "logits/rejected": -2.509213924407959,
343
+ "logps/chosen": -0.842459499835968,
344
+ "logps/rejected": -1.0031650066375732,
345
+ "loss": 0.935,
346
+ "nll_loss": 0.840924084186554,
347
+ "rewards/accuracies": 0.6625000238418579,
348
+ "rewards/chosen": -0.0842459425330162,
349
+ "rewards/margins": 0.016070572659373283,
350
+ "rewards/rejected": -0.10031652450561523,
351
+ "step": 190
352
+ },
353
+ {
354
+ "epoch": 0.95,
355
+ "grad_norm": 10.265720627085773,
356
+ "learning_rate": 3.5355339059327378e-06,
357
+ "log_odds_chosen": 0.20157551765441895,
358
+ "log_odds_ratio": -0.6874672174453735,
359
+ "logits/chosen": -2.5119268894195557,
360
+ "logits/rejected": -2.516270399093628,
361
+ "logps/chosen": -0.8817191123962402,
362
+ "logps/rejected": -1.0264997482299805,
363
+ "loss": 0.9827,
364
+ "nll_loss": 0.9070207476615906,
365
+ "rewards/accuracies": 0.6000000238418579,
366
+ "rewards/chosen": -0.08817192167043686,
367
+ "rewards/margins": 0.01447806041687727,
368
+ "rewards/rejected": -0.10264997184276581,
369
+ "step": 200
370
+ },
371
+ {
372
+ "epoch": 1.0,
373
+ "grad_norm": 10.139451950242082,
374
+ "learning_rate": 3.450327796711771e-06,
375
+ "log_odds_chosen": 0.22201482951641083,
376
+ "log_odds_ratio": -0.6832079887390137,
377
+ "logits/chosen": -2.438990831375122,
378
+ "logits/rejected": -2.4239249229431152,
379
+ "logps/chosen": -0.9554345011711121,
380
+ "logps/rejected": -1.1055123805999756,
381
+ "loss": 0.9961,
382
+ "nll_loss": 0.9768595695495605,
383
+ "rewards/accuracies": 0.5874999761581421,
384
+ "rewards/chosen": -0.09554344415664673,
385
+ "rewards/margins": 0.01500779576599598,
386
+ "rewards/rejected": -0.11055125296115875,
387
+ "step": 210
388
+ },
389
+ {
390
+ "epoch": 1.0,
391
+ "eval_log_odds_chosen": 0.3896147906780243,
392
+ "eval_log_odds_ratio": -0.6294437050819397,
393
+ "eval_logits/chosen": -2.359004259109497,
394
+ "eval_logits/rejected": -2.3542282581329346,
395
+ "eval_logps/chosen": -0.8311433792114258,
396
+ "eval_logps/rejected": -1.0547062158584595,
397
+ "eval_loss": 0.9198942184448242,
398
+ "eval_nll_loss": 0.8774265646934509,
399
+ "eval_rewards/accuracies": 0.6302083134651184,
400
+ "eval_rewards/chosen": -0.08311434835195541,
401
+ "eval_rewards/margins": 0.022356273606419563,
402
+ "eval_rewards/rejected": -0.10547062009572983,
403
+ "eval_runtime": 61.0411,
404
+ "eval_samples_per_second": 12.287,
405
+ "eval_steps_per_second": 0.393,
406
+ "step": 211
407
+ },
408
+ {
409
+ "epoch": 1.04,
410
+ "grad_norm": 8.627109146931504,
411
+ "learning_rate": 3.3709993123162106e-06,
412
+ "log_odds_chosen": 1.4988970756530762,
413
+ "log_odds_ratio": -0.30856823921203613,
414
+ "logits/chosen": -2.3857474327087402,
415
+ "logits/rejected": -2.3801207542419434,
416
+ "logps/chosen": -0.45462799072265625,
417
+ "logps/rejected": -1.2372629642486572,
418
+ "loss": 0.564,
419
+ "nll_loss": 0.48940524458885193,
420
+ "rewards/accuracies": 0.8999999761581421,
421
+ "rewards/chosen": -0.045462802052497864,
422
+ "rewards/margins": 0.07826349139213562,
423
+ "rewards/rejected": -0.12372628599405289,
424
+ "step": 220
425
+ },
426
+ {
427
+ "epoch": 1.09,
428
+ "grad_norm": 10.837022751981742,
429
+ "learning_rate": 3.296902366978936e-06,
430
+ "log_odds_chosen": 1.769687294960022,
431
+ "log_odds_ratio": -0.22798781096935272,
432
+ "logits/chosen": -2.460228204727173,
433
+ "logits/rejected": -2.4452919960021973,
434
+ "logps/chosen": -0.48005446791648865,
435
+ "logps/rejected": -1.4746625423431396,
436
+ "loss": 0.5158,
437
+ "nll_loss": 0.5252029895782471,
438
+ "rewards/accuracies": 0.9750000238418579,
439
+ "rewards/chosen": -0.04800545051693916,
440
+ "rewards/margins": 0.09946081787347794,
441
+ "rewards/rejected": -0.1474662721157074,
442
+ "step": 230
443
+ },
444
+ {
445
+ "epoch": 1.14,
446
+ "grad_norm": 9.220899180149999,
447
+ "learning_rate": 3.2274861218395142e-06,
448
+ "log_odds_chosen": 1.664380431175232,
449
+ "log_odds_ratio": -0.2567726671695709,
450
+ "logits/chosen": -2.4479331970214844,
451
+ "logits/rejected": -2.453798294067383,
452
+ "logps/chosen": -0.43317586183547974,
453
+ "logps/rejected": -1.2501039505004883,
454
+ "loss": 0.4887,
455
+ "nll_loss": 0.45028549432754517,
456
+ "rewards/accuracies": 0.9375,
457
+ "rewards/chosen": -0.04331757873296738,
458
+ "rewards/margins": 0.08169281482696533,
459
+ "rewards/rejected": -0.1250104010105133,
460
+ "step": 240
461
+ },
462
+ {
463
+ "epoch": 1.18,
464
+ "grad_norm": 9.9378304724897,
465
+ "learning_rate": 3.1622776601683796e-06,
466
+ "log_odds_chosen": 1.5019989013671875,
467
+ "log_odds_ratio": -0.298888623714447,
468
+ "logits/chosen": -2.5074825286865234,
469
+ "logits/rejected": -2.4912190437316895,
470
+ "logps/chosen": -0.4875302314758301,
471
+ "logps/rejected": -1.2663142681121826,
472
+ "loss": 0.5504,
473
+ "nll_loss": 0.5222305655479431,
474
+ "rewards/accuracies": 0.925000011920929,
475
+ "rewards/chosen": -0.04875302314758301,
476
+ "rewards/margins": 0.07787840068340302,
477
+ "rewards/rejected": -0.12663142383098602,
478
+ "step": 250
479
+ },
480
+ {
481
+ "epoch": 1.23,
482
+ "grad_norm": 10.42321736026305,
483
+ "learning_rate": 3.1008683647302113e-06,
484
+ "log_odds_chosen": 1.7215213775634766,
485
+ "log_odds_ratio": -0.2548612058162689,
486
+ "logits/chosen": -2.523898124694824,
487
+ "logits/rejected": -2.516082525253296,
488
+ "logps/chosen": -0.4069024920463562,
489
+ "logps/rejected": -1.2453140020370483,
490
+ "loss": 0.4878,
491
+ "nll_loss": 0.4438396394252777,
492
+ "rewards/accuracies": 0.925000011920929,
493
+ "rewards/chosen": -0.04069024696946144,
494
+ "rewards/margins": 0.08384115993976593,
495
+ "rewards/rejected": -0.12453138828277588,
496
+ "step": 260
497
+ },
498
+ {
499
+ "epoch": 1.28,
500
+ "grad_norm": 8.517195541285803,
501
+ "learning_rate": 3.0429030972509227e-06,
502
+ "log_odds_chosen": 1.4622563123703003,
503
+ "log_odds_ratio": -0.31491032242774963,
504
+ "logits/chosen": -2.421508312225342,
505
+ "logits/rejected": -2.4460442066192627,
506
+ "logps/chosen": -0.44591742753982544,
507
+ "logps/rejected": -1.127700924873352,
508
+ "loss": 0.5104,
509
+ "nll_loss": 0.48815250396728516,
510
+ "rewards/accuracies": 0.8999999761581421,
511
+ "rewards/chosen": -0.044591739773750305,
512
+ "rewards/margins": 0.06817835569381714,
513
+ "rewards/rejected": -0.11277009546756744,
514
+ "step": 270
515
+ },
516
+ {
517
+ "epoch": 1.33,
518
+ "grad_norm": 8.676525480706056,
519
+ "learning_rate": 2.988071523335984e-06,
520
+ "log_odds_chosen": 1.713667631149292,
521
+ "log_odds_ratio": -0.23984098434448242,
522
+ "logits/chosen": -2.433767795562744,
523
+ "logits/rejected": -2.4231998920440674,
524
+ "logps/chosen": -0.42581120133399963,
525
+ "logps/rejected": -1.254382848739624,
526
+ "loss": 0.4905,
527
+ "nll_loss": 0.4856826364994049,
528
+ "rewards/accuracies": 0.9624999761581421,
529
+ "rewards/chosen": -0.04258112236857414,
530
+ "rewards/margins": 0.0828571617603302,
531
+ "rewards/rejected": -0.12543828785419464,
532
+ "step": 280
533
+ },
534
+ {
535
+ "epoch": 1.37,
536
+ "grad_norm": 10.179660663431909,
537
+ "learning_rate": 2.9361010975735177e-06,
538
+ "log_odds_chosen": 1.6406803131103516,
539
+ "log_odds_ratio": -0.2432694435119629,
540
+ "logits/chosen": -2.40620493888855,
541
+ "logits/rejected": -2.462026357650757,
542
+ "logps/chosen": -0.43379467725753784,
543
+ "logps/rejected": -1.231093406677246,
544
+ "loss": 0.4952,
545
+ "nll_loss": 0.4767753481864929,
546
+ "rewards/accuracies": 0.9750000238418579,
547
+ "rewards/chosen": -0.043379463255405426,
548
+ "rewards/margins": 0.07972987741231918,
549
+ "rewards/rejected": -0.1231093555688858,
550
+ "step": 290
551
+ },
552
+ {
553
+ "epoch": 1.42,
554
+ "grad_norm": 10.354476045311776,
555
+ "learning_rate": 2.8867513459481293e-06,
556
+ "log_odds_chosen": 1.8753684759140015,
557
+ "log_odds_ratio": -0.24006888270378113,
558
+ "logits/chosen": -2.4737935066223145,
559
+ "logits/rejected": -2.483096122741699,
560
+ "logps/chosen": -0.3926279544830322,
561
+ "logps/rejected": -1.2253657579421997,
562
+ "loss": 0.4883,
563
+ "nll_loss": 0.4013836979866028,
564
+ "rewards/accuracies": 0.9624999761581421,
565
+ "rewards/chosen": -0.0392627976834774,
566
+ "rewards/margins": 0.08327377587556839,
567
+ "rewards/rejected": -0.12253657728433609,
568
+ "step": 300
569
+ },
570
+ {
571
+ "epoch": 1.47,
572
+ "grad_norm": 9.015044311790971,
573
+ "learning_rate": 2.839809171235324e-06,
574
+ "log_odds_chosen": 1.8872253894805908,
575
+ "log_odds_ratio": -0.22469110786914825,
576
+ "logits/chosen": -2.400991678237915,
577
+ "logits/rejected": -2.397636890411377,
578
+ "logps/chosen": -0.40627264976501465,
579
+ "logps/rejected": -1.4336292743682861,
580
+ "loss": 0.5011,
581
+ "nll_loss": 0.43345755338668823,
582
+ "rewards/accuracies": 0.987500011920929,
583
+ "rewards/chosen": -0.040627263486385345,
584
+ "rewards/margins": 0.10273567587137222,
585
+ "rewards/rejected": -0.14336295425891876,
586
+ "step": 310
587
+ },
588
+ {
589
+ "epoch": 1.52,
590
+ "grad_norm": 9.539457700499925,
591
+ "learning_rate": 2.7950849718747376e-06,
592
+ "log_odds_chosen": 1.8418588638305664,
593
+ "log_odds_ratio": -0.20935162901878357,
594
+ "logits/chosen": -2.3977291584014893,
595
+ "logits/rejected": -2.3498239517211914,
596
+ "logps/chosen": -0.4429972767829895,
597
+ "logps/rejected": -1.4245294332504272,
598
+ "loss": 0.4827,
599
+ "nll_loss": 0.4764658510684967,
600
+ "rewards/accuracies": 0.987500011920929,
601
+ "rewards/chosen": -0.04429972171783447,
602
+ "rewards/margins": 0.09815323352813721,
603
+ "rewards/rejected": -0.14245295524597168,
604
+ "step": 320
605
+ },
606
+ {
607
+ "epoch": 1.56,
608
+ "grad_norm": 9.434503294939443,
609
+ "learning_rate": 2.752409412815902e-06,
610
+ "log_odds_chosen": 1.5989511013031006,
611
+ "log_odds_ratio": -0.26907485723495483,
612
+ "logits/chosen": -2.4358298778533936,
613
+ "logits/rejected": -2.4641692638397217,
614
+ "logps/chosen": -0.4505313038825989,
615
+ "logps/rejected": -1.2999070882797241,
616
+ "loss": 0.4685,
617
+ "nll_loss": 0.49505728483200073,
618
+ "rewards/accuracies": 0.9375,
619
+ "rewards/chosen": -0.04505313187837601,
620
+ "rewards/margins": 0.08493757992982864,
621
+ "rewards/rejected": -0.12999072670936584,
622
+ "step": 330
623
+ },
624
+ {
625
+ "epoch": 1.61,
626
+ "grad_norm": 9.785466529429728,
627
+ "learning_rate": 2.711630722733202e-06,
628
+ "log_odds_chosen": 1.757354736328125,
629
+ "log_odds_ratio": -0.22594013810157776,
630
+ "logits/chosen": -2.386993885040283,
631
+ "logits/rejected": -2.3797829151153564,
632
+ "logps/chosen": -0.3836737275123596,
633
+ "logps/rejected": -1.226621389389038,
634
+ "loss": 0.5139,
635
+ "nll_loss": 0.4257877469062805,
636
+ "rewards/accuracies": 0.987500011920929,
637
+ "rewards/chosen": -0.0383673757314682,
638
+ "rewards/margins": 0.08429475873708725,
639
+ "rewards/rejected": -0.12266211211681366,
640
+ "step": 340
641
+ },
642
+ {
643
+ "epoch": 1.66,
644
+ "grad_norm": 9.783556860345605,
645
+ "learning_rate": 2.6726124191242444e-06,
646
+ "log_odds_chosen": 1.5875946283340454,
647
+ "log_odds_ratio": -0.282478392124176,
648
+ "logits/chosen": -2.4133667945861816,
649
+ "logits/rejected": -2.4031758308410645,
650
+ "logps/chosen": -0.47809138894081116,
651
+ "logps/rejected": -1.3111228942871094,
652
+ "loss": 0.4947,
653
+ "nll_loss": 0.46197766065597534,
654
+ "rewards/accuracies": 0.9750000238418579,
655
+ "rewards/chosen": -0.04780913516879082,
656
+ "rewards/margins": 0.08330314606428146,
657
+ "rewards/rejected": -0.13111227750778198,
658
+ "step": 350
659
+ },
660
+ {
661
+ "epoch": 1.71,
662
+ "grad_norm": 10.952497066310066,
663
+ "learning_rate": 2.6352313834736496e-06,
664
+ "log_odds_chosen": 1.444085717201233,
665
+ "log_odds_ratio": -0.302844375371933,
666
+ "logits/chosen": -2.35420560836792,
667
+ "logits/rejected": -2.40644907951355,
668
+ "logps/chosen": -0.4726253151893616,
669
+ "logps/rejected": -1.181214451789856,
670
+ "loss": 0.5344,
671
+ "nll_loss": 0.5417958498001099,
672
+ "rewards/accuracies": 0.8999999761581421,
673
+ "rewards/chosen": -0.0472625270485878,
674
+ "rewards/margins": 0.0708589106798172,
675
+ "rewards/rejected": -0.1181214451789856,
676
+ "step": 360
677
+ },
678
+ {
679
+ "epoch": 1.75,
680
+ "grad_norm": 9.337096555826308,
681
+ "learning_rate": 2.599376224550182e-06,
682
+ "log_odds_chosen": 1.5420761108398438,
683
+ "log_odds_ratio": -0.2706839144229889,
684
+ "logits/chosen": -2.393585681915283,
685
+ "logits/rejected": -2.3775479793548584,
686
+ "logps/chosen": -0.4294983744621277,
687
+ "logps/rejected": -1.177330493927002,
688
+ "loss": 0.489,
689
+ "nll_loss": 0.4465916156768799,
690
+ "rewards/accuracies": 0.949999988079071,
691
+ "rewards/chosen": -0.04294983297586441,
692
+ "rewards/margins": 0.07478321343660355,
693
+ "rewards/rejected": -0.11773304641246796,
694
+ "step": 370
695
+ },
696
+ {
697
+ "epoch": 1.8,
698
+ "grad_norm": 9.897099582842259,
699
+ "learning_rate": 2.564945880212886e-06,
700
+ "log_odds_chosen": 1.6586507558822632,
701
+ "log_odds_ratio": -0.2737709581851959,
702
+ "logits/chosen": -2.505715847015381,
703
+ "logits/rejected": -2.478332996368408,
704
+ "logps/chosen": -0.47623515129089355,
705
+ "logps/rejected": -1.3953008651733398,
706
+ "loss": 0.5232,
707
+ "nll_loss": 0.498232901096344,
708
+ "rewards/accuracies": 0.925000011920929,
709
+ "rewards/chosen": -0.047623515129089355,
710
+ "rewards/margins": 0.0919065773487091,
711
+ "rewards/rejected": -0.13953009247779846,
712
+ "step": 380
713
+ },
714
+ {
715
+ "epoch": 1.85,
716
+ "grad_norm": 8.744712266716895,
717
+ "learning_rate": 2.5318484177091667e-06,
718
+ "log_odds_chosen": 1.4230409860610962,
719
+ "log_odds_ratio": -0.29352661967277527,
720
+ "logits/chosen": -2.455392360687256,
721
+ "logits/rejected": -2.4511191844940186,
722
+ "logps/chosen": -0.5203135013580322,
723
+ "logps/rejected": -1.2620891332626343,
724
+ "loss": 0.5658,
725
+ "nll_loss": 0.547887921333313,
726
+ "rewards/accuracies": 0.925000011920929,
727
+ "rewards/chosen": -0.05203135684132576,
728
+ "rewards/margins": 0.0741775631904602,
729
+ "rewards/rejected": -0.12620893120765686,
730
+ "step": 390
731
+ },
732
+ {
733
+ "epoch": 1.9,
734
+ "grad_norm": 9.30311341527825,
735
+ "learning_rate": 2.5e-06,
736
+ "log_odds_chosen": 1.7381293773651123,
737
+ "log_odds_ratio": -0.25197070837020874,
738
+ "logits/chosen": -2.512232780456543,
739
+ "logits/rejected": -2.5167810916900635,
740
+ "logps/chosen": -0.4353383183479309,
741
+ "logps/rejected": -1.34853196144104,
742
+ "loss": 0.5158,
743
+ "nll_loss": 0.4797208309173584,
744
+ "rewards/accuracies": 0.9375,
745
+ "rewards/chosen": -0.04353383183479309,
746
+ "rewards/margins": 0.09131935983896255,
747
+ "rewards/rejected": -0.13485319912433624,
748
+ "step": 400
749
+ },
750
+ {
751
+ "epoch": 1.94,
752
+ "grad_norm": 9.272816070528886,
753
+ "learning_rate": 2.4693239916239746e-06,
754
+ "log_odds_chosen": 1.5647149085998535,
755
+ "log_odds_ratio": -0.29222574830055237,
756
+ "logits/chosen": -2.4848999977111816,
757
+ "logits/rejected": -2.4973931312561035,
758
+ "logps/chosen": -0.4486003816127777,
759
+ "logps/rejected": -1.2298743724822998,
760
+ "loss": 0.4991,
761
+ "nll_loss": 0.49500808119773865,
762
+ "rewards/accuracies": 0.9125000238418579,
763
+ "rewards/chosen": -0.04486004263162613,
764
+ "rewards/margins": 0.07812739908695221,
765
+ "rewards/rejected": -0.12298743426799774,
766
+ "step": 410
767
+ },
768
+ {
769
+ "epoch": 1.99,
770
+ "grad_norm": 11.956370778331848,
771
+ "learning_rate": 2.4397501823713327e-06,
772
+ "log_odds_chosen": 1.8623790740966797,
773
+ "log_odds_ratio": -0.23835237324237823,
774
+ "logits/chosen": -2.491649866104126,
775
+ "logits/rejected": -2.4964215755462646,
776
+ "logps/chosen": -0.36661866307258606,
777
+ "logps/rejected": -1.2916817665100098,
778
+ "loss": 0.4999,
779
+ "nll_loss": 0.4245728850364685,
780
+ "rewards/accuracies": 0.925000011920929,
781
+ "rewards/chosen": -0.03666186332702637,
782
+ "rewards/margins": 0.09250631183385849,
783
+ "rewards/rejected": -0.12916818261146545,
784
+ "step": 420
785
+ },
786
+ {
787
+ "epoch": 2.0,
788
+ "eval_log_odds_chosen": 0.4993869364261627,
789
+ "eval_log_odds_ratio": -0.6011503338813782,
790
+ "eval_logits/chosen": -2.51769757270813,
791
+ "eval_logits/rejected": -2.518415689468384,
792
+ "eval_logps/chosen": -0.8936966061592102,
793
+ "eval_logps/rejected": -1.183638334274292,
794
+ "eval_loss": 0.9812777638435364,
795
+ "eval_nll_loss": 0.940436840057373,
796
+ "eval_rewards/accuracies": 0.6614583134651184,
797
+ "eval_rewards/chosen": -0.08936966210603714,
798
+ "eval_rewards/margins": 0.028994165360927582,
799
+ "eval_rewards/rejected": -0.11836383491754532,
800
+ "eval_runtime": 60.9721,
801
+ "eval_samples_per_second": 12.301,
802
+ "eval_steps_per_second": 0.394,
803
+ "step": 422
804
+ },
805
+ {
806
+ "epoch": 2.04,
807
+ "grad_norm": 11.137133583839702,
808
+ "learning_rate": 2.411214110852061e-06,
809
+ "log_odds_chosen": 2.8829193115234375,
810
+ "log_odds_ratio": -0.11160006374120712,
811
+ "logits/chosen": -2.481720209121704,
812
+ "logits/rejected": -2.5044140815734863,
813
+ "logps/chosen": -0.21169082820415497,
814
+ "logps/rejected": -1.3710792064666748,
815
+ "loss": 0.2521,
816
+ "nll_loss": 0.2048369199037552,
817
+ "rewards/accuracies": 0.9750000238418579,
818
+ "rewards/chosen": -0.021169083192944527,
819
+ "rewards/margins": 0.11593882739543915,
820
+ "rewards/rejected": -0.13710792362689972,
821
+ "step": 430
822
+ },
823
+ {
824
+ "epoch": 2.09,
825
+ "grad_norm": 9.190185309499267,
826
+ "learning_rate": 2.3836564731139807e-06,
827
+ "log_odds_chosen": 3.475831985473633,
828
+ "log_odds_ratio": -0.0667722150683403,
829
+ "logits/chosen": -2.468839645385742,
830
+ "logits/rejected": -2.504293441772461,
831
+ "logps/chosen": -0.13927796483039856,
832
+ "logps/rejected": -1.5609562397003174,
833
+ "loss": 0.1889,
834
+ "nll_loss": 0.142722949385643,
835
+ "rewards/accuracies": 0.987500011920929,
836
+ "rewards/chosen": -0.013927794992923737,
837
+ "rewards/margins": 0.1421678364276886,
838
+ "rewards/rejected": -0.15609565377235413,
839
+ "step": 440
840
+ },
841
+ {
842
+ "epoch": 2.13,
843
+ "grad_norm": 9.564387172693797,
844
+ "learning_rate": 2.357022603955159e-06,
845
+ "log_odds_chosen": 3.917105197906494,
846
+ "log_odds_ratio": -0.037336982786655426,
847
+ "logits/chosen": -2.5166149139404297,
848
+ "logits/rejected": -2.545985698699951,
849
+ "logps/chosen": -0.13963064551353455,
850
+ "logps/rejected": -1.9594411849975586,
851
+ "loss": 0.181,
852
+ "nll_loss": 0.15659931302070618,
853
+ "rewards/accuracies": 1.0,
854
+ "rewards/chosen": -0.013963064178824425,
855
+ "rewards/margins": 0.18198105692863464,
856
+ "rewards/rejected": -0.19594410061836243,
857
+ "step": 450
858
+ },
859
+ {
860
+ "epoch": 2.18,
861
+ "grad_norm": 9.849074304214747,
862
+ "learning_rate": 2.3312620206007847e-06,
863
+ "log_odds_chosen": 3.3254103660583496,
864
+ "log_odds_ratio": -0.06604982912540436,
865
+ "logits/chosen": -2.4885592460632324,
866
+ "logits/rejected": -2.5000545978546143,
867
+ "logps/chosen": -0.17423538863658905,
868
+ "logps/rejected": -1.5009019374847412,
869
+ "loss": 0.1952,
870
+ "nll_loss": 0.21155783534049988,
871
+ "rewards/accuracies": 1.0,
872
+ "rewards/chosen": -0.017423538491129875,
873
+ "rewards/margins": 0.13266664743423462,
874
+ "rewards/rejected": -0.15009018778800964,
875
+ "step": 460
876
+ },
877
+ {
878
+ "epoch": 2.23,
879
+ "grad_norm": 9.877044500517252,
880
+ "learning_rate": 2.3063280200722128e-06,
881
+ "log_odds_chosen": 3.6100239753723145,
882
+ "log_odds_ratio": -0.05039358139038086,
883
+ "logits/chosen": -2.5267927646636963,
884
+ "logits/rejected": -2.5161995887756348,
885
+ "logps/chosen": -0.16191843152046204,
886
+ "logps/rejected": -1.8569713830947876,
887
+ "loss": 0.173,
888
+ "nll_loss": 0.16810880601406097,
889
+ "rewards/accuracies": 1.0,
890
+ "rewards/chosen": -0.016191843897104263,
891
+ "rewards/margins": 0.1695052832365036,
892
+ "rewards/rejected": -0.18569712340831757,
893
+ "step": 470
894
+ },
895
+ {
896
+ "epoch": 2.27,
897
+ "grad_norm": 8.163874170175868,
898
+ "learning_rate": 2.2821773229381924e-06,
899
+ "log_odds_chosen": 3.612773895263672,
900
+ "log_odds_ratio": -0.05763618275523186,
901
+ "logits/chosen": -2.5359530448913574,
902
+ "logits/rejected": -2.539139747619629,
903
+ "logps/chosen": -0.17719587683677673,
904
+ "logps/rejected": -1.8914750814437866,
905
+ "loss": 0.1808,
906
+ "nll_loss": 0.17015354335308075,
907
+ "rewards/accuracies": 0.987500011920929,
908
+ "rewards/chosen": -0.017719587311148643,
909
+ "rewards/margins": 0.171427920460701,
910
+ "rewards/rejected": -0.18914751708507538,
911
+ "step": 480
912
+ },
913
+ {
914
+ "epoch": 2.32,
915
+ "grad_norm": 9.330097236244276,
916
+ "learning_rate": 2.2587697572631284e-06,
917
+ "log_odds_chosen": 3.656383514404297,
918
+ "log_odds_ratio": -0.05575472116470337,
919
+ "logits/chosen": -2.450357437133789,
920
+ "logits/rejected": -2.4894728660583496,
921
+ "logps/chosen": -0.12832140922546387,
922
+ "logps/rejected": -1.6764119863510132,
923
+ "loss": 0.1743,
924
+ "nll_loss": 0.1293896734714508,
925
+ "rewards/accuracies": 0.987500011920929,
926
+ "rewards/chosen": -0.012832140550017357,
927
+ "rewards/margins": 0.15480905771255493,
928
+ "rewards/rejected": -0.16764120757579803,
929
+ "step": 490
930
+ },
931
+ {
932
+ "epoch": 2.37,
933
+ "grad_norm": 8.638065176436028,
934
+ "learning_rate": 2.23606797749979e-06,
935
+ "log_odds_chosen": 3.5653939247131348,
936
+ "log_odds_ratio": -0.04521235078573227,
937
+ "logits/chosen": -2.4729416370391846,
938
+ "logits/rejected": -2.4972083568573,
939
+ "logps/chosen": -0.14347168803215027,
940
+ "logps/rejected": -1.709639549255371,
941
+ "loss": 0.1687,
942
+ "nll_loss": 0.15304037928581238,
943
+ "rewards/accuracies": 1.0,
944
+ "rewards/chosen": -0.014347168616950512,
945
+ "rewards/margins": 0.15661677718162537,
946
+ "rewards/rejected": -0.17096397280693054,
947
+ "step": 500
948
+ },
949
+ {
950
+ "epoch": 2.42,
951
+ "grad_norm": 9.821968133030161,
952
+ "learning_rate": 2.2140372138502386e-06,
953
+ "log_odds_chosen": 3.384936809539795,
954
+ "log_odds_ratio": -0.06072957068681717,
955
+ "logits/chosen": -2.477996349334717,
956
+ "logits/rejected": -2.482419729232788,
957
+ "logps/chosen": -0.17441074550151825,
958
+ "logps/rejected": -1.6888253688812256,
959
+ "loss": 0.191,
960
+ "nll_loss": 0.20637556910514832,
961
+ "rewards/accuracies": 1.0,
962
+ "rewards/chosen": -0.017441075295209885,
963
+ "rewards/margins": 0.15144145488739014,
964
+ "rewards/rejected": -0.16888253390789032,
965
+ "step": 510
966
+ },
967
+ {
968
+ "epoch": 2.46,
969
+ "grad_norm": 9.230367667532265,
970
+ "learning_rate": 2.1926450482675734e-06,
971
+ "log_odds_chosen": 3.698854446411133,
972
+ "log_odds_ratio": -0.04557035490870476,
973
+ "logits/chosen": -2.493439197540283,
974
+ "logits/rejected": -2.507626533508301,
975
+ "logps/chosen": -0.15794190764427185,
976
+ "logps/rejected": -1.836679458618164,
977
+ "loss": 0.1823,
978
+ "nll_loss": 0.1679120659828186,
979
+ "rewards/accuracies": 1.0,
980
+ "rewards/chosen": -0.015794191509485245,
981
+ "rewards/margins": 0.16787376999855042,
982
+ "rewards/rejected": -0.18366795778274536,
983
+ "step": 520
984
+ },
985
+ {
986
+ "epoch": 2.51,
987
+ "grad_norm": 9.220963630868974,
988
+ "learning_rate": 2.1718612138153473e-06,
989
+ "log_odds_chosen": 3.2164981365203857,
990
+ "log_odds_ratio": -0.08045772463083267,
991
+ "logits/chosen": -2.4849820137023926,
992
+ "logits/rejected": -2.513134717941284,
993
+ "logps/chosen": -0.20204570889472961,
994
+ "logps/rejected": -1.6124742031097412,
995
+ "loss": 0.1955,
996
+ "nll_loss": 0.1980821043252945,
997
+ "rewards/accuracies": 0.987500011920929,
998
+ "rewards/chosen": -0.0202045738697052,
999
+ "rewards/margins": 0.14104285836219788,
1000
+ "rewards/rejected": -0.16124743223190308,
1001
+ "step": 530
1002
+ },
1003
+ {
1004
+ "epoch": 2.56,
1005
+ "grad_norm": 10.33252022252946,
1006
+ "learning_rate": 2.151657414559676e-06,
1007
+ "log_odds_chosen": 3.355163097381592,
1008
+ "log_odds_ratio": -0.05846261978149414,
1009
+ "logits/chosen": -2.4703683853149414,
1010
+ "logits/rejected": -2.4815986156463623,
1011
+ "logps/chosen": -0.16165122389793396,
1012
+ "logps/rejected": -1.595573902130127,
1013
+ "loss": 0.1944,
1014
+ "nll_loss": 0.1843605637550354,
1015
+ "rewards/accuracies": 1.0,
1016
+ "rewards/chosen": -0.016165124252438545,
1017
+ "rewards/margins": 0.14339227974414825,
1018
+ "rewards/rejected": -0.15955740213394165,
1019
+ "step": 540
1020
+ },
1021
+ {
1022
+ "epoch": 2.61,
1023
+ "grad_norm": 10.384732401123731,
1024
+ "learning_rate": 2.132007163556104e-06,
1025
+ "log_odds_chosen": 3.7364859580993652,
1026
+ "log_odds_ratio": -0.050393521785736084,
1027
+ "logits/chosen": -2.488259792327881,
1028
+ "logits/rejected": -2.49784517288208,
1029
+ "logps/chosen": -0.1458165943622589,
1030
+ "logps/rejected": -1.7980148792266846,
1031
+ "loss": 0.1656,
1032
+ "nll_loss": 0.16633237898349762,
1033
+ "rewards/accuracies": 0.987500011920929,
1034
+ "rewards/chosen": -0.014581659808754921,
1035
+ "rewards/margins": 0.16521981358528137,
1036
+ "rewards/rejected": -0.17980147898197174,
1037
+ "step": 550
1038
+ },
1039
+ {
1040
+ "epoch": 2.65,
1041
+ "grad_norm": 8.88379907455603,
1042
+ "learning_rate": 2.1128856368212917e-06,
1043
+ "log_odds_chosen": 3.3721110820770264,
1044
+ "log_odds_ratio": -0.05570857599377632,
1045
+ "logits/chosen": -2.4560656547546387,
1046
+ "logits/rejected": -2.459944725036621,
1047
+ "logps/chosen": -0.14691922068595886,
1048
+ "logps/rejected": -1.4988293647766113,
1049
+ "loss": 0.1941,
1050
+ "nll_loss": 0.16041871905326843,
1051
+ "rewards/accuracies": 1.0,
1052
+ "rewards/chosen": -0.014691921882331371,
1053
+ "rewards/margins": 0.13519100844860077,
1054
+ "rewards/rejected": -0.1498829424381256,
1055
+ "step": 560
1056
+ },
1057
+ {
1058
+ "epoch": 2.7,
1059
+ "grad_norm": 8.564539697443587,
1060
+ "learning_rate": 2.0942695414584777e-06,
1061
+ "log_odds_chosen": 3.540961503982544,
1062
+ "log_odds_ratio": -0.050423987209796906,
1063
+ "logits/chosen": -2.4706387519836426,
1064
+ "logits/rejected": -2.478842258453369,
1065
+ "logps/chosen": -0.15486404299736023,
1066
+ "logps/rejected": -1.6839189529418945,
1067
+ "loss": 0.1827,
1068
+ "nll_loss": 0.16179664433002472,
1069
+ "rewards/accuracies": 1.0,
1070
+ "rewards/chosen": -0.015486404299736023,
1071
+ "rewards/margins": 0.15290549397468567,
1072
+ "rewards/rejected": -0.1683918982744217,
1073
+ "step": 570
1074
+ },
1075
+ {
1076
+ "epoch": 2.75,
1077
+ "grad_norm": 8.776123151000377,
1078
+ "learning_rate": 2.0761369963434992e-06,
1079
+ "log_odds_chosen": 3.2635269165039062,
1080
+ "log_odds_ratio": -0.07641835510730743,
1081
+ "logits/chosen": -2.49845552444458,
1082
+ "logits/rejected": -2.5065455436706543,
1083
+ "logps/chosen": -0.21721407771110535,
1084
+ "logps/rejected": -1.628244161605835,
1085
+ "loss": 0.2085,
1086
+ "nll_loss": 0.26145899295806885,
1087
+ "rewards/accuracies": 0.987500011920929,
1088
+ "rewards/chosen": -0.021721405908465385,
1089
+ "rewards/margins": 0.14110301434993744,
1090
+ "rewards/rejected": -0.16282442212104797,
1091
+ "step": 580
1092
+ },
1093
+ {
1094
+ "epoch": 2.8,
1095
+ "grad_norm": 8.360109402740795,
1096
+ "learning_rate": 2.058467423981546e-06,
1097
+ "log_odds_chosen": 3.5911190509796143,
1098
+ "log_odds_ratio": -0.04746793955564499,
1099
+ "logits/chosen": -2.4481544494628906,
1100
+ "logits/rejected": -2.4686474800109863,
1101
+ "logps/chosen": -0.18100793659687042,
1102
+ "logps/rejected": -1.7877342700958252,
1103
+ "loss": 0.1973,
1104
+ "nll_loss": 0.19511719048023224,
1105
+ "rewards/accuracies": 1.0,
1106
+ "rewards/chosen": -0.018100792542099953,
1107
+ "rewards/margins": 0.1606726348400116,
1108
+ "rewards/rejected": -0.178773432970047,
1109
+ "step": 590
1110
+ },
1111
+ {
1112
+ "epoch": 2.84,
1113
+ "grad_norm": 8.230484923418327,
1114
+ "learning_rate": 2.0412414523193154e-06,
1115
+ "log_odds_chosen": 3.52643084526062,
1116
+ "log_odds_ratio": -0.05582552030682564,
1117
+ "logits/chosen": -2.43485689163208,
1118
+ "logits/rejected": -2.452052116394043,
1119
+ "logps/chosen": -0.16544193029403687,
1120
+ "logps/rejected": -1.6863620281219482,
1121
+ "loss": 0.1863,
1122
+ "nll_loss": 0.17863860726356506,
1123
+ "rewards/accuracies": 1.0,
1124
+ "rewards/chosen": -0.016544191166758537,
1125
+ "rewards/margins": 0.15209202468395233,
1126
+ "rewards/rejected": -0.16863620281219482,
1127
+ "step": 600
1128
+ },
1129
+ {
1130
+ "epoch": 2.89,
1131
+ "grad_norm": 9.657714694141937,
1132
+ "learning_rate": 2.0244408254472904e-06,
1133
+ "log_odds_chosen": 3.441495418548584,
1134
+ "log_odds_ratio": -0.050580836832523346,
1135
+ "logits/chosen": -2.3969621658325195,
1136
+ "logits/rejected": -2.4254117012023926,
1137
+ "logps/chosen": -0.1414327472448349,
1138
+ "logps/rejected": -1.5730435848236084,
1139
+ "loss": 0.1825,
1140
+ "nll_loss": 0.16234278678894043,
1141
+ "rewards/accuracies": 1.0,
1142
+ "rewards/chosen": -0.01414327509701252,
1143
+ "rewards/margins": 0.1431610882282257,
1144
+ "rewards/rejected": -0.15730436146259308,
1145
+ "step": 610
1146
+ },
1147
+ {
1148
+ "epoch": 2.94,
1149
+ "grad_norm": 9.418203083185244,
1150
+ "learning_rate": 2.0080483222562476e-06,
1151
+ "log_odds_chosen": 3.510298252105713,
1152
+ "log_odds_ratio": -0.04852527379989624,
1153
+ "logits/chosen": -2.365495443344116,
1154
+ "logits/rejected": -2.3816208839416504,
1155
+ "logps/chosen": -0.14439718425273895,
1156
+ "logps/rejected": -1.5629642009735107,
1157
+ "loss": 0.1816,
1158
+ "nll_loss": 0.1693393439054489,
1159
+ "rewards/accuracies": 1.0,
1160
+ "rewards/chosen": -0.014439716935157776,
1161
+ "rewards/margins": 0.14185671508312225,
1162
+ "rewards/rejected": -0.15629643201828003,
1163
+ "step": 620
1164
+ },
1165
+ {
1166
+ "epoch": 2.99,
1167
+ "grad_norm": 9.905181462463698,
1168
+ "learning_rate": 1.9920476822239895e-06,
1169
+ "log_odds_chosen": 3.6201272010803223,
1170
+ "log_odds_ratio": -0.04453429579734802,
1171
+ "logits/chosen": -2.415280818939209,
1172
+ "logits/rejected": -2.403134822845459,
1173
+ "logps/chosen": -0.1919378936290741,
1174
+ "logps/rejected": -1.9056262969970703,
1175
+ "loss": 0.1899,
1176
+ "nll_loss": 0.1791400909423828,
1177
+ "rewards/accuracies": 1.0,
1178
+ "rewards/chosen": -0.01919379085302353,
1179
+ "rewards/margins": 0.17136886715888977,
1180
+ "rewards/rejected": -0.1905626356601715,
1181
+ "step": 630
1182
+ },
1183
+ {
1184
+ "epoch": 3.0,
1185
+ "eval_log_odds_chosen": 0.539490818977356,
1186
+ "eval_log_odds_ratio": -0.6108235716819763,
1187
+ "eval_logits/chosen": -2.4178202152252197,
1188
+ "eval_logits/rejected": -2.420111894607544,
1189
+ "eval_logps/chosen": -1.1129909753799438,
1190
+ "eval_logps/rejected": -1.4606622457504272,
1191
+ "eval_loss": 1.2152478694915771,
1192
+ "eval_nll_loss": 1.1710882186889648,
1193
+ "eval_rewards/accuracies": 0.65625,
1194
+ "eval_rewards/chosen": -0.11129910498857498,
1195
+ "eval_rewards/margins": 0.03476712107658386,
1196
+ "eval_rewards/rejected": -0.14606623351573944,
1197
+ "eval_runtime": 60.9563,
1198
+ "eval_samples_per_second": 12.304,
1199
+ "eval_steps_per_second": 0.394,
1200
+ "step": 633
1201
+ },
1202
+ {
1203
+ "epoch": 3.0,
1204
+ "step": 633,
1205
+ "total_flos": 0.0,
1206
+ "train_loss": 0.5526923979816467,
1207
+ "train_runtime": 6310.4087,
1208
+ "train_samples_per_second": 3.209,
1209
+ "train_steps_per_second": 0.1
1210
+ }
1211
+ ],
1212
+ "logging_steps": 10,
1213
+ "max_steps": 633,
1214
+ "num_input_tokens_seen": 0,
1215
+ "num_train_epochs": 3,
1216
+ "save_steps": 500,
1217
+ "total_flos": 0.0,
1218
+ "train_batch_size": 8,
1219
+ "trial_name": null,
1220
+ "trial_params": null
1221
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acaf9c7a2cb692271963d3f4fb8c92c54b23a09fd264a214d4e80f2eba74fd39
3
+ size 6456