xiaoxiaolin commited on
Commit
11d14d8
1 Parent(s): dacf41f

Upload folder using huggingface_hub

Browse files
config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/data1/ranw/ckpt/webshop_sft_model/qagent-vanilla-Q-BON-iter1-epoch3-Llama-2-7b-chat-hf-webshop",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 4096,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 11008,
14
+ "max_position_embeddings": 4096,
15
+ "mlp_bias": false,
16
+ "model_type": "llama",
17
+ "num_attention_heads": 32,
18
+ "num_hidden_layers": 32,
19
+ "num_key_value_heads": 32,
20
+ "pretraining_tp": 1,
21
+ "rms_norm_eps": 1e-05,
22
+ "rope_scaling": null,
23
+ "rope_theta": 10000.0,
24
+ "tie_word_embeddings": false,
25
+ "torch_dtype": "float32",
26
+ "transformers_version": "4.44.2",
27
+ "use_cache": true,
28
+ "vocab_size": 32000
29
+ }
generation_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 1,
3
+ "do_sample": true,
4
+ "eos_token_id": 2,
5
+ "max_length": 4096,
6
+ "pad_token_id": 0,
7
+ "temperature": 0.6,
8
+ "top_p": 0.9,
9
+ "transformers_version": "4.44.2"
10
+ }
model-00001-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a2311ffb89e88b710c086082830ffc7cef36875fce081257b9ba3688b1d9b35
3
+ size 4840396416
model-00002-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f81605c423cf8193e037ef3c85e3c45ad777737a57ce1f9e1dddd6f5abce00f
3
+ size 4857206856
model-00003-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fee72e2690f06bda9c81c385cba698e2112f7d6556adf6a0742b549bb4b362b
3
+ size 4857206904
model-00004-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3dd2125fe7441d70bdc85086d5ca4fc179c5dcecb03abdbd65afdb11694175e2
3
+ size 4857206904
model-00005-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b4724097fb6c0446a844f1c4035d49457539e0a62829d6dc2535ea3971e2816
3
+ size 4857206904
model-00006-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cfab7f1f27da3ea40d72c1e1876d58ab962b92dbd91c6c84195da6ce7a06f30
3
+ size 2684472112
model.safetensors.index.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 26953662464
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00006-of-00006.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00006.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00006.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00006.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00006.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00006.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00006.safetensors",
13
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
14
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
15
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
16
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
17
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00006.safetensors",
18
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00006.safetensors",
19
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00006.safetensors",
20
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00006.safetensors",
21
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00006.safetensors",
22
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
23
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
24
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
25
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
26
+ "model.layers.10.input_layernorm.weight": "model-00002-of-00006.safetensors",
27
+ "model.layers.10.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
28
+ "model.layers.10.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
29
+ "model.layers.10.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
30
+ "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
31
+ "model.layers.10.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
32
+ "model.layers.10.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
33
+ "model.layers.10.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
34
+ "model.layers.10.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
35
+ "model.layers.11.input_layernorm.weight": "model-00003-of-00006.safetensors",
36
+ "model.layers.11.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
37
+ "model.layers.11.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
38
+ "model.layers.11.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
39
+ "model.layers.11.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
40
+ "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
41
+ "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
42
+ "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
43
+ "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
44
+ "model.layers.12.input_layernorm.weight": "model-00003-of-00006.safetensors",
45
+ "model.layers.12.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
46
+ "model.layers.12.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
47
+ "model.layers.12.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
48
+ "model.layers.12.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
49
+ "model.layers.12.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
50
+ "model.layers.12.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
51
+ "model.layers.12.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
52
+ "model.layers.12.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
53
+ "model.layers.13.input_layernorm.weight": "model-00003-of-00006.safetensors",
54
+ "model.layers.13.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
55
+ "model.layers.13.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
56
+ "model.layers.13.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
57
+ "model.layers.13.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
58
+ "model.layers.13.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
59
+ "model.layers.13.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
60
+ "model.layers.13.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
61
+ "model.layers.13.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
62
+ "model.layers.14.input_layernorm.weight": "model-00003-of-00006.safetensors",
63
+ "model.layers.14.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
64
+ "model.layers.14.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
65
+ "model.layers.14.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
66
+ "model.layers.14.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
67
+ "model.layers.14.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
68
+ "model.layers.14.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
69
+ "model.layers.14.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
70
+ "model.layers.14.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
71
+ "model.layers.15.input_layernorm.weight": "model-00003-of-00006.safetensors",
72
+ "model.layers.15.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
73
+ "model.layers.15.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
74
+ "model.layers.15.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
75
+ "model.layers.15.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
76
+ "model.layers.15.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
77
+ "model.layers.15.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
78
+ "model.layers.15.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
79
+ "model.layers.15.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
80
+ "model.layers.16.input_layernorm.weight": "model-00003-of-00006.safetensors",
81
+ "model.layers.16.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
82
+ "model.layers.16.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
83
+ "model.layers.16.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
84
+ "model.layers.16.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
85
+ "model.layers.16.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
86
+ "model.layers.16.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
87
+ "model.layers.16.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
88
+ "model.layers.16.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
89
+ "model.layers.17.input_layernorm.weight": "model-00004-of-00006.safetensors",
90
+ "model.layers.17.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
91
+ "model.layers.17.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
92
+ "model.layers.17.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
93
+ "model.layers.17.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
94
+ "model.layers.17.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
95
+ "model.layers.17.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
96
+ "model.layers.17.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
97
+ "model.layers.17.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
98
+ "model.layers.18.input_layernorm.weight": "model-00004-of-00006.safetensors",
99
+ "model.layers.18.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
100
+ "model.layers.18.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
101
+ "model.layers.18.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
102
+ "model.layers.18.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
103
+ "model.layers.18.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
104
+ "model.layers.18.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
105
+ "model.layers.18.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
106
+ "model.layers.18.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
107
+ "model.layers.19.input_layernorm.weight": "model-00004-of-00006.safetensors",
108
+ "model.layers.19.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
109
+ "model.layers.19.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
110
+ "model.layers.19.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
111
+ "model.layers.19.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
112
+ "model.layers.19.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
113
+ "model.layers.19.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
114
+ "model.layers.19.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
115
+ "model.layers.19.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
116
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00006.safetensors",
117
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00006.safetensors",
118
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00006.safetensors",
119
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00006.safetensors",
120
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00006.safetensors",
121
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
122
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
123
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
124
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
125
+ "model.layers.20.input_layernorm.weight": "model-00004-of-00006.safetensors",
126
+ "model.layers.20.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
127
+ "model.layers.20.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
128
+ "model.layers.20.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
129
+ "model.layers.20.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
130
+ "model.layers.20.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
131
+ "model.layers.20.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
132
+ "model.layers.20.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
133
+ "model.layers.20.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
134
+ "model.layers.21.input_layernorm.weight": "model-00004-of-00006.safetensors",
135
+ "model.layers.21.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
136
+ "model.layers.21.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
137
+ "model.layers.21.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
138
+ "model.layers.21.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
139
+ "model.layers.21.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
140
+ "model.layers.21.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
141
+ "model.layers.21.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
142
+ "model.layers.21.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
143
+ "model.layers.22.input_layernorm.weight": "model-00004-of-00006.safetensors",
144
+ "model.layers.22.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
145
+ "model.layers.22.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
146
+ "model.layers.22.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
147
+ "model.layers.22.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
148
+ "model.layers.22.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
149
+ "model.layers.22.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
150
+ "model.layers.22.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
151
+ "model.layers.22.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
152
+ "model.layers.23.input_layernorm.weight": "model-00005-of-00006.safetensors",
153
+ "model.layers.23.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
154
+ "model.layers.23.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
155
+ "model.layers.23.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
156
+ "model.layers.23.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
157
+ "model.layers.23.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
158
+ "model.layers.23.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
159
+ "model.layers.23.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
160
+ "model.layers.23.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
161
+ "model.layers.24.input_layernorm.weight": "model-00005-of-00006.safetensors",
162
+ "model.layers.24.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
163
+ "model.layers.24.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
164
+ "model.layers.24.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
165
+ "model.layers.24.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
166
+ "model.layers.24.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
167
+ "model.layers.24.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
168
+ "model.layers.24.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
169
+ "model.layers.24.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
170
+ "model.layers.25.input_layernorm.weight": "model-00005-of-00006.safetensors",
171
+ "model.layers.25.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
172
+ "model.layers.25.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
173
+ "model.layers.25.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
174
+ "model.layers.25.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
175
+ "model.layers.25.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
176
+ "model.layers.25.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
177
+ "model.layers.25.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
178
+ "model.layers.25.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
179
+ "model.layers.26.input_layernorm.weight": "model-00005-of-00006.safetensors",
180
+ "model.layers.26.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
181
+ "model.layers.26.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
182
+ "model.layers.26.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
183
+ "model.layers.26.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
184
+ "model.layers.26.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
185
+ "model.layers.26.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
186
+ "model.layers.26.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
187
+ "model.layers.26.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
188
+ "model.layers.27.input_layernorm.weight": "model-00005-of-00006.safetensors",
189
+ "model.layers.27.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
190
+ "model.layers.27.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
191
+ "model.layers.27.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
192
+ "model.layers.27.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
193
+ "model.layers.27.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
194
+ "model.layers.27.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
195
+ "model.layers.27.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
196
+ "model.layers.27.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
197
+ "model.layers.28.input_layernorm.weight": "model-00005-of-00006.safetensors",
198
+ "model.layers.28.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
199
+ "model.layers.28.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
200
+ "model.layers.28.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
201
+ "model.layers.28.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
202
+ "model.layers.28.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
203
+ "model.layers.28.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
204
+ "model.layers.28.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
205
+ "model.layers.28.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
206
+ "model.layers.29.input_layernorm.weight": "model-00006-of-00006.safetensors",
207
+ "model.layers.29.mlp.down_proj.weight": "model-00006-of-00006.safetensors",
208
+ "model.layers.29.mlp.gate_proj.weight": "model-00006-of-00006.safetensors",
209
+ "model.layers.29.mlp.up_proj.weight": "model-00006-of-00006.safetensors",
210
+ "model.layers.29.post_attention_layernorm.weight": "model-00006-of-00006.safetensors",
211
+ "model.layers.29.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
212
+ "model.layers.29.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
213
+ "model.layers.29.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
214
+ "model.layers.29.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
215
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00006.safetensors",
216
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00006.safetensors",
217
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00006.safetensors",
218
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00006.safetensors",
219
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00006.safetensors",
220
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
221
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
222
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
223
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
224
+ "model.layers.30.input_layernorm.weight": "model-00006-of-00006.safetensors",
225
+ "model.layers.30.mlp.down_proj.weight": "model-00006-of-00006.safetensors",
226
+ "model.layers.30.mlp.gate_proj.weight": "model-00006-of-00006.safetensors",
227
+ "model.layers.30.mlp.up_proj.weight": "model-00006-of-00006.safetensors",
228
+ "model.layers.30.post_attention_layernorm.weight": "model-00006-of-00006.safetensors",
229
+ "model.layers.30.self_attn.k_proj.weight": "model-00006-of-00006.safetensors",
230
+ "model.layers.30.self_attn.o_proj.weight": "model-00006-of-00006.safetensors",
231
+ "model.layers.30.self_attn.q_proj.weight": "model-00006-of-00006.safetensors",
232
+ "model.layers.30.self_attn.v_proj.weight": "model-00006-of-00006.safetensors",
233
+ "model.layers.31.input_layernorm.weight": "model-00006-of-00006.safetensors",
234
+ "model.layers.31.mlp.down_proj.weight": "model-00006-of-00006.safetensors",
235
+ "model.layers.31.mlp.gate_proj.weight": "model-00006-of-00006.safetensors",
236
+ "model.layers.31.mlp.up_proj.weight": "model-00006-of-00006.safetensors",
237
+ "model.layers.31.post_attention_layernorm.weight": "model-00006-of-00006.safetensors",
238
+ "model.layers.31.self_attn.k_proj.weight": "model-00006-of-00006.safetensors",
239
+ "model.layers.31.self_attn.o_proj.weight": "model-00006-of-00006.safetensors",
240
+ "model.layers.31.self_attn.q_proj.weight": "model-00006-of-00006.safetensors",
241
+ "model.layers.31.self_attn.v_proj.weight": "model-00006-of-00006.safetensors",
242
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00006.safetensors",
243
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00006.safetensors",
244
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00006.safetensors",
245
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00006.safetensors",
246
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00006.safetensors",
247
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
248
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
249
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
250
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
251
+ "model.layers.5.input_layernorm.weight": "model-00002-of-00006.safetensors",
252
+ "model.layers.5.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
253
+ "model.layers.5.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
254
+ "model.layers.5.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
255
+ "model.layers.5.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
256
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
257
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
258
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
259
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
260
+ "model.layers.6.input_layernorm.weight": "model-00002-of-00006.safetensors",
261
+ "model.layers.6.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
262
+ "model.layers.6.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
263
+ "model.layers.6.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
264
+ "model.layers.6.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
265
+ "model.layers.6.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
266
+ "model.layers.6.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
267
+ "model.layers.6.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
268
+ "model.layers.6.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
269
+ "model.layers.7.input_layernorm.weight": "model-00002-of-00006.safetensors",
270
+ "model.layers.7.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
271
+ "model.layers.7.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
272
+ "model.layers.7.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
273
+ "model.layers.7.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
274
+ "model.layers.7.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
275
+ "model.layers.7.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
276
+ "model.layers.7.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
277
+ "model.layers.7.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
278
+ "model.layers.8.input_layernorm.weight": "model-00002-of-00006.safetensors",
279
+ "model.layers.8.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
280
+ "model.layers.8.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
281
+ "model.layers.8.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
282
+ "model.layers.8.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
283
+ "model.layers.8.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
284
+ "model.layers.8.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
285
+ "model.layers.8.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
286
+ "model.layers.8.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
287
+ "model.layers.9.input_layernorm.weight": "model-00002-of-00006.safetensors",
288
+ "model.layers.9.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
289
+ "model.layers.9.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
290
+ "model.layers.9.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
291
+ "model.layers.9.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
292
+ "model.layers.9.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
293
+ "model.layers.9.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
294
+ "model.layers.9.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
295
+ "model.layers.9.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
296
+ "model.norm.weight": "model-00006-of-00006.safetensors"
297
+ }
298
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": true,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "bos_token": "<s>",
32
+ "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% set system_message = false %}{% endif %}{% for message in loop_messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if loop.index0 == 0 and system_message != false %}{% set content = '<<SYS>>\\n' + system_message + '\\n<</SYS>>\\n\\n' + message['content'] %}{% else %}{% set content = message['content'] %}{% endif %}{% if message['role'] == 'user' %}{{ bos_token + '[INST] ' + content.strip() + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + content.strip() + ' ' + eos_token }}{% endif %}{% endfor %}",
33
+ "clean_up_tokenization_spaces": false,
34
+ "eos_token": "</s>",
35
+ "legacy": false,
36
+ "model_max_length": 4096,
37
+ "pad_token": "<unk>",
38
+ "padding_side": "right",
39
+ "sp_model_kwargs": {},
40
+ "spaces_between_special_tokens": false,
41
+ "tokenizer_class": "LlamaTokenizer",
42
+ "unk_token": "<unk>",
43
+ "use_default_system_prompt": false
44
+ }
trainer_state.json ADDED
@@ -0,0 +1,1722 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 2.992,
5
+ "eval_steps": 500,
6
+ "global_step": 561,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.02666666666666667,
13
+ "grad_norm": 18.322052001953125,
14
+ "learning_rate": 8.771929824561403e-08,
15
+ "logits/chosen": -0.8889046907424927,
16
+ "logits/rejected": -0.8736551403999329,
17
+ "logps/chosen": -81.54634094238281,
18
+ "logps/rejected": -75.81903839111328,
19
+ "loss": 0.6912,
20
+ "rewards/accuracies": 0.3499999940395355,
21
+ "rewards/chosen": 0.0006732392939738929,
22
+ "rewards/margins": 0.005079272203147411,
23
+ "rewards/rejected": -0.004406032618135214,
24
+ "step": 5
25
+ },
26
+ {
27
+ "epoch": 0.05333333333333334,
28
+ "grad_norm": 19.94486427307129,
29
+ "learning_rate": 1.7543859649122805e-07,
30
+ "logits/chosen": -0.9008264541625977,
31
+ "logits/rejected": -0.8997122645378113,
32
+ "logps/chosen": -56.543792724609375,
33
+ "logps/rejected": -70.56910705566406,
34
+ "loss": 0.6741,
35
+ "rewards/accuracies": 0.75,
36
+ "rewards/chosen": 0.040357910096645355,
37
+ "rewards/margins": 0.049730073660612106,
38
+ "rewards/rejected": -0.009372168220579624,
39
+ "step": 10
40
+ },
41
+ {
42
+ "epoch": 0.08,
43
+ "grad_norm": 16.07955551147461,
44
+ "learning_rate": 2.631578947368421e-07,
45
+ "logits/chosen": -0.9306875467300415,
46
+ "logits/rejected": -0.9334227442741394,
47
+ "logps/chosen": -70.51209259033203,
48
+ "logps/rejected": -84.35716247558594,
49
+ "loss": 0.6422,
50
+ "rewards/accuracies": 0.75,
51
+ "rewards/chosen": 0.06031479313969612,
52
+ "rewards/margins": 0.09791518747806549,
53
+ "rewards/rejected": -0.03760039433836937,
54
+ "step": 15
55
+ },
56
+ {
57
+ "epoch": 0.10666666666666667,
58
+ "grad_norm": 16.454225540161133,
59
+ "learning_rate": 3.508771929824561e-07,
60
+ "logits/chosen": -0.8976303339004517,
61
+ "logits/rejected": -0.8699060678482056,
62
+ "logps/chosen": -72.55310821533203,
63
+ "logps/rejected": -87.03462219238281,
64
+ "loss": 0.6069,
65
+ "rewards/accuracies": 0.6499999761581421,
66
+ "rewards/chosen": 0.05961589142680168,
67
+ "rewards/margins": 0.28127092123031616,
68
+ "rewards/rejected": -0.2216549664735794,
69
+ "step": 20
70
+ },
71
+ {
72
+ "epoch": 0.13333333333333333,
73
+ "grad_norm": 14.617012977600098,
74
+ "learning_rate": 4.3859649122807013e-07,
75
+ "logits/chosen": -0.9105884432792664,
76
+ "logits/rejected": -0.9121261835098267,
77
+ "logps/chosen": -81.52780151367188,
78
+ "logps/rejected": -91.52813720703125,
79
+ "loss": 0.6019,
80
+ "rewards/accuracies": 0.699999988079071,
81
+ "rewards/chosen": 0.12029141187667847,
82
+ "rewards/margins": 0.3524962365627289,
83
+ "rewards/rejected": -0.2322048395872116,
84
+ "step": 25
85
+ },
86
+ {
87
+ "epoch": 0.16,
88
+ "grad_norm": 13.667928695678711,
89
+ "learning_rate": 5.263157894736842e-07,
90
+ "logits/chosen": -0.8984580039978027,
91
+ "logits/rejected": -0.9042614102363586,
92
+ "logps/chosen": -74.46806335449219,
93
+ "logps/rejected": -76.2645034790039,
94
+ "loss": 0.5363,
95
+ "rewards/accuracies": 0.8500000238418579,
96
+ "rewards/chosen": 0.09522397816181183,
97
+ "rewards/margins": 0.6430191993713379,
98
+ "rewards/rejected": -0.547795295715332,
99
+ "step": 30
100
+ },
101
+ {
102
+ "epoch": 0.18666666666666668,
103
+ "grad_norm": 16.28107452392578,
104
+ "learning_rate": 6.140350877192982e-07,
105
+ "logits/chosen": -0.8978121876716614,
106
+ "logits/rejected": -0.8999565243721008,
107
+ "logps/chosen": -82.0078353881836,
108
+ "logps/rejected": -80.48029327392578,
109
+ "loss": 0.5021,
110
+ "rewards/accuracies": 0.6000000238418579,
111
+ "rewards/chosen": 0.27644580602645874,
112
+ "rewards/margins": 0.4465056359767914,
113
+ "rewards/rejected": -0.17005985975265503,
114
+ "step": 35
115
+ },
116
+ {
117
+ "epoch": 0.21333333333333335,
118
+ "grad_norm": 16.186918258666992,
119
+ "learning_rate": 7.017543859649122e-07,
120
+ "logits/chosen": -0.8643107414245605,
121
+ "logits/rejected": -0.8585313558578491,
122
+ "logps/chosen": -103.23567962646484,
123
+ "logps/rejected": -90.96733856201172,
124
+ "loss": 0.4986,
125
+ "rewards/accuracies": 0.6499999761581421,
126
+ "rewards/chosen": -0.10163059085607529,
127
+ "rewards/margins": 0.47456789016723633,
128
+ "rewards/rejected": -0.5761984586715698,
129
+ "step": 40
130
+ },
131
+ {
132
+ "epoch": 0.24,
133
+ "grad_norm": 14.479534149169922,
134
+ "learning_rate": 7.894736842105263e-07,
135
+ "logits/chosen": -0.9103773236274719,
136
+ "logits/rejected": -0.9051542282104492,
137
+ "logps/chosen": -67.92506408691406,
138
+ "logps/rejected": -77.37336730957031,
139
+ "loss": 0.4875,
140
+ "rewards/accuracies": 0.699999988079071,
141
+ "rewards/chosen": 0.29286879301071167,
142
+ "rewards/margins": 0.8011099100112915,
143
+ "rewards/rejected": -0.5082410573959351,
144
+ "step": 45
145
+ },
146
+ {
147
+ "epoch": 0.26666666666666666,
148
+ "grad_norm": 12.070276260375977,
149
+ "learning_rate": 8.771929824561403e-07,
150
+ "logits/chosen": -0.8517910242080688,
151
+ "logits/rejected": -0.8563734889030457,
152
+ "logps/chosen": -92.3311767578125,
153
+ "logps/rejected": -102.72621154785156,
154
+ "loss": 0.4479,
155
+ "rewards/accuracies": 0.6499999761581421,
156
+ "rewards/chosen": -0.06940027326345444,
157
+ "rewards/margins": 1.23981511592865,
158
+ "rewards/rejected": -1.3092153072357178,
159
+ "step": 50
160
+ },
161
+ {
162
+ "epoch": 0.29333333333333333,
163
+ "grad_norm": 15.573699951171875,
164
+ "learning_rate": 9.649122807017545e-07,
165
+ "logits/chosen": -0.9419733285903931,
166
+ "logits/rejected": -0.9363008737564087,
167
+ "logps/chosen": -76.72134399414062,
168
+ "logps/rejected": -82.520751953125,
169
+ "loss": 0.527,
170
+ "rewards/accuracies": 0.699999988079071,
171
+ "rewards/chosen": 0.09638428688049316,
172
+ "rewards/margins": 0.900995135307312,
173
+ "rewards/rejected": -0.8046110272407532,
174
+ "step": 55
175
+ },
176
+ {
177
+ "epoch": 0.32,
178
+ "grad_norm": 13.533390998840332,
179
+ "learning_rate": 1e-06,
180
+ "logits/chosen": -0.9142201542854309,
181
+ "logits/rejected": -0.9173645973205566,
182
+ "logps/chosen": -79.86390686035156,
183
+ "logps/rejected": -71.9579086303711,
184
+ "loss": 0.5225,
185
+ "rewards/accuracies": 0.6499999761581421,
186
+ "rewards/chosen": 0.173385351896286,
187
+ "rewards/margins": 1.1095057725906372,
188
+ "rewards/rejected": -0.9361203908920288,
189
+ "step": 60
190
+ },
191
+ {
192
+ "epoch": 0.3466666666666667,
193
+ "grad_norm": 12.57950210571289,
194
+ "learning_rate": 1e-06,
195
+ "logits/chosen": -0.920464038848877,
196
+ "logits/rejected": -0.9172080159187317,
197
+ "logps/chosen": -80.74758911132812,
198
+ "logps/rejected": -76.41374969482422,
199
+ "loss": 0.5248,
200
+ "rewards/accuracies": 0.699999988079071,
201
+ "rewards/chosen": 0.2707827687263489,
202
+ "rewards/margins": 0.8655017614364624,
203
+ "rewards/rejected": -0.5947191119194031,
204
+ "step": 65
205
+ },
206
+ {
207
+ "epoch": 0.37333333333333335,
208
+ "grad_norm": 17.20355224609375,
209
+ "learning_rate": 1e-06,
210
+ "logits/chosen": -0.8898583650588989,
211
+ "logits/rejected": -0.8803698420524597,
212
+ "logps/chosen": -80.18648529052734,
213
+ "logps/rejected": -109.63679504394531,
214
+ "loss": 0.3912,
215
+ "rewards/accuracies": 0.800000011920929,
216
+ "rewards/chosen": -0.030266547575592995,
217
+ "rewards/margins": 1.6781041622161865,
218
+ "rewards/rejected": -1.7083708047866821,
219
+ "step": 70
220
+ },
221
+ {
222
+ "epoch": 0.4,
223
+ "grad_norm": 14.076622009277344,
224
+ "learning_rate": 1e-06,
225
+ "logits/chosen": -0.9825113415718079,
226
+ "logits/rejected": -0.9686470031738281,
227
+ "logps/chosen": -72.63887786865234,
228
+ "logps/rejected": -98.7930908203125,
229
+ "loss": 0.4875,
230
+ "rewards/accuracies": 0.800000011920929,
231
+ "rewards/chosen": 0.3350667357444763,
232
+ "rewards/margins": 1.8856172561645508,
233
+ "rewards/rejected": -1.5505504608154297,
234
+ "step": 75
235
+ },
236
+ {
237
+ "epoch": 0.4266666666666667,
238
+ "grad_norm": 13.67685317993164,
239
+ "learning_rate": 1e-06,
240
+ "logits/chosen": -0.9678022265434265,
241
+ "logits/rejected": -0.9671600461006165,
242
+ "logps/chosen": -47.99736785888672,
243
+ "logps/rejected": -77.59354400634766,
244
+ "loss": 0.4304,
245
+ "rewards/accuracies": 0.8999999761581421,
246
+ "rewards/chosen": 0.5903635621070862,
247
+ "rewards/margins": 1.1374304294586182,
248
+ "rewards/rejected": -0.547066867351532,
249
+ "step": 80
250
+ },
251
+ {
252
+ "epoch": 0.4533333333333333,
253
+ "grad_norm": 12.650314331054688,
254
+ "learning_rate": 1e-06,
255
+ "logits/chosen": -0.9299849271774292,
256
+ "logits/rejected": -0.9215339422225952,
257
+ "logps/chosen": -84.56604766845703,
258
+ "logps/rejected": -96.4300308227539,
259
+ "loss": 0.3804,
260
+ "rewards/accuracies": 0.6000000238418579,
261
+ "rewards/chosen": 0.6257888078689575,
262
+ "rewards/margins": 1.4266914129257202,
263
+ "rewards/rejected": -0.8009026646614075,
264
+ "step": 85
265
+ },
266
+ {
267
+ "epoch": 0.48,
268
+ "grad_norm": 12.806818962097168,
269
+ "learning_rate": 1e-06,
270
+ "logits/chosen": -0.9175389409065247,
271
+ "logits/rejected": -0.9051604270935059,
272
+ "logps/chosen": -96.42292022705078,
273
+ "logps/rejected": -106.83795166015625,
274
+ "loss": 0.4908,
275
+ "rewards/accuracies": 0.800000011920929,
276
+ "rewards/chosen": -0.018072601407766342,
277
+ "rewards/margins": 1.4671040773391724,
278
+ "rewards/rejected": -1.4851768016815186,
279
+ "step": 90
280
+ },
281
+ {
282
+ "epoch": 0.5066666666666667,
283
+ "grad_norm": 15.225826263427734,
284
+ "learning_rate": 1e-06,
285
+ "logits/chosen": -0.9506580233573914,
286
+ "logits/rejected": -0.9402367472648621,
287
+ "logps/chosen": -83.20651245117188,
288
+ "logps/rejected": -106.87049865722656,
289
+ "loss": 0.4715,
290
+ "rewards/accuracies": 0.75,
291
+ "rewards/chosen": 0.14173248410224915,
292
+ "rewards/margins": 1.5761946439743042,
293
+ "rewards/rejected": -1.434462308883667,
294
+ "step": 95
295
+ },
296
+ {
297
+ "epoch": 0.5333333333333333,
298
+ "grad_norm": 17.279712677001953,
299
+ "learning_rate": 1e-06,
300
+ "logits/chosen": -0.8913367390632629,
301
+ "logits/rejected": -0.8918878436088562,
302
+ "logps/chosen": -46.1912956237793,
303
+ "logps/rejected": -58.76585006713867,
304
+ "loss": 0.4605,
305
+ "rewards/accuracies": 0.699999988079071,
306
+ "rewards/chosen": 0.3738502860069275,
307
+ "rewards/margins": 1.1125816106796265,
308
+ "rewards/rejected": -0.7387313842773438,
309
+ "step": 100
310
+ },
311
+ {
312
+ "epoch": 0.56,
313
+ "grad_norm": 15.407678604125977,
314
+ "learning_rate": 1e-06,
315
+ "logits/chosen": -0.9308265447616577,
316
+ "logits/rejected": -0.9138677716255188,
317
+ "logps/chosen": -89.54922485351562,
318
+ "logps/rejected": -90.88710021972656,
319
+ "loss": 0.4469,
320
+ "rewards/accuracies": 0.8999999761581421,
321
+ "rewards/chosen": 0.10435737669467926,
322
+ "rewards/margins": 1.630505919456482,
323
+ "rewards/rejected": -1.5261485576629639,
324
+ "step": 105
325
+ },
326
+ {
327
+ "epoch": 0.5866666666666667,
328
+ "grad_norm": 14.938887596130371,
329
+ "learning_rate": 1e-06,
330
+ "logits/chosen": -0.965082049369812,
331
+ "logits/rejected": -0.9715268015861511,
332
+ "logps/chosen": -42.88313674926758,
333
+ "logps/rejected": -61.1027717590332,
334
+ "loss": 0.4947,
335
+ "rewards/accuracies": 0.6000000238418579,
336
+ "rewards/chosen": 0.0917281061410904,
337
+ "rewards/margins": 0.6923992037773132,
338
+ "rewards/rejected": -0.6006711721420288,
339
+ "step": 110
340
+ },
341
+ {
342
+ "epoch": 0.6133333333333333,
343
+ "grad_norm": 15.150486946105957,
344
+ "learning_rate": 1e-06,
345
+ "logits/chosen": -1.0348211526870728,
346
+ "logits/rejected": -1.0350416898727417,
347
+ "logps/chosen": -58.58684158325195,
348
+ "logps/rejected": -69.80925750732422,
349
+ "loss": 0.4301,
350
+ "rewards/accuracies": 0.949999988079071,
351
+ "rewards/chosen": 0.741683304309845,
352
+ "rewards/margins": 1.3792873620986938,
353
+ "rewards/rejected": -0.6376041173934937,
354
+ "step": 115
355
+ },
356
+ {
357
+ "epoch": 0.64,
358
+ "grad_norm": 15.18857192993164,
359
+ "learning_rate": 1e-06,
360
+ "logits/chosen": -0.9605352282524109,
361
+ "logits/rejected": -0.9632300138473511,
362
+ "logps/chosen": -66.8796615600586,
363
+ "logps/rejected": -71.91226959228516,
364
+ "loss": 0.4272,
365
+ "rewards/accuracies": 0.75,
366
+ "rewards/chosen": 0.334277480840683,
367
+ "rewards/margins": 1.3439124822616577,
368
+ "rewards/rejected": -1.0096348524093628,
369
+ "step": 120
370
+ },
371
+ {
372
+ "epoch": 0.6666666666666666,
373
+ "grad_norm": 17.82939338684082,
374
+ "learning_rate": 1e-06,
375
+ "logits/chosen": -0.9578378796577454,
376
+ "logits/rejected": -0.9434881210327148,
377
+ "logps/chosen": -91.96797943115234,
378
+ "logps/rejected": -118.94937896728516,
379
+ "loss": 0.469,
380
+ "rewards/accuracies": 0.800000011920929,
381
+ "rewards/chosen": 0.11757056415081024,
382
+ "rewards/margins": 2.0075435638427734,
383
+ "rewards/rejected": -1.8899730443954468,
384
+ "step": 125
385
+ },
386
+ {
387
+ "epoch": 0.6933333333333334,
388
+ "grad_norm": 14.026658058166504,
389
+ "learning_rate": 1e-06,
390
+ "logits/chosen": -0.9356282353401184,
391
+ "logits/rejected": -0.9223386645317078,
392
+ "logps/chosen": -84.6489486694336,
393
+ "logps/rejected": -97.97027587890625,
394
+ "loss": 0.4071,
395
+ "rewards/accuracies": 0.800000011920929,
396
+ "rewards/chosen": -0.1822044849395752,
397
+ "rewards/margins": 1.0983351469039917,
398
+ "rewards/rejected": -1.2805397510528564,
399
+ "step": 130
400
+ },
401
+ {
402
+ "epoch": 0.72,
403
+ "grad_norm": 16.48579978942871,
404
+ "learning_rate": 1e-06,
405
+ "logits/chosen": -0.9593449831008911,
406
+ "logits/rejected": -0.9539397358894348,
407
+ "logps/chosen": -87.32046508789062,
408
+ "logps/rejected": -104.6070327758789,
409
+ "loss": 0.4735,
410
+ "rewards/accuracies": 0.8500000238418579,
411
+ "rewards/chosen": 0.11158242076635361,
412
+ "rewards/margins": 1.4765504598617554,
413
+ "rewards/rejected": -1.3649680614471436,
414
+ "step": 135
415
+ },
416
+ {
417
+ "epoch": 0.7466666666666667,
418
+ "grad_norm": 15.402965545654297,
419
+ "learning_rate": 1e-06,
420
+ "logits/chosen": -0.8937503695487976,
421
+ "logits/rejected": -0.8833414912223816,
422
+ "logps/chosen": -56.55735397338867,
423
+ "logps/rejected": -64.20155334472656,
424
+ "loss": 0.444,
425
+ "rewards/accuracies": 0.699999988079071,
426
+ "rewards/chosen": -0.04381413012742996,
427
+ "rewards/margins": 1.1586054563522339,
428
+ "rewards/rejected": -1.202419638633728,
429
+ "step": 140
430
+ },
431
+ {
432
+ "epoch": 0.7733333333333333,
433
+ "grad_norm": 13.747047424316406,
434
+ "learning_rate": 1e-06,
435
+ "logits/chosen": -0.9597466588020325,
436
+ "logits/rejected": -0.9614435434341431,
437
+ "logps/chosen": -63.605621337890625,
438
+ "logps/rejected": -67.96479797363281,
439
+ "loss": 0.3972,
440
+ "rewards/accuracies": 0.699999988079071,
441
+ "rewards/chosen": 0.4212762713432312,
442
+ "rewards/margins": 0.9387065768241882,
443
+ "rewards/rejected": -0.5174302458763123,
444
+ "step": 145
445
+ },
446
+ {
447
+ "epoch": 0.8,
448
+ "grad_norm": 12.082706451416016,
449
+ "learning_rate": 1e-06,
450
+ "logits/chosen": -0.8831542730331421,
451
+ "logits/rejected": -0.8782299757003784,
452
+ "logps/chosen": -89.6635971069336,
453
+ "logps/rejected": -102.15290832519531,
454
+ "loss": 0.4195,
455
+ "rewards/accuracies": 0.75,
456
+ "rewards/chosen": -0.23645827174186707,
457
+ "rewards/margins": 1.130458116531372,
458
+ "rewards/rejected": -1.3669164180755615,
459
+ "step": 150
460
+ },
461
+ {
462
+ "epoch": 0.8266666666666667,
463
+ "grad_norm": 11.634166717529297,
464
+ "learning_rate": 1e-06,
465
+ "logits/chosen": -0.9502588510513306,
466
+ "logits/rejected": -0.9416471719741821,
467
+ "logps/chosen": -68.0828857421875,
468
+ "logps/rejected": -100.38713073730469,
469
+ "loss": 0.3315,
470
+ "rewards/accuracies": 0.800000011920929,
471
+ "rewards/chosen": 0.19413085281848907,
472
+ "rewards/margins": 2.068560838699341,
473
+ "rewards/rejected": -1.8744302988052368,
474
+ "step": 155
475
+ },
476
+ {
477
+ "epoch": 0.8533333333333334,
478
+ "grad_norm": 15.321958541870117,
479
+ "learning_rate": 1e-06,
480
+ "logits/chosen": -1.0011286735534668,
481
+ "logits/rejected": -1.0069797039031982,
482
+ "logps/chosen": -58.62431716918945,
483
+ "logps/rejected": -81.33189392089844,
484
+ "loss": 0.4751,
485
+ "rewards/accuracies": 0.6000000238418579,
486
+ "rewards/chosen": 0.06445435434579849,
487
+ "rewards/margins": 1.0574184656143188,
488
+ "rewards/rejected": -0.9929640889167786,
489
+ "step": 160
490
+ },
491
+ {
492
+ "epoch": 0.88,
493
+ "grad_norm": 13.829776763916016,
494
+ "learning_rate": 1e-06,
495
+ "logits/chosen": -0.9930332899093628,
496
+ "logits/rejected": -0.9925543069839478,
497
+ "logps/chosen": -65.17711639404297,
498
+ "logps/rejected": -79.74466705322266,
499
+ "loss": 0.4647,
500
+ "rewards/accuracies": 0.800000011920929,
501
+ "rewards/chosen": 0.07586468756198883,
502
+ "rewards/margins": 0.9637776613235474,
503
+ "rewards/rejected": -0.8879130482673645,
504
+ "step": 165
505
+ },
506
+ {
507
+ "epoch": 0.9066666666666666,
508
+ "grad_norm": 15.529192924499512,
509
+ "learning_rate": 1e-06,
510
+ "logits/chosen": -0.9089356660842896,
511
+ "logits/rejected": -0.9082063436508179,
512
+ "logps/chosen": -73.97811126708984,
513
+ "logps/rejected": -95.45903778076172,
514
+ "loss": 0.4392,
515
+ "rewards/accuracies": 0.699999988079071,
516
+ "rewards/chosen": -0.33137935400009155,
517
+ "rewards/margins": 1.3767789602279663,
518
+ "rewards/rejected": -1.7081581354141235,
519
+ "step": 170
520
+ },
521
+ {
522
+ "epoch": 0.9333333333333333,
523
+ "grad_norm": 16.684486389160156,
524
+ "learning_rate": 1e-06,
525
+ "logits/chosen": -0.97789067029953,
526
+ "logits/rejected": -0.9672166705131531,
527
+ "logps/chosen": -68.05913543701172,
528
+ "logps/rejected": -76.677490234375,
529
+ "loss": 0.3687,
530
+ "rewards/accuracies": 0.699999988079071,
531
+ "rewards/chosen": 0.15937092900276184,
532
+ "rewards/margins": 1.4713166952133179,
533
+ "rewards/rejected": -1.3119456768035889,
534
+ "step": 175
535
+ },
536
+ {
537
+ "epoch": 0.96,
538
+ "grad_norm": 10.531698226928711,
539
+ "learning_rate": 1e-06,
540
+ "logits/chosen": -0.9809746742248535,
541
+ "logits/rejected": -0.963492214679718,
542
+ "logps/chosen": -82.75611877441406,
543
+ "logps/rejected": -115.19696044921875,
544
+ "loss": 0.3942,
545
+ "rewards/accuracies": 1.0,
546
+ "rewards/chosen": 0.1854589432477951,
547
+ "rewards/margins": 2.5827653408050537,
548
+ "rewards/rejected": -2.397306442260742,
549
+ "step": 180
550
+ },
551
+ {
552
+ "epoch": 0.9866666666666667,
553
+ "grad_norm": 13.4787015914917,
554
+ "learning_rate": 1e-06,
555
+ "logits/chosen": -0.96388179063797,
556
+ "logits/rejected": -0.9607783555984497,
557
+ "logps/chosen": -73.0042953491211,
558
+ "logps/rejected": -70.76305389404297,
559
+ "loss": 0.3791,
560
+ "rewards/accuracies": 0.550000011920929,
561
+ "rewards/chosen": 0.26272085309028625,
562
+ "rewards/margins": 1.4707610607147217,
563
+ "rewards/rejected": -1.2080401182174683,
564
+ "step": 185
565
+ },
566
+ {
567
+ "epoch": 1.0133333333333334,
568
+ "grad_norm": 10.95159912109375,
569
+ "learning_rate": 1e-06,
570
+ "logits/chosen": -0.952096164226532,
571
+ "logits/rejected": -0.9481242895126343,
572
+ "logps/chosen": -65.6482162475586,
573
+ "logps/rejected": -94.6727523803711,
574
+ "loss": 0.3602,
575
+ "rewards/accuracies": 0.8999999761581421,
576
+ "rewards/chosen": 0.6763356924057007,
577
+ "rewards/margins": 1.9549248218536377,
578
+ "rewards/rejected": -1.2785890102386475,
579
+ "step": 190
580
+ },
581
+ {
582
+ "epoch": 1.04,
583
+ "grad_norm": 4.499619960784912,
584
+ "learning_rate": 1e-06,
585
+ "logits/chosen": -0.9216955900192261,
586
+ "logits/rejected": -0.9177412986755371,
587
+ "logps/chosen": -57.536109924316406,
588
+ "logps/rejected": -89.177734375,
589
+ "loss": 0.1967,
590
+ "rewards/accuracies": 0.8999999761581421,
591
+ "rewards/chosen": 0.7885057330131531,
592
+ "rewards/margins": 1.7107608318328857,
593
+ "rewards/rejected": -0.9222550392150879,
594
+ "step": 195
595
+ },
596
+ {
597
+ "epoch": 1.0666666666666667,
598
+ "grad_norm": 7.151574611663818,
599
+ "learning_rate": 1e-06,
600
+ "logits/chosen": -0.9569188952445984,
601
+ "logits/rejected": -0.9455480575561523,
602
+ "logps/chosen": -53.17787551879883,
603
+ "logps/rejected": -78.4444351196289,
604
+ "loss": 0.2083,
605
+ "rewards/accuracies": 0.8999999761581421,
606
+ "rewards/chosen": 0.8187260627746582,
607
+ "rewards/margins": 2.6014370918273926,
608
+ "rewards/rejected": -1.7827112674713135,
609
+ "step": 200
610
+ },
611
+ {
612
+ "epoch": 1.0933333333333333,
613
+ "grad_norm": 7.790274620056152,
614
+ "learning_rate": 1e-06,
615
+ "logits/chosen": -0.9871414303779602,
616
+ "logits/rejected": -0.9741198420524597,
617
+ "logps/chosen": -88.66045379638672,
618
+ "logps/rejected": -130.34814453125,
619
+ "loss": 0.1628,
620
+ "rewards/accuracies": 1.0,
621
+ "rewards/chosen": 0.14811792969703674,
622
+ "rewards/margins": 3.1860337257385254,
623
+ "rewards/rejected": -3.0379159450531006,
624
+ "step": 205
625
+ },
626
+ {
627
+ "epoch": 1.12,
628
+ "grad_norm": 5.487607955932617,
629
+ "learning_rate": 1e-06,
630
+ "logits/chosen": -1.034203052520752,
631
+ "logits/rejected": -1.020320177078247,
632
+ "logps/chosen": -71.07328796386719,
633
+ "logps/rejected": -107.5151138305664,
634
+ "loss": 0.1747,
635
+ "rewards/accuracies": 0.8999999761581421,
636
+ "rewards/chosen": 0.050754476338624954,
637
+ "rewards/margins": 3.4751205444335938,
638
+ "rewards/rejected": -3.4243664741516113,
639
+ "step": 210
640
+ },
641
+ {
642
+ "epoch": 1.1466666666666667,
643
+ "grad_norm": 4.493232727050781,
644
+ "learning_rate": 1e-06,
645
+ "logits/chosen": -0.9926775693893433,
646
+ "logits/rejected": -0.9901949167251587,
647
+ "logps/chosen": -85.27278900146484,
648
+ "logps/rejected": -123.84268951416016,
649
+ "loss": 0.1556,
650
+ "rewards/accuracies": 1.0,
651
+ "rewards/chosen": -0.030883288010954857,
652
+ "rewards/margins": 3.6463749408721924,
653
+ "rewards/rejected": -3.677258253097534,
654
+ "step": 215
655
+ },
656
+ {
657
+ "epoch": 1.1733333333333333,
658
+ "grad_norm": 8.204482078552246,
659
+ "learning_rate": 1e-06,
660
+ "logits/chosen": -1.012113332748413,
661
+ "logits/rejected": -1.012290596961975,
662
+ "logps/chosen": -85.76399993896484,
663
+ "logps/rejected": -119.85150146484375,
664
+ "loss": 0.1645,
665
+ "rewards/accuracies": 1.0,
666
+ "rewards/chosen": -0.43645215034484863,
667
+ "rewards/margins": 3.1804490089416504,
668
+ "rewards/rejected": -3.616900682449341,
669
+ "step": 220
670
+ },
671
+ {
672
+ "epoch": 1.2,
673
+ "grad_norm": 8.91741943359375,
674
+ "learning_rate": 1e-06,
675
+ "logits/chosen": -0.9512453079223633,
676
+ "logits/rejected": -0.9447379112243652,
677
+ "logps/chosen": -99.74169921875,
678
+ "logps/rejected": -125.4989013671875,
679
+ "loss": 0.2412,
680
+ "rewards/accuracies": 0.8999999761581421,
681
+ "rewards/chosen": -1.6342785358428955,
682
+ "rewards/margins": 3.1997177600860596,
683
+ "rewards/rejected": -4.833995819091797,
684
+ "step": 225
685
+ },
686
+ {
687
+ "epoch": 1.2266666666666666,
688
+ "grad_norm": 8.655539512634277,
689
+ "learning_rate": 1e-06,
690
+ "logits/chosen": -0.9638906717300415,
691
+ "logits/rejected": -0.9550336599349976,
692
+ "logps/chosen": -90.74691009521484,
693
+ "logps/rejected": -129.67747497558594,
694
+ "loss": 0.1931,
695
+ "rewards/accuracies": 0.8999999761581421,
696
+ "rewards/chosen": -0.6765635013580322,
697
+ "rewards/margins": 3.924309492111206,
698
+ "rewards/rejected": -4.600872993469238,
699
+ "step": 230
700
+ },
701
+ {
702
+ "epoch": 1.2533333333333334,
703
+ "grad_norm": 4.054169654846191,
704
+ "learning_rate": 1e-06,
705
+ "logits/chosen": -1.030977487564087,
706
+ "logits/rejected": -1.0393321514129639,
707
+ "logps/chosen": -54.60747528076172,
708
+ "logps/rejected": -86.23637390136719,
709
+ "loss": 0.2012,
710
+ "rewards/accuracies": 0.949999988079071,
711
+ "rewards/chosen": 0.5984514951705933,
712
+ "rewards/margins": 3.5396370887756348,
713
+ "rewards/rejected": -2.941185712814331,
714
+ "step": 235
715
+ },
716
+ {
717
+ "epoch": 1.28,
718
+ "grad_norm": 6.45543909072876,
719
+ "learning_rate": 1e-06,
720
+ "logits/chosen": -1.000688910484314,
721
+ "logits/rejected": -0.9959505200386047,
722
+ "logps/chosen": -51.59619140625,
723
+ "logps/rejected": -78.5389633178711,
724
+ "loss": 0.1828,
725
+ "rewards/accuracies": 0.8999999761581421,
726
+ "rewards/chosen": 0.38332298398017883,
727
+ "rewards/margins": 2.686915874481201,
728
+ "rewards/rejected": -2.3035929203033447,
729
+ "step": 240
730
+ },
731
+ {
732
+ "epoch": 1.3066666666666666,
733
+ "grad_norm": 7.938249111175537,
734
+ "learning_rate": 1e-06,
735
+ "logits/chosen": -0.9679718017578125,
736
+ "logits/rejected": -0.9519965052604675,
737
+ "logps/chosen": -91.22569274902344,
738
+ "logps/rejected": -120.7100601196289,
739
+ "loss": 0.1685,
740
+ "rewards/accuracies": 1.0,
741
+ "rewards/chosen": -0.6173192262649536,
742
+ "rewards/margins": 3.579179286956787,
743
+ "rewards/rejected": -4.196498870849609,
744
+ "step": 245
745
+ },
746
+ {
747
+ "epoch": 1.3333333333333333,
748
+ "grad_norm": 5.844974040985107,
749
+ "learning_rate": 1e-06,
750
+ "logits/chosen": -0.9622223973274231,
751
+ "logits/rejected": -0.9677648544311523,
752
+ "logps/chosen": -92.24385070800781,
753
+ "logps/rejected": -129.18057250976562,
754
+ "loss": 0.1531,
755
+ "rewards/accuracies": 0.949999988079071,
756
+ "rewards/chosen": -0.8045636415481567,
757
+ "rewards/margins": 2.904844284057617,
758
+ "rewards/rejected": -3.7094082832336426,
759
+ "step": 250
760
+ },
761
+ {
762
+ "epoch": 1.3599999999999999,
763
+ "grad_norm": 8.192586898803711,
764
+ "learning_rate": 1e-06,
765
+ "logits/chosen": -0.9564617276191711,
766
+ "logits/rejected": -0.953482985496521,
767
+ "logps/chosen": -92.21900939941406,
768
+ "logps/rejected": -115.93299865722656,
769
+ "loss": 0.1831,
770
+ "rewards/accuracies": 0.800000011920929,
771
+ "rewards/chosen": -0.9339162707328796,
772
+ "rewards/margins": 3.1692094802856445,
773
+ "rewards/rejected": -4.103126049041748,
774
+ "step": 255
775
+ },
776
+ {
777
+ "epoch": 1.3866666666666667,
778
+ "grad_norm": 11.612703323364258,
779
+ "learning_rate": 1e-06,
780
+ "logits/chosen": -0.9484894871711731,
781
+ "logits/rejected": -0.9465099573135376,
782
+ "logps/chosen": -88.50215148925781,
783
+ "logps/rejected": -122.58497619628906,
784
+ "loss": 0.2373,
785
+ "rewards/accuracies": 1.0,
786
+ "rewards/chosen": -0.8389456868171692,
787
+ "rewards/margins": 4.004569053649902,
788
+ "rewards/rejected": -4.843514442443848,
789
+ "step": 260
790
+ },
791
+ {
792
+ "epoch": 1.4133333333333333,
793
+ "grad_norm": 3.950273036956787,
794
+ "learning_rate": 1e-06,
795
+ "logits/chosen": -1.0333240032196045,
796
+ "logits/rejected": -1.026308298110962,
797
+ "logps/chosen": -81.88277435302734,
798
+ "logps/rejected": -115.6543960571289,
799
+ "loss": 0.1738,
800
+ "rewards/accuracies": 0.949999988079071,
801
+ "rewards/chosen": -0.251314640045166,
802
+ "rewards/margins": 3.181288242340088,
803
+ "rewards/rejected": -3.432603120803833,
804
+ "step": 265
805
+ },
806
+ {
807
+ "epoch": 1.44,
808
+ "grad_norm": 9.259121894836426,
809
+ "learning_rate": 1e-06,
810
+ "logits/chosen": -0.9529396891593933,
811
+ "logits/rejected": -0.9476019740104675,
812
+ "logps/chosen": -93.74283599853516,
813
+ "logps/rejected": -117.82435607910156,
814
+ "loss": 0.1717,
815
+ "rewards/accuracies": 0.949999988079071,
816
+ "rewards/chosen": -0.17003877460956573,
817
+ "rewards/margins": 3.552433490753174,
818
+ "rewards/rejected": -3.7224719524383545,
819
+ "step": 270
820
+ },
821
+ {
822
+ "epoch": 1.4666666666666668,
823
+ "grad_norm": 17.85709571838379,
824
+ "learning_rate": 1e-06,
825
+ "logits/chosen": -1.0109868049621582,
826
+ "logits/rejected": -1.0117450952529907,
827
+ "logps/chosen": -93.20274353027344,
828
+ "logps/rejected": -124.9032211303711,
829
+ "loss": 0.2349,
830
+ "rewards/accuracies": 1.0,
831
+ "rewards/chosen": -0.15422244369983673,
832
+ "rewards/margins": 3.6027817726135254,
833
+ "rewards/rejected": -3.757004499435425,
834
+ "step": 275
835
+ },
836
+ {
837
+ "epoch": 1.4933333333333334,
838
+ "grad_norm": 12.522117614746094,
839
+ "learning_rate": 1e-06,
840
+ "logits/chosen": -1.0297268629074097,
841
+ "logits/rejected": -1.0234237909317017,
842
+ "logps/chosen": -50.927391052246094,
843
+ "logps/rejected": -76.98394775390625,
844
+ "loss": 0.2888,
845
+ "rewards/accuracies": 0.949999988079071,
846
+ "rewards/chosen": 0.37279123067855835,
847
+ "rewards/margins": 2.9437782764434814,
848
+ "rewards/rejected": -2.5709872245788574,
849
+ "step": 280
850
+ },
851
+ {
852
+ "epoch": 1.52,
853
+ "grad_norm": 18.62285804748535,
854
+ "learning_rate": 1e-06,
855
+ "logits/chosen": -0.97401362657547,
856
+ "logits/rejected": -0.9802001714706421,
857
+ "logps/chosen": -63.81083297729492,
858
+ "logps/rejected": -86.34494018554688,
859
+ "loss": 0.2402,
860
+ "rewards/accuracies": 0.8500000238418579,
861
+ "rewards/chosen": -0.40913981199264526,
862
+ "rewards/margins": 2.3060526847839355,
863
+ "rewards/rejected": -2.7151923179626465,
864
+ "step": 285
865
+ },
866
+ {
867
+ "epoch": 1.5466666666666666,
868
+ "grad_norm": 14.160591125488281,
869
+ "learning_rate": 1e-06,
870
+ "logits/chosen": -1.0169212818145752,
871
+ "logits/rejected": -1.0096107721328735,
872
+ "logps/chosen": -66.8814926147461,
873
+ "logps/rejected": -97.15755462646484,
874
+ "loss": 0.1656,
875
+ "rewards/accuracies": 0.949999988079071,
876
+ "rewards/chosen": 0.21899671852588654,
877
+ "rewards/margins": 3.427858352661133,
878
+ "rewards/rejected": -3.2088615894317627,
879
+ "step": 290
880
+ },
881
+ {
882
+ "epoch": 1.5733333333333333,
883
+ "grad_norm": 5.416359901428223,
884
+ "learning_rate": 1e-06,
885
+ "logits/chosen": -0.9776760935783386,
886
+ "logits/rejected": -0.9770166277885437,
887
+ "logps/chosen": -67.19379425048828,
888
+ "logps/rejected": -101.93242645263672,
889
+ "loss": 0.2106,
890
+ "rewards/accuracies": 0.8999999761581421,
891
+ "rewards/chosen": -0.027063405141234398,
892
+ "rewards/margins": 3.1595966815948486,
893
+ "rewards/rejected": -3.1866602897644043,
894
+ "step": 295
895
+ },
896
+ {
897
+ "epoch": 1.6,
898
+ "grad_norm": 7.018192291259766,
899
+ "learning_rate": 1e-06,
900
+ "logits/chosen": -1.023803949356079,
901
+ "logits/rejected": -1.0103000402450562,
902
+ "logps/chosen": -63.242576599121094,
903
+ "logps/rejected": -96.89302062988281,
904
+ "loss": 0.1886,
905
+ "rewards/accuracies": 0.8500000238418579,
906
+ "rewards/chosen": 0.034027088433504105,
907
+ "rewards/margins": 3.0590438842773438,
908
+ "rewards/rejected": -3.0250167846679688,
909
+ "step": 300
910
+ },
911
+ {
912
+ "epoch": 1.6266666666666667,
913
+ "grad_norm": 7.443521499633789,
914
+ "learning_rate": 1e-06,
915
+ "logits/chosen": -1.0116432905197144,
916
+ "logits/rejected": -1.023240327835083,
917
+ "logps/chosen": -84.08271026611328,
918
+ "logps/rejected": -159.3538360595703,
919
+ "loss": 0.1687,
920
+ "rewards/accuracies": 0.8500000238418579,
921
+ "rewards/chosen": -0.561992347240448,
922
+ "rewards/margins": 4.012616157531738,
923
+ "rewards/rejected": -4.57460880279541,
924
+ "step": 305
925
+ },
926
+ {
927
+ "epoch": 1.6533333333333333,
928
+ "grad_norm": 18.155902862548828,
929
+ "learning_rate": 1e-06,
930
+ "logits/chosen": -1.025331974029541,
931
+ "logits/rejected": -1.021927833557129,
932
+ "logps/chosen": -73.14906311035156,
933
+ "logps/rejected": -121.94499206542969,
934
+ "loss": 0.1807,
935
+ "rewards/accuracies": 0.949999988079071,
936
+ "rewards/chosen": -0.2425353080034256,
937
+ "rewards/margins": 3.504972457885742,
938
+ "rewards/rejected": -3.7475082874298096,
939
+ "step": 310
940
+ },
941
+ {
942
+ "epoch": 1.6800000000000002,
943
+ "grad_norm": 11.773829460144043,
944
+ "learning_rate": 1e-06,
945
+ "logits/chosen": -0.9748228192329407,
946
+ "logits/rejected": -0.9767504930496216,
947
+ "logps/chosen": -88.07221984863281,
948
+ "logps/rejected": -147.24069213867188,
949
+ "loss": 0.1454,
950
+ "rewards/accuracies": 0.949999988079071,
951
+ "rewards/chosen": -0.8688488006591797,
952
+ "rewards/margins": 4.263859748840332,
953
+ "rewards/rejected": -5.1327080726623535,
954
+ "step": 315
955
+ },
956
+ {
957
+ "epoch": 1.7066666666666666,
958
+ "grad_norm": 11.361552238464355,
959
+ "learning_rate": 1e-06,
960
+ "logits/chosen": -0.9369309544563293,
961
+ "logits/rejected": -0.9380790591239929,
962
+ "logps/chosen": -96.55718231201172,
963
+ "logps/rejected": -119.29121398925781,
964
+ "loss": 0.2044,
965
+ "rewards/accuracies": 0.8999999761581421,
966
+ "rewards/chosen": -0.5562452077865601,
967
+ "rewards/margins": 3.295253038406372,
968
+ "rewards/rejected": -3.8514983654022217,
969
+ "step": 320
970
+ },
971
+ {
972
+ "epoch": 1.7333333333333334,
973
+ "grad_norm": 10.781094551086426,
974
+ "learning_rate": 1e-06,
975
+ "logits/chosen": -0.99322509765625,
976
+ "logits/rejected": -0.9963001012802124,
977
+ "logps/chosen": -94.03239440917969,
978
+ "logps/rejected": -129.18307495117188,
979
+ "loss": 0.228,
980
+ "rewards/accuracies": 0.949999988079071,
981
+ "rewards/chosen": -1.0290601253509521,
982
+ "rewards/margins": 3.1235461235046387,
983
+ "rewards/rejected": -4.152606010437012,
984
+ "step": 325
985
+ },
986
+ {
987
+ "epoch": 1.76,
988
+ "grad_norm": 13.973287582397461,
989
+ "learning_rate": 1e-06,
990
+ "logits/chosen": -0.9588006138801575,
991
+ "logits/rejected": -0.9579130411148071,
992
+ "logps/chosen": -93.3379135131836,
993
+ "logps/rejected": -145.30789184570312,
994
+ "loss": 0.1606,
995
+ "rewards/accuracies": 1.0,
996
+ "rewards/chosen": -0.4644792675971985,
997
+ "rewards/margins": 4.343870639801025,
998
+ "rewards/rejected": -4.808349609375,
999
+ "step": 330
1000
+ },
1001
+ {
1002
+ "epoch": 1.7866666666666666,
1003
+ "grad_norm": 5.7010931968688965,
1004
+ "learning_rate": 1e-06,
1005
+ "logits/chosen": -0.9664214253425598,
1006
+ "logits/rejected": -0.9622098207473755,
1007
+ "logps/chosen": -77.8204574584961,
1008
+ "logps/rejected": -139.20196533203125,
1009
+ "loss": 0.1036,
1010
+ "rewards/accuracies": 0.8999999761581421,
1011
+ "rewards/chosen": -0.5566782355308533,
1012
+ "rewards/margins": 4.256915092468262,
1013
+ "rewards/rejected": -4.81359338760376,
1014
+ "step": 335
1015
+ },
1016
+ {
1017
+ "epoch": 1.8133333333333335,
1018
+ "grad_norm": 15.839357376098633,
1019
+ "learning_rate": 1e-06,
1020
+ "logits/chosen": -0.9637653231620789,
1021
+ "logits/rejected": -0.9616791009902954,
1022
+ "logps/chosen": -97.48753356933594,
1023
+ "logps/rejected": -127.541259765625,
1024
+ "loss": 0.1682,
1025
+ "rewards/accuracies": 0.949999988079071,
1026
+ "rewards/chosen": -0.562965452671051,
1027
+ "rewards/margins": 3.636152744293213,
1028
+ "rewards/rejected": -4.199118614196777,
1029
+ "step": 340
1030
+ },
1031
+ {
1032
+ "epoch": 1.8399999999999999,
1033
+ "grad_norm": 7.101598739624023,
1034
+ "learning_rate": 1e-06,
1035
+ "logits/chosen": -1.0025413036346436,
1036
+ "logits/rejected": -1.0012590885162354,
1037
+ "logps/chosen": -79.85330963134766,
1038
+ "logps/rejected": -98.39595794677734,
1039
+ "loss": 0.1832,
1040
+ "rewards/accuracies": 0.949999988079071,
1041
+ "rewards/chosen": 0.4088008403778076,
1042
+ "rewards/margins": 3.5112171173095703,
1043
+ "rewards/rejected": -3.102416515350342,
1044
+ "step": 345
1045
+ },
1046
+ {
1047
+ "epoch": 1.8666666666666667,
1048
+ "grad_norm": 12.390422821044922,
1049
+ "learning_rate": 1e-06,
1050
+ "logits/chosen": -1.0001270771026611,
1051
+ "logits/rejected": -1.0041121244430542,
1052
+ "logps/chosen": -47.137535095214844,
1053
+ "logps/rejected": -87.8708267211914,
1054
+ "loss": 0.1838,
1055
+ "rewards/accuracies": 0.8999999761581421,
1056
+ "rewards/chosen": -0.07667568325996399,
1057
+ "rewards/margins": 3.5933804512023926,
1058
+ "rewards/rejected": -3.670056104660034,
1059
+ "step": 350
1060
+ },
1061
+ {
1062
+ "epoch": 1.8933333333333333,
1063
+ "grad_norm": 6.624851226806641,
1064
+ "learning_rate": 1e-06,
1065
+ "logits/chosen": -1.0326136350631714,
1066
+ "logits/rejected": -1.0377240180969238,
1067
+ "logps/chosen": -78.13581848144531,
1068
+ "logps/rejected": -121.1558837890625,
1069
+ "loss": 0.1175,
1070
+ "rewards/accuracies": 0.949999988079071,
1071
+ "rewards/chosen": -0.22754593193531036,
1072
+ "rewards/margins": 3.7420687675476074,
1073
+ "rewards/rejected": -3.9696147441864014,
1074
+ "step": 355
1075
+ },
1076
+ {
1077
+ "epoch": 1.92,
1078
+ "grad_norm": 9.035065650939941,
1079
+ "learning_rate": 1e-06,
1080
+ "logits/chosen": -0.959603488445282,
1081
+ "logits/rejected": -0.9709904789924622,
1082
+ "logps/chosen": -76.04480743408203,
1083
+ "logps/rejected": -128.50241088867188,
1084
+ "loss": 0.1593,
1085
+ "rewards/accuracies": 0.8999999761581421,
1086
+ "rewards/chosen": -0.2988315522670746,
1087
+ "rewards/margins": 3.922844648361206,
1088
+ "rewards/rejected": -4.221675395965576,
1089
+ "step": 360
1090
+ },
1091
+ {
1092
+ "epoch": 1.9466666666666668,
1093
+ "grad_norm": 5.306190013885498,
1094
+ "learning_rate": 1e-06,
1095
+ "logits/chosen": -0.9443675875663757,
1096
+ "logits/rejected": -0.9363471865653992,
1097
+ "logps/chosen": -99.31537628173828,
1098
+ "logps/rejected": -136.64498901367188,
1099
+ "loss": 0.1425,
1100
+ "rewards/accuracies": 0.8999999761581421,
1101
+ "rewards/chosen": -1.223646879196167,
1102
+ "rewards/margins": 3.9859280586242676,
1103
+ "rewards/rejected": -5.209575653076172,
1104
+ "step": 365
1105
+ },
1106
+ {
1107
+ "epoch": 1.9733333333333334,
1108
+ "grad_norm": 12.788728713989258,
1109
+ "learning_rate": 1e-06,
1110
+ "logits/chosen": -0.9819198846817017,
1111
+ "logits/rejected": -0.9810832142829895,
1112
+ "logps/chosen": -79.12177276611328,
1113
+ "logps/rejected": -109.90110778808594,
1114
+ "loss": 0.209,
1115
+ "rewards/accuracies": 0.8999999761581421,
1116
+ "rewards/chosen": -0.3604089915752411,
1117
+ "rewards/margins": 3.0035877227783203,
1118
+ "rewards/rejected": -3.3639965057373047,
1119
+ "step": 370
1120
+ },
1121
+ {
1122
+ "epoch": 2.0,
1123
+ "grad_norm": 13.587727546691895,
1124
+ "learning_rate": 1e-06,
1125
+ "logits/chosen": -0.9736833572387695,
1126
+ "logits/rejected": -0.9709518551826477,
1127
+ "logps/chosen": -111.94571685791016,
1128
+ "logps/rejected": -123.80615234375,
1129
+ "loss": 0.1791,
1130
+ "rewards/accuracies": 0.949999988079071,
1131
+ "rewards/chosen": -1.4601728916168213,
1132
+ "rewards/margins": 3.195746660232544,
1133
+ "rewards/rejected": -4.655919551849365,
1134
+ "step": 375
1135
+ },
1136
+ {
1137
+ "epoch": 2.026666666666667,
1138
+ "grad_norm": 2.3324151039123535,
1139
+ "learning_rate": 1e-06,
1140
+ "logits/chosen": -0.9637060165405273,
1141
+ "logits/rejected": -0.9625332951545715,
1142
+ "logps/chosen": -64.67164611816406,
1143
+ "logps/rejected": -107.06028747558594,
1144
+ "loss": 0.0579,
1145
+ "rewards/accuracies": 0.949999988079071,
1146
+ "rewards/chosen": -0.13231408596038818,
1147
+ "rewards/margins": 4.237712383270264,
1148
+ "rewards/rejected": -4.370026588439941,
1149
+ "step": 380
1150
+ },
1151
+ {
1152
+ "epoch": 2.0533333333333332,
1153
+ "grad_norm": 2.0872855186462402,
1154
+ "learning_rate": 1e-06,
1155
+ "logits/chosen": -0.9391244649887085,
1156
+ "logits/rejected": -0.9333323240280151,
1157
+ "logps/chosen": -53.344078063964844,
1158
+ "logps/rejected": -118.3141098022461,
1159
+ "loss": 0.0571,
1160
+ "rewards/accuracies": 0.8999999761581421,
1161
+ "rewards/chosen": -0.09113900363445282,
1162
+ "rewards/margins": 4.982771873474121,
1163
+ "rewards/rejected": -5.073911190032959,
1164
+ "step": 385
1165
+ },
1166
+ {
1167
+ "epoch": 2.08,
1168
+ "grad_norm": 3.842839479446411,
1169
+ "learning_rate": 1e-06,
1170
+ "logits/chosen": -1.0063703060150146,
1171
+ "logits/rejected": -1.0035226345062256,
1172
+ "logps/chosen": -85.92240905761719,
1173
+ "logps/rejected": -147.6622314453125,
1174
+ "loss": 0.0788,
1175
+ "rewards/accuracies": 1.0,
1176
+ "rewards/chosen": -0.7325506210327148,
1177
+ "rewards/margins": 5.368863105773926,
1178
+ "rewards/rejected": -6.101413249969482,
1179
+ "step": 390
1180
+ },
1181
+ {
1182
+ "epoch": 2.1066666666666665,
1183
+ "grad_norm": 2.944322347640991,
1184
+ "learning_rate": 1e-06,
1185
+ "logits/chosen": -0.9857255816459656,
1186
+ "logits/rejected": -0.9889955520629883,
1187
+ "logps/chosen": -86.60405731201172,
1188
+ "logps/rejected": -146.21925354003906,
1189
+ "loss": 0.0604,
1190
+ "rewards/accuracies": 1.0,
1191
+ "rewards/chosen": -0.44693589210510254,
1192
+ "rewards/margins": 5.308945655822754,
1193
+ "rewards/rejected": -5.755881309509277,
1194
+ "step": 395
1195
+ },
1196
+ {
1197
+ "epoch": 2.1333333333333333,
1198
+ "grad_norm": 1.7155674695968628,
1199
+ "learning_rate": 1e-06,
1200
+ "logits/chosen": -0.994691014289856,
1201
+ "logits/rejected": -0.9987820386886597,
1202
+ "logps/chosen": -74.14541625976562,
1203
+ "logps/rejected": -137.02389526367188,
1204
+ "loss": 0.0644,
1205
+ "rewards/accuracies": 1.0,
1206
+ "rewards/chosen": -0.06674417108297348,
1207
+ "rewards/margins": 6.20205020904541,
1208
+ "rewards/rejected": -6.268795013427734,
1209
+ "step": 400
1210
+ },
1211
+ {
1212
+ "epoch": 2.16,
1213
+ "grad_norm": 1.7242580652236938,
1214
+ "learning_rate": 1e-06,
1215
+ "logits/chosen": -0.9776423573493958,
1216
+ "logits/rejected": -0.9695422053337097,
1217
+ "logps/chosen": -79.72328186035156,
1218
+ "logps/rejected": -142.98239135742188,
1219
+ "loss": 0.0439,
1220
+ "rewards/accuracies": 0.949999988079071,
1221
+ "rewards/chosen": -0.8770833015441895,
1222
+ "rewards/margins": 6.5918288230896,
1223
+ "rewards/rejected": -7.4689130783081055,
1224
+ "step": 405
1225
+ },
1226
+ {
1227
+ "epoch": 2.1866666666666665,
1228
+ "grad_norm": 2.5285656452178955,
1229
+ "learning_rate": 1e-06,
1230
+ "logits/chosen": -0.943740963935852,
1231
+ "logits/rejected": -0.9352688789367676,
1232
+ "logps/chosen": -92.97003936767578,
1233
+ "logps/rejected": -145.69248962402344,
1234
+ "loss": 0.0871,
1235
+ "rewards/accuracies": 0.949999988079071,
1236
+ "rewards/chosen": -1.433349370956421,
1237
+ "rewards/margins": 5.817093372344971,
1238
+ "rewards/rejected": -7.250443458557129,
1239
+ "step": 410
1240
+ },
1241
+ {
1242
+ "epoch": 2.2133333333333334,
1243
+ "grad_norm": 6.4492597579956055,
1244
+ "learning_rate": 1e-06,
1245
+ "logits/chosen": -0.9524329304695129,
1246
+ "logits/rejected": -0.9578531384468079,
1247
+ "logps/chosen": -79.1036148071289,
1248
+ "logps/rejected": -126.85892486572266,
1249
+ "loss": 0.0692,
1250
+ "rewards/accuracies": 0.800000011920929,
1251
+ "rewards/chosen": -1.3942753076553345,
1252
+ "rewards/margins": 5.110877990722656,
1253
+ "rewards/rejected": -6.505153656005859,
1254
+ "step": 415
1255
+ },
1256
+ {
1257
+ "epoch": 2.24,
1258
+ "grad_norm": 3.9530835151672363,
1259
+ "learning_rate": 1e-06,
1260
+ "logits/chosen": -0.9325224757194519,
1261
+ "logits/rejected": -0.9292308688163757,
1262
+ "logps/chosen": -92.8379135131836,
1263
+ "logps/rejected": -124.98873138427734,
1264
+ "loss": 0.0744,
1265
+ "rewards/accuracies": 0.949999988079071,
1266
+ "rewards/chosen": -1.4230420589447021,
1267
+ "rewards/margins": 4.427516937255859,
1268
+ "rewards/rejected": -5.850558757781982,
1269
+ "step": 420
1270
+ },
1271
+ {
1272
+ "epoch": 2.2666666666666666,
1273
+ "grad_norm": 3.244112491607666,
1274
+ "learning_rate": 1e-06,
1275
+ "logits/chosen": -0.9953921437263489,
1276
+ "logits/rejected": -0.9901760220527649,
1277
+ "logps/chosen": -90.95811462402344,
1278
+ "logps/rejected": -153.2058563232422,
1279
+ "loss": 0.0505,
1280
+ "rewards/accuracies": 1.0,
1281
+ "rewards/chosen": -1.1016261577606201,
1282
+ "rewards/margins": 5.2358317375183105,
1283
+ "rewards/rejected": -6.33745813369751,
1284
+ "step": 425
1285
+ },
1286
+ {
1287
+ "epoch": 2.2933333333333334,
1288
+ "grad_norm": 3.9486782550811768,
1289
+ "learning_rate": 1e-06,
1290
+ "logits/chosen": -0.9572321176528931,
1291
+ "logits/rejected": -0.963973343372345,
1292
+ "logps/chosen": -107.50956726074219,
1293
+ "logps/rejected": -146.10031127929688,
1294
+ "loss": 0.0639,
1295
+ "rewards/accuracies": 0.949999988079071,
1296
+ "rewards/chosen": -0.6813853979110718,
1297
+ "rewards/margins": 5.3139801025390625,
1298
+ "rewards/rejected": -5.995365619659424,
1299
+ "step": 430
1300
+ },
1301
+ {
1302
+ "epoch": 2.32,
1303
+ "grad_norm": 3.333291530609131,
1304
+ "learning_rate": 1e-06,
1305
+ "logits/chosen": -0.9470894932746887,
1306
+ "logits/rejected": -0.9312503933906555,
1307
+ "logps/chosen": -92.68052673339844,
1308
+ "logps/rejected": -140.1178436279297,
1309
+ "loss": 0.0572,
1310
+ "rewards/accuracies": 1.0,
1311
+ "rewards/chosen": -0.656975507736206,
1312
+ "rewards/margins": 5.908856391906738,
1313
+ "rewards/rejected": -6.565831661224365,
1314
+ "step": 435
1315
+ },
1316
+ {
1317
+ "epoch": 2.3466666666666667,
1318
+ "grad_norm": 9.420110702514648,
1319
+ "learning_rate": 1e-06,
1320
+ "logits/chosen": -0.905168354511261,
1321
+ "logits/rejected": -0.905125617980957,
1322
+ "logps/chosen": -82.85682678222656,
1323
+ "logps/rejected": -134.2998046875,
1324
+ "loss": 0.1159,
1325
+ "rewards/accuracies": 0.8999999761581421,
1326
+ "rewards/chosen": -1.2471545934677124,
1327
+ "rewards/margins": 5.145517826080322,
1328
+ "rewards/rejected": -6.392672538757324,
1329
+ "step": 440
1330
+ },
1331
+ {
1332
+ "epoch": 2.3733333333333335,
1333
+ "grad_norm": 2.3797969818115234,
1334
+ "learning_rate": 1e-06,
1335
+ "logits/chosen": -0.9869009256362915,
1336
+ "logits/rejected": -0.9733754396438599,
1337
+ "logps/chosen": -124.99552917480469,
1338
+ "logps/rejected": -176.72329711914062,
1339
+ "loss": 0.0451,
1340
+ "rewards/accuracies": 1.0,
1341
+ "rewards/chosen": -2.0458180904388428,
1342
+ "rewards/margins": 6.536404609680176,
1343
+ "rewards/rejected": -8.582223892211914,
1344
+ "step": 445
1345
+ },
1346
+ {
1347
+ "epoch": 2.4,
1348
+ "grad_norm": 6.350354194641113,
1349
+ "learning_rate": 1e-06,
1350
+ "logits/chosen": -0.9659305810928345,
1351
+ "logits/rejected": -0.9653702974319458,
1352
+ "logps/chosen": -71.59037017822266,
1353
+ "logps/rejected": -125.0820083618164,
1354
+ "loss": 0.0761,
1355
+ "rewards/accuracies": 0.949999988079071,
1356
+ "rewards/chosen": -0.7029816508293152,
1357
+ "rewards/margins": 5.654323577880859,
1358
+ "rewards/rejected": -6.35730504989624,
1359
+ "step": 450
1360
+ },
1361
+ {
1362
+ "epoch": 2.4266666666666667,
1363
+ "grad_norm": 3.5830423831939697,
1364
+ "learning_rate": 1e-06,
1365
+ "logits/chosen": -0.963516891002655,
1366
+ "logits/rejected": -0.9658119082450867,
1367
+ "logps/chosen": -102.05826568603516,
1368
+ "logps/rejected": -157.795654296875,
1369
+ "loss": 0.0456,
1370
+ "rewards/accuracies": 0.949999988079071,
1371
+ "rewards/chosen": -0.9112430810928345,
1372
+ "rewards/margins": 5.601228713989258,
1373
+ "rewards/rejected": -6.512471675872803,
1374
+ "step": 455
1375
+ },
1376
+ {
1377
+ "epoch": 2.453333333333333,
1378
+ "grad_norm": 1.1380105018615723,
1379
+ "learning_rate": 1e-06,
1380
+ "logits/chosen": -0.9547368884086609,
1381
+ "logits/rejected": -0.9605048298835754,
1382
+ "logps/chosen": -108.07603454589844,
1383
+ "logps/rejected": -171.28262329101562,
1384
+ "loss": 0.0595,
1385
+ "rewards/accuracies": 0.949999988079071,
1386
+ "rewards/chosen": -1.361283779144287,
1387
+ "rewards/margins": 6.42099142074585,
1388
+ "rewards/rejected": -7.782275199890137,
1389
+ "step": 460
1390
+ },
1391
+ {
1392
+ "epoch": 2.48,
1393
+ "grad_norm": 1.1063487529754639,
1394
+ "learning_rate": 1e-06,
1395
+ "logits/chosen": -0.914827823638916,
1396
+ "logits/rejected": -0.9137645959854126,
1397
+ "logps/chosen": -85.6763687133789,
1398
+ "logps/rejected": -160.23040771484375,
1399
+ "loss": 0.0519,
1400
+ "rewards/accuracies": 0.8999999761581421,
1401
+ "rewards/chosen": -1.18534255027771,
1402
+ "rewards/margins": 6.746701240539551,
1403
+ "rewards/rejected": -7.932042598724365,
1404
+ "step": 465
1405
+ },
1406
+ {
1407
+ "epoch": 2.506666666666667,
1408
+ "grad_norm": 1.4177899360656738,
1409
+ "learning_rate": 1e-06,
1410
+ "logits/chosen": -0.941451370716095,
1411
+ "logits/rejected": -0.9365523457527161,
1412
+ "logps/chosen": -77.06974029541016,
1413
+ "logps/rejected": -124.00151062011719,
1414
+ "loss": 0.055,
1415
+ "rewards/accuracies": 0.8500000238418579,
1416
+ "rewards/chosen": -0.9638670086860657,
1417
+ "rewards/margins": 4.646106719970703,
1418
+ "rewards/rejected": -5.609973907470703,
1419
+ "step": 470
1420
+ },
1421
+ {
1422
+ "epoch": 2.533333333333333,
1423
+ "grad_norm": 3.090573787689209,
1424
+ "learning_rate": 1e-06,
1425
+ "logits/chosen": -0.9199577569961548,
1426
+ "logits/rejected": -0.9237100481987,
1427
+ "logps/chosen": -86.69384765625,
1428
+ "logps/rejected": -112.1557846069336,
1429
+ "loss": 0.0502,
1430
+ "rewards/accuracies": 0.949999988079071,
1431
+ "rewards/chosen": -1.1065205335617065,
1432
+ "rewards/margins": 4.314031600952148,
1433
+ "rewards/rejected": -5.4205522537231445,
1434
+ "step": 475
1435
+ },
1436
+ {
1437
+ "epoch": 2.56,
1438
+ "grad_norm": 7.529189109802246,
1439
+ "learning_rate": 1e-06,
1440
+ "logits/chosen": -0.9499224424362183,
1441
+ "logits/rejected": -0.9461296200752258,
1442
+ "logps/chosen": -98.30276489257812,
1443
+ "logps/rejected": -169.53909301757812,
1444
+ "loss": 0.0692,
1445
+ "rewards/accuracies": 0.949999988079071,
1446
+ "rewards/chosen": -0.8648740649223328,
1447
+ "rewards/margins": 7.417494773864746,
1448
+ "rewards/rejected": -8.282369613647461,
1449
+ "step": 480
1450
+ },
1451
+ {
1452
+ "epoch": 2.586666666666667,
1453
+ "grad_norm": 4.119349479675293,
1454
+ "learning_rate": 1e-06,
1455
+ "logits/chosen": -0.9638736844062805,
1456
+ "logits/rejected": -0.9807848930358887,
1457
+ "logps/chosen": -82.62672424316406,
1458
+ "logps/rejected": -177.9151611328125,
1459
+ "loss": 0.0633,
1460
+ "rewards/accuracies": 1.0,
1461
+ "rewards/chosen": -1.7524802684783936,
1462
+ "rewards/margins": 7.268258094787598,
1463
+ "rewards/rejected": -9.02073860168457,
1464
+ "step": 485
1465
+ },
1466
+ {
1467
+ "epoch": 2.6133333333333333,
1468
+ "grad_norm": 3.560368299484253,
1469
+ "learning_rate": 1e-06,
1470
+ "logits/chosen": -0.9251823425292969,
1471
+ "logits/rejected": -0.9247922897338867,
1472
+ "logps/chosen": -87.5182876586914,
1473
+ "logps/rejected": -169.91879272460938,
1474
+ "loss": 0.1587,
1475
+ "rewards/accuracies": 0.949999988079071,
1476
+ "rewards/chosen": -1.2214720249176025,
1477
+ "rewards/margins": 6.715635776519775,
1478
+ "rewards/rejected": -7.937108039855957,
1479
+ "step": 490
1480
+ },
1481
+ {
1482
+ "epoch": 2.64,
1483
+ "grad_norm": 1.6422832012176514,
1484
+ "learning_rate": 1e-06,
1485
+ "logits/chosen": -0.9349163174629211,
1486
+ "logits/rejected": -0.9354310035705566,
1487
+ "logps/chosen": -70.24932861328125,
1488
+ "logps/rejected": -146.46348571777344,
1489
+ "loss": 0.0691,
1490
+ "rewards/accuracies": 0.949999988079071,
1491
+ "rewards/chosen": -0.6841042637825012,
1492
+ "rewards/margins": 6.8183794021606445,
1493
+ "rewards/rejected": -7.502483367919922,
1494
+ "step": 495
1495
+ },
1496
+ {
1497
+ "epoch": 2.6666666666666665,
1498
+ "grad_norm": 2.153717041015625,
1499
+ "learning_rate": 1e-06,
1500
+ "logits/chosen": -0.9372892379760742,
1501
+ "logits/rejected": -0.9442557096481323,
1502
+ "logps/chosen": -74.71461486816406,
1503
+ "logps/rejected": -149.56219482421875,
1504
+ "loss": 0.0355,
1505
+ "rewards/accuracies": 0.949999988079071,
1506
+ "rewards/chosen": -0.513080894947052,
1507
+ "rewards/margins": 6.21486759185791,
1508
+ "rewards/rejected": -6.727948188781738,
1509
+ "step": 500
1510
+ },
1511
+ {
1512
+ "epoch": 2.6933333333333334,
1513
+ "grad_norm": 2.593008279800415,
1514
+ "learning_rate": 1e-06,
1515
+ "logits/chosen": -0.9294622540473938,
1516
+ "logits/rejected": -0.9304571151733398,
1517
+ "logps/chosen": -90.94770050048828,
1518
+ "logps/rejected": -147.8937225341797,
1519
+ "loss": 0.0517,
1520
+ "rewards/accuracies": 0.949999988079071,
1521
+ "rewards/chosen": -1.548829197883606,
1522
+ "rewards/margins": 6.368669033050537,
1523
+ "rewards/rejected": -7.917498588562012,
1524
+ "step": 505
1525
+ },
1526
+ {
1527
+ "epoch": 2.7199999999999998,
1528
+ "grad_norm": 12.588436126708984,
1529
+ "learning_rate": 1e-06,
1530
+ "logits/chosen": -0.9092562794685364,
1531
+ "logits/rejected": -0.9060165286064148,
1532
+ "logps/chosen": -83.18638610839844,
1533
+ "logps/rejected": -130.56942749023438,
1534
+ "loss": 0.0565,
1535
+ "rewards/accuracies": 0.949999988079071,
1536
+ "rewards/chosen": -1.80876886844635,
1537
+ "rewards/margins": 5.5646467208862305,
1538
+ "rewards/rejected": -7.373415946960449,
1539
+ "step": 510
1540
+ },
1541
+ {
1542
+ "epoch": 2.7466666666666666,
1543
+ "grad_norm": 3.6610798835754395,
1544
+ "learning_rate": 1e-06,
1545
+ "logits/chosen": -1.00751531124115,
1546
+ "logits/rejected": -1.0096744298934937,
1547
+ "logps/chosen": -91.13447570800781,
1548
+ "logps/rejected": -148.82164001464844,
1549
+ "loss": 0.0285,
1550
+ "rewards/accuracies": 1.0,
1551
+ "rewards/chosen": -1.2526090145111084,
1552
+ "rewards/margins": 6.453673362731934,
1553
+ "rewards/rejected": -7.706282138824463,
1554
+ "step": 515
1555
+ },
1556
+ {
1557
+ "epoch": 2.7733333333333334,
1558
+ "grad_norm": 1.0663306713104248,
1559
+ "learning_rate": 1e-06,
1560
+ "logits/chosen": -0.9299656748771667,
1561
+ "logits/rejected": -0.9350908994674683,
1562
+ "logps/chosen": -94.07073211669922,
1563
+ "logps/rejected": -164.70330810546875,
1564
+ "loss": 0.0384,
1565
+ "rewards/accuracies": 1.0,
1566
+ "rewards/chosen": -1.654295563697815,
1567
+ "rewards/margins": 6.949212074279785,
1568
+ "rewards/rejected": -8.603507995605469,
1569
+ "step": 520
1570
+ },
1571
+ {
1572
+ "epoch": 2.8,
1573
+ "grad_norm": 6.753617286682129,
1574
+ "learning_rate": 1e-06,
1575
+ "logits/chosen": -0.9451917409896851,
1576
+ "logits/rejected": -0.9450856447219849,
1577
+ "logps/chosen": -101.37158203125,
1578
+ "logps/rejected": -149.12391662597656,
1579
+ "loss": 0.0611,
1580
+ "rewards/accuracies": 0.949999988079071,
1581
+ "rewards/chosen": -2.0039877891540527,
1582
+ "rewards/margins": 6.462770938873291,
1583
+ "rewards/rejected": -8.46675968170166,
1584
+ "step": 525
1585
+ },
1586
+ {
1587
+ "epoch": 2.8266666666666667,
1588
+ "grad_norm": 9.86711597442627,
1589
+ "learning_rate": 1e-06,
1590
+ "logits/chosen": -0.9280322194099426,
1591
+ "logits/rejected": -0.9299230575561523,
1592
+ "logps/chosen": -94.41612243652344,
1593
+ "logps/rejected": -168.96092224121094,
1594
+ "loss": 0.0731,
1595
+ "rewards/accuracies": 0.949999988079071,
1596
+ "rewards/chosen": -1.973374366760254,
1597
+ "rewards/margins": 6.930324554443359,
1598
+ "rewards/rejected": -8.90369987487793,
1599
+ "step": 530
1600
+ },
1601
+ {
1602
+ "epoch": 2.8533333333333335,
1603
+ "grad_norm": 10.411066055297852,
1604
+ "learning_rate": 1e-06,
1605
+ "logits/chosen": -0.9307094812393188,
1606
+ "logits/rejected": -0.9296162724494934,
1607
+ "logps/chosen": -87.57542419433594,
1608
+ "logps/rejected": -134.86257934570312,
1609
+ "loss": 0.0633,
1610
+ "rewards/accuracies": 0.949999988079071,
1611
+ "rewards/chosen": -1.4023369550704956,
1612
+ "rewards/margins": 5.565145969390869,
1613
+ "rewards/rejected": -6.967482566833496,
1614
+ "step": 535
1615
+ },
1616
+ {
1617
+ "epoch": 2.88,
1618
+ "grad_norm": 1.9066485166549683,
1619
+ "learning_rate": 1e-06,
1620
+ "logits/chosen": -0.9522913098335266,
1621
+ "logits/rejected": -0.9543063044548035,
1622
+ "logps/chosen": -90.59437561035156,
1623
+ "logps/rejected": -152.72274780273438,
1624
+ "loss": 0.1226,
1625
+ "rewards/accuracies": 0.949999988079071,
1626
+ "rewards/chosen": -2.1479926109313965,
1627
+ "rewards/margins": 6.265242576599121,
1628
+ "rewards/rejected": -8.41323471069336,
1629
+ "step": 540
1630
+ },
1631
+ {
1632
+ "epoch": 2.9066666666666667,
1633
+ "grad_norm": 1.3979381322860718,
1634
+ "learning_rate": 1e-06,
1635
+ "logits/chosen": -0.9681941866874695,
1636
+ "logits/rejected": -0.9737881422042847,
1637
+ "logps/chosen": -101.29634094238281,
1638
+ "logps/rejected": -180.4349822998047,
1639
+ "loss": 0.0586,
1640
+ "rewards/accuracies": 0.949999988079071,
1641
+ "rewards/chosen": -2.112703323364258,
1642
+ "rewards/margins": 6.271733283996582,
1643
+ "rewards/rejected": -8.38443660736084,
1644
+ "step": 545
1645
+ },
1646
+ {
1647
+ "epoch": 2.9333333333333336,
1648
+ "grad_norm": 7.896867275238037,
1649
+ "learning_rate": 1e-06,
1650
+ "logits/chosen": -0.9267778396606445,
1651
+ "logits/rejected": -0.9226524233818054,
1652
+ "logps/chosen": -86.86418914794922,
1653
+ "logps/rejected": -140.9725341796875,
1654
+ "loss": 0.0609,
1655
+ "rewards/accuracies": 1.0,
1656
+ "rewards/chosen": -1.6871792078018188,
1657
+ "rewards/margins": 5.857264995574951,
1658
+ "rewards/rejected": -7.544443607330322,
1659
+ "step": 550
1660
+ },
1661
+ {
1662
+ "epoch": 2.96,
1663
+ "grad_norm": 2.1880979537963867,
1664
+ "learning_rate": 1e-06,
1665
+ "logits/chosen": -0.9265161752700806,
1666
+ "logits/rejected": -0.9284146428108215,
1667
+ "logps/chosen": -86.45133972167969,
1668
+ "logps/rejected": -158.9849853515625,
1669
+ "loss": 0.0429,
1670
+ "rewards/accuracies": 1.0,
1671
+ "rewards/chosen": -1.7444225549697876,
1672
+ "rewards/margins": 7.387899875640869,
1673
+ "rewards/rejected": -9.132322311401367,
1674
+ "step": 555
1675
+ },
1676
+ {
1677
+ "epoch": 2.986666666666667,
1678
+ "grad_norm": 0.5332115888595581,
1679
+ "learning_rate": 1e-06,
1680
+ "logits/chosen": -0.9316242933273315,
1681
+ "logits/rejected": -0.9294928312301636,
1682
+ "logps/chosen": -89.85722351074219,
1683
+ "logps/rejected": -164.3010711669922,
1684
+ "loss": 0.0779,
1685
+ "rewards/accuracies": 1.0,
1686
+ "rewards/chosen": -1.4647209644317627,
1687
+ "rewards/margins": 7.412230491638184,
1688
+ "rewards/rejected": -8.876951217651367,
1689
+ "step": 560
1690
+ },
1691
+ {
1692
+ "epoch": 2.992,
1693
+ "step": 561,
1694
+ "total_flos": 0.0,
1695
+ "train_loss": 0.24264521944982376,
1696
+ "train_runtime": 1521.7244,
1697
+ "train_samples_per_second": 11.827,
1698
+ "train_steps_per_second": 0.369
1699
+ }
1700
+ ],
1701
+ "logging_steps": 5,
1702
+ "max_steps": 561,
1703
+ "num_input_tokens_seen": 0,
1704
+ "num_train_epochs": 3,
1705
+ "save_steps": 500,
1706
+ "stateful_callbacks": {
1707
+ "TrainerControl": {
1708
+ "args": {
1709
+ "should_epoch_stop": false,
1710
+ "should_evaluate": false,
1711
+ "should_log": false,
1712
+ "should_save": false,
1713
+ "should_training_stop": false
1714
+ },
1715
+ "attributes": {}
1716
+ }
1717
+ },
1718
+ "total_flos": 0.0,
1719
+ "train_batch_size": 2,
1720
+ "trial_name": null,
1721
+ "trial_params": null
1722
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51de14b19924dcc5d2c57da516ff26e0a60d43c8dff8e2e5af86c6e68c41881d
3
+ size 5560