beamaia commited on
Commit
f8c44d0
1 Parent(s): da20946

Training in progress, step 100, checkpoint

Browse files
checkpoint-100/adapter_config.json CHANGED
@@ -20,10 +20,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_proj",
24
- "k_proj",
25
  "v_proj",
26
- "o_proj"
 
 
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
23
  "v_proj",
24
+ "o_proj",
25
+ "k_proj",
26
+ "q_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
checkpoint-100/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d1d26bc61e21d0b93fd9b446c555c662c0dced2ed04a45c6e79b26eeb3ee52e9
3
  size 6849416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7cd70ce8242302b84492e870615155acf7a71718e051dfbaf76425667d74d71
3
  size 6849416
checkpoint-100/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a3d921e02a11aa9aed101e1624589a67a8ae88f444b6f2746ea10128ca55dc8
3
  size 13846266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cf2261968b8d2e33b229a1338632b305db9baac4defb430ea87ed364d9acf5f
3
  size 13846266
checkpoint-100/trainer_state.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
- "best_metric": 0.3468475937843323,
3
- "best_model_checkpoint": "./zephyr/08-04-24-Weni-WeniGPT-Agents-Zephyr-1.0.18-KTO_Hyperparameter search, altering lora params for KTO task.-2_max_steps-145_batch_16_2024-04-08_ppid_9/checkpoint-100",
4
  "epoch": 0.684931506849315,
5
  "eval_steps": 50,
6
  "global_step": 100,
@@ -10,95 +10,95 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.14,
13
- "grad_norm": 3.3949286937713623,
14
  "learning_rate": 0.00018142857142857142,
15
- "loss": 0.4993,
16
  "step": 20,
17
- "train/kl": 2.0773630142211914,
18
- "train/logps/chosen": -280.7225560897436,
19
- "train/logps/rejected": -280.629668445122,
20
- "train/rewards/chosen": -0.4658302894005409,
21
- "train/rewards/margins": 0.008156592134686835,
22
- "train/rewards/rejected": -0.4739868815352277
23
  },
24
  {
25
  "epoch": 0.27,
26
- "grad_norm": 2.1768083572387695,
27
  "learning_rate": 0.00015285714285714287,
28
- "loss": 0.412,
29
  "step": 40,
30
- "train/kl": 11.56242561340332,
31
- "train/logps/chosen": -279.2820556640625,
32
- "train/logps/rejected": -285.233056640625,
33
- "train/rewards/chosen": 1.6047859191894531,
34
- "train/rewards/margins": 1.591878777742386,
35
- "train/rewards/rejected": 0.01290714144706726
36
  },
37
  {
38
  "epoch": 0.34,
39
- "eval/kl": 34.189842224121094,
40
- "eval/logps/chosen": -240.1980633802817,
41
- "eval/logps/rejected": -241.65251681170886,
42
- "eval/rewards/chosen": 4.346213757152289,
43
- "eval/rewards/margins": 2.0745202015376725,
44
- "eval/rewards/rejected": 2.2716935556146165,
45
- "eval_loss": 0.3987465500831604,
46
- "eval_runtime": 139.8252,
47
- "eval_samples_per_second": 2.146,
48
- "eval_steps_per_second": 0.536,
49
  "step": 50
50
  },
51
  {
52
  "epoch": 0.41,
53
- "grad_norm": 3.7959017753601074,
54
  "learning_rate": 0.00012428571428571428,
55
- "loss": 0.409,
56
  "step": 60,
57
- "train/kl": 25.960163116455078,
58
- "train/logps/chosen": -228.58444552951389,
59
- "train/logps/rejected": -259.24209872159093,
60
- "train/rewards/chosen": 3.7380718655056424,
61
- "train/rewards/margins": 2.2719342395512747,
62
- "train/rewards/rejected": 1.466137625954368
63
  },
64
  {
65
  "epoch": 0.55,
66
- "grad_norm": 5.033076763153076,
67
  "learning_rate": 9.571428571428573e-05,
68
- "loss": 0.3576,
69
  "step": 80,
70
- "train/kl": 5.953394412994385,
71
- "train/logps/chosen": -279.7611678685897,
72
- "train/logps/rejected": -306.2880144817073,
73
- "train/rewards/chosen": 1.7610896183894231,
74
- "train/rewards/margins": 3.6376326169126703,
75
- "train/rewards/rejected": -1.876542998523247
76
  },
77
  {
78
  "epoch": 0.68,
79
- "grad_norm": 6.184478759765625,
80
  "learning_rate": 6.714285714285714e-05,
81
- "loss": 0.3304,
82
  "step": 100,
83
- "train/kl": 0.5991002321243286,
84
- "train/logps/chosen": -307.0083233173077,
85
- "train/logps/rejected": -345.8705592105263,
86
- "train/rewards/chosen": -1.1401430570162259,
87
- "train/rewards/margins": 4.161937278098906,
88
- "train/rewards/rejected": -5.302080335115131
89
  },
90
  {
91
  "epoch": 0.68,
92
- "eval/kl": 3.7238929271698,
93
- "eval/logps/chosen": -265.10247029049293,
94
- "eval/logps/rejected": -277.03555181962025,
95
- "eval/rewards/chosen": 1.855774355606294,
96
- "eval/rewards/margins": 3.1223836440754393,
97
- "eval/rewards/rejected": -1.2666092884691456,
98
- "eval_loss": 0.3468475937843323,
99
- "eval_runtime": 139.8048,
100
- "eval_samples_per_second": 2.146,
101
- "eval_steps_per_second": 0.536,
102
  "step": 100
103
  }
104
  ],
 
1
  {
2
+ "best_metric": 0.35275527834892273,
3
+ "best_model_checkpoint": "./zephyr/09-04-24-Weni-WeniGPT-Agents-Zephyr-1.0.18-KTO_Hyperparameter search, altering lora params for KTO task.-2_max_steps-145_batch_16_2024-04-09_ppid_9/checkpoint-100",
4
  "epoch": 0.684931506849315,
5
  "eval_steps": 50,
6
  "global_step": 100,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.14,
13
+ "grad_norm": 6.346695899963379,
14
  "learning_rate": 0.00018142857142857142,
15
+ "loss": 0.467,
16
  "step": 20,
17
+ "train/kl": 6.117425918579102,
18
+ "train/logps/chosen": -259.1842447916667,
19
+ "train/logps/rejected": -290.1948988970588,
20
+ "train/rewards/chosen": 0.930299072265625,
21
+ "train/rewards/margins": 0.5634524266860065,
22
+ "train/rewards/rejected": 0.36684664557961855
23
  },
24
  {
25
  "epoch": 0.27,
26
+ "grad_norm": 4.604153633117676,
27
  "learning_rate": 0.00015285714285714287,
28
+ "loss": 0.4224,
29
  "step": 40,
30
+ "train/kl": 6.080809116363525,
31
+ "train/logps/chosen": -274.32459677419354,
32
+ "train/logps/rejected": -291.6558948863636,
33
+ "train/rewards/chosen": 0.9791939027847782,
34
+ "train/rewards/margins": 1.6478286295692244,
35
+ "train/rewards/rejected": -0.668634726784446
36
  },
37
  {
38
  "epoch": 0.34,
39
+ "eval/kl": 7.511639595031738,
40
+ "eval/logps/chosen": -263.8732394366197,
41
+ "eval/logps/rejected": -268.6064082278481,
42
+ "eval/rewards/chosen": 1.9797810299295775,
43
+ "eval/rewards/margins": 2.403955568071142,
44
+ "eval/rewards/rejected": -0.42417453814156447,
45
+ "eval_loss": 0.3916032016277313,
46
+ "eval_runtime": 141.678,
47
+ "eval_samples_per_second": 2.117,
48
+ "eval_steps_per_second": 0.529,
49
  "step": 50
50
  },
51
  {
52
  "epoch": 0.41,
53
+ "grad_norm": 5.5113444328308105,
54
  "learning_rate": 0.00012428571428571428,
55
+ "loss": 0.3832,
56
  "step": 60,
57
+ "train/kl": 16.34114646911621,
58
+ "train/logps/chosen": -240.63917267628204,
59
+ "train/logps/rejected": -277.463486089939,
60
+ "train/rewards/chosen": 3.0404166197165465,
61
+ "train/rewards/margins": 2.3469540618075815,
62
+ "train/rewards/rejected": 0.6934625579089653
63
  },
64
  {
65
  "epoch": 0.55,
66
+ "grad_norm": 2.7619431018829346,
67
  "learning_rate": 9.571428571428573e-05,
68
+ "loss": 0.3233,
69
  "step": 80,
70
+ "train/kl": 0.7871202826499939,
71
+ "train/logps/chosen": -294.5624213506711,
72
+ "train/logps/rejected": -336.1468612938597,
73
+ "train/rewards/chosen": 0.5479572987396445,
74
+ "train/rewards/margins": 5.502926202933724,
75
+ "train/rewards/rejected": -4.954968904194079
76
  },
77
  {
78
  "epoch": 0.68,
79
+ "grad_norm": 4.996425151824951,
80
  "learning_rate": 6.714285714285714e-05,
81
+ "loss": 0.3749,
82
  "step": 100,
83
+ "train/kl": 6.590612888336182,
84
+ "train/logps/chosen": -265.2372325922819,
85
+ "train/logps/rejected": -300.5130665204678,
86
+ "train/rewards/chosen": 2.635832920970533,
87
+ "train/rewards/margins": 3.7929440163766914,
88
+ "train/rewards/rejected": -1.1571110954061585
89
  },
90
  {
91
  "epoch": 0.68,
92
+ "eval/kl": 2.193509817123413,
93
+ "eval/logps/chosen": -268.1729478433099,
94
+ "eval/logps/rejected": -291.57960838607596,
95
+ "eval/rewards/chosen": 1.5498130690883583,
96
+ "eval/rewards/margins": 4.271308299075008,
97
+ "eval/rewards/rejected": -2.7214952299866497,
98
+ "eval_loss": 0.35275527834892273,
99
+ "eval_runtime": 141.6617,
100
+ "eval_samples_per_second": 2.118,
101
+ "eval_steps_per_second": 0.529,
102
  "step": 100
103
  }
104
  ],
checkpoint-100/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:38ae2e87f29eb2d5b20b8a1633c876370c1a196bac5b23847645ad024eea5fb3
3
  size 5688
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab74b0678263b82e3cef84b95bcb8d0582ce59316d01cb7ca3ea92557a3bcdfe
3
  size 5688