Hanzalwi commited on
Commit
0f87e5f
1 Parent(s): 3fa806e

Training in progress, step 300, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -16,8 +16,8 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "q_proj",
20
- "v_proj"
21
  ],
22
  "task_type": "CAUSAL_LM"
23
  }
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
+ "v_proj",
20
+ "q_proj"
21
  ],
22
  "task_type": "CAUSAL_LM"
23
  }
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:867173bcaddd4298f59f7d72b16f31a68f27518e15f271a9040271ca66f6319e
3
  size 12595704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c8b46775cd1518360b0322d9acadb985d980bc6360007c1b29cefe44d9dd337
3
  size 12595704
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f4b1d6ed1999ea1b1bbcec41b01bea8534d890f34f41175ccb3137f2d3ef69cc
3
- size 25222277
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4b65442bbf6d1467cc9417ee8a2d402be15409442f366e77e6f6ef6d1af2390
3
+ size 25222341
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:85658496e06077f87b750aeb031c8edf46d72979c48ec879f8596b632cba3381
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cfb045d2aace0aa6b48f6b2659a883d08eab7c39a2384b021a9dd36bdce66b0
3
  size 14575
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7aa2c8b84e17817e6a4dcba5955fca913e266fdcd47f5594a29933ebd4972a01
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3e02e4b8a74e841ee5a1ce8c453a9becb15f5be6cbcdd100b814e93235dab36
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 1.1465147733688354,
3
- "best_model_checkpoint": "./outputs/checkpoint-200",
4
- "epoch": 0.26666666666666666,
5
  "eval_steps": 100,
6
- "global_step": 200,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -35,6 +35,20 @@
35
  "eval_samples_per_second": 10.395,
36
  "eval_steps_per_second": 1.303,
37
  "step": 200
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38
  }
39
  ],
40
  "logging_steps": 100,
@@ -42,7 +56,7 @@
42
  "num_input_tokens_seen": 0,
43
  "num_train_epochs": 3,
44
  "save_steps": 100,
45
- "total_flos": 2.8242451495747584e+16,
46
  "trial_name": null,
47
  "trial_params": null
48
  }
 
1
  {
2
+ "best_metric": 1.1323238611221313,
3
+ "best_model_checkpoint": "./outputs/checkpoint-300",
4
+ "epoch": 0.4,
5
  "eval_steps": 100,
6
+ "global_step": 300,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
35
  "eval_samples_per_second": 10.395,
36
  "eval_steps_per_second": 1.303,
37
  "step": 200
38
+ },
39
+ {
40
+ "epoch": 0.4,
41
+ "learning_rate": 0.0002,
42
+ "loss": 0.9916,
43
+ "step": 300
44
+ },
45
+ {
46
+ "epoch": 0.4,
47
+ "eval_loss": 1.1323238611221313,
48
+ "eval_runtime": 186.616,
49
+ "eval_samples_per_second": 10.342,
50
+ "eval_steps_per_second": 1.297,
51
+ "step": 300
52
  }
53
  ],
54
  "logging_steps": 100,
 
56
  "num_input_tokens_seen": 0,
57
  "num_train_epochs": 3,
58
  "save_steps": 100,
59
+ "total_flos": 4.242711954033869e+16,
60
  "trial_name": null,
61
  "trial_params": null
62
  }
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef6793b9c786624948fe20bfafe4d99e555c2e1c3dcf3c8e95e0213265b53906
3
  size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a44655c832d56011e28a0581b59475c0dfdadb4c68de0ecc5ac3364c1002507d
3
  size 4283