duyvt6663 commited on
Commit
7ad7bca
·
1 Parent(s): c8ed599

Training in progress, step 50, checkpoint

Browse files
checkpoint-50/README.md CHANGED
@@ -216,4 +216,23 @@ The following `bitsandbytes` quantization config was used during training:
216
  ### Framework versions
217
 
218
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
219
  - PEFT 0.6.0
 
216
  ### Framework versions
217
 
218
 
219
+ - PEFT 0.6.0
220
+ ## Training procedure
221
+
222
+
223
+ The following `bitsandbytes` quantization config was used during training:
224
+ - quant_method: bitsandbytes
225
+ - load_in_8bit: True
226
+ - load_in_4bit: False
227
+ - llm_int8_threshold: 6.0
228
+ - llm_int8_skip_modules: None
229
+ - llm_int8_enable_fp32_cpu_offload: False
230
+ - llm_int8_has_fp16_weight: False
231
+ - bnb_4bit_quant_type: fp4
232
+ - bnb_4bit_use_double_quant: False
233
+ - bnb_4bit_compute_dtype: float32
234
+
235
+ ### Framework versions
236
+
237
+
238
  - PEFT 0.6.0
checkpoint-50/adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "/kaggle/input/vietcuna-for-binary-classification/kalapa-vietcuna-3b",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -9,10 +9,10 @@
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "lora_alpha": 32,
12
- "lora_dropout": 0.1,
13
  "modules_to_save": null,
14
  "peft_type": "LORA",
15
- "r": 8,
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "vietcuna-3b-v2/kalapa-vietcuna-3b/",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "lora_alpha": 32,
12
+ "lora_dropout": 0.05,
13
  "modules_to_save": null,
14
  "peft_type": "LORA",
15
+ "r": 16,
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
checkpoint-50/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c0e2c08e5163bb3f67a5ce7acd4dd75e8daa2a7869bb3a6fd5bbe81296f5a646
3
- size 9859800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4166664b368ce4cd6e33f9c9e5988f4cd0f67acb476bfaf70382c8384056788a
3
+ size 19690328
checkpoint-50/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:649f6f10acc825530de20a2818f6a35ceaec27f008838bbd89d1eafa771845a3
3
- size 42724
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b45841b33a84da2a1b486d9551765151b191da5b910c16259aacd2114b08f224
3
+ size 38087098
checkpoint-50/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:be8ed0ba649837443f79e95d66f2cdebd362fce4049ee79d5b9831b2c22ccf76
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f197d949afeca5f73d2f9c4e8008213842dfb0a96b35f9276fd815dc3b955de7
3
  size 14244
checkpoint-50/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a6a36ba0ea017dad0467a59150d5a3a03523965e4b322c77d048735654bcbefb
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:609839f04f1b0fbf1991d0fad32a3ffd669d0403ec0a02950e111710295f6725
3
  size 1064
checkpoint-50/special_tokens_map.json CHANGED
@@ -2,6 +2,6 @@
2
  "bos_token": "<s>",
3
  "eos_token": "</s>",
4
  "pad_token": "<pad>",
5
- "sep_token": "<s>",
6
  "unk_token": "<unk>"
7
  }
 
2
  "bos_token": "<s>",
3
  "eos_token": "</s>",
4
  "pad_token": "<pad>",
5
+ "sep_token": "\n\n",
6
  "unk_token": "<unk>"
7
  }
checkpoint-50/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
checkpoint-50/tokenizer_config.json CHANGED
@@ -39,7 +39,7 @@
39
  "eos_token": "</s>",
40
  "model_max_length": 1000000000000000019884624838656,
41
  "pad_token": "<pad>",
42
- "sep_token": "<s>",
43
  "tokenizer_class": "BloomTokenizer",
44
  "unk_token": "<unk>"
45
  }
 
39
  "eos_token": "</s>",
40
  "model_max_length": 1000000000000000019884624838656,
41
  "pad_token": "<pad>",
42
+ "sep_token": "\n\n",
43
  "tokenizer_class": "BloomTokenizer",
44
  "unk_token": "<unk>"
45
  }
checkpoint-50/trainer_state.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "best_metric": 0.522564172744751,
3
  "best_model_checkpoint": "output/checkpoint-50",
4
- "epoch": 0.34904013961605584,
5
  "eval_steps": 50,
6
  "global_step": 50,
7
  "is_hyper_param_search": false,
@@ -10,31 +10,31 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.01,
13
- "learning_rate": 1.0000000000000001e-07,
14
- "loss": 0.4196,
15
  "step": 1
16
  },
17
  {
18
- "epoch": 0.35,
19
- "learning_rate": 4.6e-06,
20
- "loss": 0.4279,
21
  "step": 50
22
  },
23
  {
24
- "epoch": 0.35,
25
- "eval_accuracy": 0.7657142857142857,
26
- "eval_loss": 0.522564172744751,
27
- "eval_runtime": 29.0326,
28
- "eval_samples_per_second": 6.028,
29
- "eval_steps_per_second": 1.516,
30
  "step": 50
31
  }
32
  ],
33
  "logging_steps": 50,
34
- "max_steps": 500,
35
- "num_train_epochs": 4,
36
  "save_steps": 50,
37
- "total_flos": 1.891662110539776e+16,
38
  "trial_name": null,
39
  "trial_params": null
40
  }
 
1
  {
2
+ "best_metric": 6.581231117248535,
3
  "best_model_checkpoint": "output/checkpoint-50",
4
+ "epoch": 0.2552648372686662,
5
  "eval_steps": 50,
6
  "global_step": 50,
7
  "is_hyper_param_search": false,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.01,
13
+ "learning_rate": 2.777777777777778e-07,
14
+ "loss": 9.5569,
15
  "step": 1
16
  },
17
  {
18
+ "epoch": 0.26,
19
+ "learning_rate": 1.25e-05,
20
+ "loss": 6.958,
21
  "step": 50
22
  },
23
  {
24
+ "epoch": 0.26,
25
+ "eval_accuracy": 0.42857142857142855,
26
+ "eval_loss": 6.581231117248535,
27
+ "eval_runtime": 28.2966,
28
+ "eval_samples_per_second": 6.184,
29
+ "eval_steps_per_second": 1.555,
30
  "step": 50
31
  }
32
  ],
33
  "logging_steps": 50,
34
+ "max_steps": 1000,
35
+ "num_train_epochs": 6,
36
  "save_steps": 50,
37
+ "total_flos": 1.854149272129536e+16,
38
  "trial_name": null,
39
  "trial_params": null
40
  }
checkpoint-50/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1b0a94354c0f3dfee9c93e225dbd48ad6f856f98977ef38ff3415910328e64a
3
  size 4600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a0ef36d7c31265f68a3edc8ae53ce9272dcc90a79f261e29a2a46e040c77a33
3
  size 4600