hasibirok0 commited on
Commit
dda2274
·
verified ·
1 Parent(s): e6a192e

Model save

Browse files
README.md CHANGED
@@ -36,11 +36,11 @@ More information needed
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 0.0002
39
- - train_batch_size: 4
40
  - eval_batch_size: 8
41
  - seed: 42
42
  - gradient_accumulation_steps: 4
43
- - total_train_batch_size: 16
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
  - lr_scheduler_warmup_steps: 20
 
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 0.0002
39
+ - train_batch_size: 32
40
  - eval_batch_size: 8
41
  - seed: 42
42
  - gradient_accumulation_steps: 4
43
+ - total_train_batch_size: 128
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
  - lr_scheduler_warmup_steps: 20
model_fold_0_batch0/adapter_config.json CHANGED
@@ -40,9 +40,9 @@
40
  "rank_pattern": {},
41
  "revision": null,
42
  "target_modules": [
43
- "k_proj",
44
  "q_proj",
45
- "v_proj"
46
  ],
47
  "task_type": "SEQ_CLS",
48
  "use_dora": false,
 
40
  "rank_pattern": {},
41
  "revision": null,
42
  "target_modules": [
43
+ "v_proj",
44
  "q_proj",
45
+ "k_proj"
46
  ],
47
  "task_type": "SEQ_CLS",
48
  "use_dora": false,
model_fold_0_batch0/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:835c437387d81519e42e460a80666370769d2657c67434b42e0ba960ffa64056
3
  size 4780504
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2e88a260e9d2641c42dfd7058e465462713bcdae4023b5d890356f1e64a778f
3
  size 4780504
model_fold_0_batch0/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e45a8de8819d4c04f3cdeed9685c2d41fe73a64e3b265506968b9f80ebf26556
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b093f9a99f5bcfbe28a31dcc057e5e9dff06005a60705aed8ee673e56a7ec907
3
  size 5496