Trace2333 commited on
Commit
764f6c5
1 Parent(s): b67686a

Upload folder using huggingface_hub

Browse files
adapter_config.json CHANGED
@@ -10,13 +10,13 @@
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
- "lora_alpha": 128,
14
  "lora_dropout": 0.05,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
- "r": 64,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
+ "lora_alpha": 256,
14
  "lora_dropout": 0.05,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
+ "r": 128,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c48cb05c67bf41c2af83c72ccfe523351ecca46abbc4bce6afbe009b9e84feae
3
- size 100733002
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86e817f53b90e2acbc47ab21f81b0adaa0b0455b71456c3f95227bf66a19ea14
3
+ size 201396298
checkpoint-4000/adapter_config.json CHANGED
@@ -10,19 +10,19 @@
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
- "lora_alpha": 128,
14
  "lora_dropout": 0.05,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
- "r": 64,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
  "q_proj",
24
- "v_proj",
25
- "k_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
 
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
+ "lora_alpha": 256,
14
  "lora_dropout": 0.05,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
+ "r": 128,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
  "q_proj",
24
+ "k_proj",
25
+ "v_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
checkpoint-4000/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e53bfad9dca96f934e2680455db914abbf960ae3dd89fbd1b00b3b9e3b5d3add
3
- size 100689344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60b4cd08ef271a49996ec4eec8fb1a4ee566bbcc6a2044e102650927c7c09668
3
+ size 201353048
checkpoint-4000/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c460cffe1c96697bddb93f322307c8d0a7539ac8052e11582e900a302058d831
3
- size 201488570
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:911b2e0d9926846486752e434911d475aed4356a916ce82a98d944df3cbedc5a
3
+ size 402815162
checkpoint-4000/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c8a5cc3d153afab4a6d4bbcb5da0d17f103aee9d0b00b17dda415d486d2d1f51
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3d436aff47ad46c88070ead13f8260f1feb126c4a3f787bc8d32902754201b0
3
  size 14244
checkpoint-4000/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d765a7c1042c67d780e30984f9532deaeaa178dcc686b873b8d93b468ca0b826
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05e74c150ec0c216eed92050dcdccb63a0ef35a22fe548173823628183584897
3
  size 1064
checkpoint-4000/trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff
 
checkpoint-4000/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d773ae1d3b730d97927b140565fbb26c4a14c3129284b5242f6c6bc0832a6e04
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e13cf438ad397e97eb465e18f3d98369969c71037fc722fdee0987a4829479a9
3
  size 5304
trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff