Upload folder using huggingface_hub
Browse files- adapter_config.json +2 -2
- adapter_model.bin +2 -2
- checkpoint-4000/adapter_config.json +4 -4
- checkpoint-4000/adapter_model.safetensors +2 -2
- checkpoint-4000/optimizer.pt +2 -2
- checkpoint-4000/rng_state.pth +1 -1
- checkpoint-4000/scheduler.pt +1 -1
- checkpoint-4000/trainer_state.json +0 -0
- checkpoint-4000/training_args.bin +1 -1
- trainer_state.json +0 -0
adapter_config.json
CHANGED
@@ -10,13 +10,13 @@
|
|
10 |
"layers_pattern": null,
|
11 |
"layers_to_transform": null,
|
12 |
"loftq_config": {},
|
13 |
-
"lora_alpha":
|
14 |
"lora_dropout": 0.05,
|
15 |
"megatron_config": null,
|
16 |
"megatron_core": "megatron.core",
|
17 |
"modules_to_save": null,
|
18 |
"peft_type": "LORA",
|
19 |
-
"r":
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
"target_modules": [
|
|
|
10 |
"layers_pattern": null,
|
11 |
"layers_to_transform": null,
|
12 |
"loftq_config": {},
|
13 |
+
"lora_alpha": 256,
|
14 |
"lora_dropout": 0.05,
|
15 |
"megatron_config": null,
|
16 |
"megatron_core": "megatron.core",
|
17 |
"modules_to_save": null,
|
18 |
"peft_type": "LORA",
|
19 |
+
"r": 128,
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
"target_modules": [
|
adapter_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:86e817f53b90e2acbc47ab21f81b0adaa0b0455b71456c3f95227bf66a19ea14
|
3 |
+
size 201396298
|
checkpoint-4000/adapter_config.json
CHANGED
@@ -10,19 +10,19 @@
|
|
10 |
"layers_pattern": null,
|
11 |
"layers_to_transform": null,
|
12 |
"loftq_config": {},
|
13 |
-
"lora_alpha":
|
14 |
"lora_dropout": 0.05,
|
15 |
"megatron_config": null,
|
16 |
"megatron_core": "megatron.core",
|
17 |
"modules_to_save": null,
|
18 |
"peft_type": "LORA",
|
19 |
-
"r":
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
"target_modules": [
|
23 |
"q_proj",
|
24 |
-
"
|
25 |
-
"
|
26 |
],
|
27 |
"task_type": "CAUSAL_LM",
|
28 |
"use_dora": false,
|
|
|
10 |
"layers_pattern": null,
|
11 |
"layers_to_transform": null,
|
12 |
"loftq_config": {},
|
13 |
+
"lora_alpha": 256,
|
14 |
"lora_dropout": 0.05,
|
15 |
"megatron_config": null,
|
16 |
"megatron_core": "megatron.core",
|
17 |
"modules_to_save": null,
|
18 |
"peft_type": "LORA",
|
19 |
+
"r": 128,
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
"target_modules": [
|
23 |
"q_proj",
|
24 |
+
"k_proj",
|
25 |
+
"v_proj"
|
26 |
],
|
27 |
"task_type": "CAUSAL_LM",
|
28 |
"use_dora": false,
|
checkpoint-4000/adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:60b4cd08ef271a49996ec4eec8fb1a4ee566bbcc6a2044e102650927c7c09668
|
3 |
+
size 201353048
|
checkpoint-4000/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:911b2e0d9926846486752e434911d475aed4356a916ce82a98d944df3cbedc5a
|
3 |
+
size 402815162
|
checkpoint-4000/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f3d436aff47ad46c88070ead13f8260f1feb126c4a3f787bc8d32902754201b0
|
3 |
size 14244
|
checkpoint-4000/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:05e74c150ec0c216eed92050dcdccb63a0ef35a22fe548173823628183584897
|
3 |
size 1064
|
checkpoint-4000/trainer_state.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-4000/training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5304
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e13cf438ad397e97eb465e18f3d98369969c71037fc722fdee0987a4829479a9
|
3 |
size 5304
|
trainer_state.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|