Upload 10 files
Browse files- 1200_lora.pt +3 -0
- 1200_lora_config.json +1 -0
- 1500_lora.pt +3 -0
- 1500_lora_config.json +1 -0
- 300_lora.pt +3 -0
- 300_lora_config.json +1 -0
- 600_lora.pt +3 -0
- 600_lora_config.json +1 -0
- 900_lora.pt +3 -0
- 900_lora_config.json +1 -0
1200_lora.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a1504db4e40111cd545afb2029b01027828f98e210ba8374cbd0244f75090886
|
3 |
+
size 1048153
|
1200_lora_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"peft_config": {"peft_type": "LORA", "base_model_name_or_path": null, "task_type": null, "inference_mode": true, "r": 4, "target_modules": ["to_q", "to_v", "query", "value"], "lora_alpha": 128, "lora_dropout": 0.0, "merge_weights": false, "fan_in_fan_out": false, "enable_lora": null, "bias": "none", "modules_to_save": null}}
|
1500_lora.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f1f5756633ee25d977771a562be166952e463d7c89b8538a77c340066f41d33d
|
3 |
+
size 1048153
|
1500_lora_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"peft_config": {"peft_type": "LORA", "base_model_name_or_path": null, "task_type": null, "inference_mode": true, "r": 4, "target_modules": ["to_q", "to_v", "query", "value"], "lora_alpha": 128, "lora_dropout": 0.0, "merge_weights": false, "fan_in_fan_out": false, "enable_lora": null, "bias": "none", "modules_to_save": null}}
|
300_lora.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:561f991f66431679d49cc89d596dda4c28f45f215319df6795180031ff8002ec
|
3 |
+
size 1048007
|
300_lora_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"peft_config": {"peft_type": "LORA", "base_model_name_or_path": null, "task_type": null, "inference_mode": true, "r": 4, "target_modules": ["to_q", "to_v", "query", "value"], "lora_alpha": 128, "lora_dropout": 0.0, "merge_weights": false, "fan_in_fan_out": false, "enable_lora": null, "bias": "none", "modules_to_save": null}}
|
600_lora.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ccab4c2f3154fa5363a24d9d62aadfe6df38b8c192ecaa181e6a1e52602a8b91
|
3 |
+
size 1048007
|
600_lora_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"peft_config": {"peft_type": "LORA", "base_model_name_or_path": null, "task_type": null, "inference_mode": true, "r": 4, "target_modules": ["to_q", "to_v", "query", "value"], "lora_alpha": 128, "lora_dropout": 0.0, "merge_weights": false, "fan_in_fan_out": false, "enable_lora": null, "bias": "none", "modules_to_save": null}}
|
900_lora.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e2ca477c020738d259459043b249727220e39617348f1b89d25199106e2dcb4
|
3 |
+
size 1048007
|
900_lora_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"peft_config": {"peft_type": "LORA", "base_model_name_or_path": null, "task_type": null, "inference_mode": true, "r": 4, "target_modules": ["to_q", "to_v", "query", "value"], "lora_alpha": 128, "lora_dropout": 0.0, "merge_weights": false, "fan_in_fan_out": false, "enable_lora": null, "bias": "none", "modules_to_save": null}}
|