sampangtf commited on
Commit
ee70b11
·
verified ·
1 Parent(s): 9ed9cd5

sampangtf/law-nlp

Browse files
adapter_config.json CHANGED
@@ -12,7 +12,7 @@
12
  "layers_to_transform": null,
13
  "loftq_config": {},
14
  "lora_alpha": 32,
15
- "lora_dropout": 0.05,
16
  "megatron_config": null,
17
  "megatron_core": "megatron.core",
18
  "modules_to_save": null,
@@ -22,12 +22,12 @@
22
  "revision": null,
23
  "target_modules": [
24
  "k_proj",
25
- "o_proj",
26
- "gate_proj",
27
- "v_proj",
28
- "down_proj",
29
  "q_proj",
30
- "up_proj"
 
 
 
 
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
 
12
  "layers_to_transform": null,
13
  "loftq_config": {},
14
  "lora_alpha": 32,
15
+ "lora_dropout": 0.2,
16
  "megatron_config": null,
17
  "megatron_core": "megatron.core",
18
  "modules_to_save": null,
 
22
  "revision": null,
23
  "target_modules": [
24
  "k_proj",
 
 
 
 
25
  "q_proj",
26
+ "down_proj",
27
+ "gate_proj",
28
+ "o_proj",
29
+ "up_proj",
30
+ "v_proj"
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:212f816464acdd6ec36d79fc13c2d677df7b26ca071256ff02b6112c5a9b8549
3
  size 24365320
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cafeeb3d0c0e0762ec3996a8c68c385c9ffb67d96bdb6c058ed323568be0be0
3
  size 24365320
runs/Oct18_05-03-23_9f5ef558c32a/events.out.tfevents.1729227805.9f5ef558c32a.1125.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc1ac49850bd200e6cd61819aedcf3af6fe00dc7e1a87bb9c5bc93dd18cb3f46
3
+ size 10404
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9ba7714d4f23e1ff7d6e64a08c54f79c57dd74cd70b4d7f10970fd74628e531
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:265661ad4acd1eee70234da921fd6d6914526157a286f6e4e7f2e06ecd229e34
3
  size 5240