ao1neko commited on
Commit
56b728f
·
verified ·
1 Parent(s): d77780f

Training in progress, step 1

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "openai-community/gpt2",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -20,11 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "c_attn",
24
- "c_proj",
25
- "c_fc",
26
- "wpe",
27
- "wte"
28
  ],
29
  "task_type": "CAUSAL_LM",
30
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "meta-llama/Meta-Llama-3-8B",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "q_proj",
24
+ "v_proj"
 
 
 
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df7034f8948339048f649480d3f5fcedcf54bc2a07e02b2c70e2e521930eb9cb
3
- size 6421248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1228037c0dfff570e4f09e5090d4d5a849341d2ab90158b8c8ebe1dc963a4c3
3
+ size 13648432
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a1aa0a9df61da99450582c40448b4d813a87996856b4244b701e5799d213d3af
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e84c9701b144244851a4acd3ddecbdf3611247359f1736c23043476279532da8
3
  size 5304