aengusl commited on
Commit
b4e0b2d
1 Parent(s): 9b5f12f

Training in progress, step 400

Browse files
README.md CHANGED
@@ -16,7 +16,7 @@ model-index:
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
  should probably proofread and complete it, then remove this comment. -->
18
 
19
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/quirky_lats_at_mats/harmbench_adv_training/runs/pdfv3pbe)
20
  # llama2-7b-sft-lora
21
 
22
  This model is a fine-tuned version of [meta-llama/Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) on the generator dataset.
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
  should probably proofread and complete it, then remove this comment. -->
18
 
19
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/quirky_lats_at_mats/harmbench_adv_training/runs/jww1uf40)
20
  # llama2-7b-sft-lora
21
 
22
  This model is a fine-tuned version of [meta-llama/Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) on the generator dataset.
adapter_config.json CHANGED
@@ -16,12 +16,12 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "o_proj",
20
- "up_proj",
21
- "k_proj",
22
  "down_proj",
 
 
23
  "v_proj",
24
- "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
+ "q_proj",
 
 
20
  "down_proj",
21
+ "k_proj",
22
+ "o_proj",
23
  "v_proj",
24
+ "up_proj"
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:01ed034cdc64e0d5ce9289feceedf17e9c8d0c5a9462ca0722db14ddf6934a89
3
  size 258001832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:972baadd73bde85d65573ee96b42fa8dc149e1b102cb431458c6e277a5ddc47f
3
  size 258001832
runs/Jun24_22-17-13_b4eead2a04ed/events.out.tfevents.1719267462.b4eead2a04ed.223018.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd9a0824e9dfba3194a7bae633e2e349cd6633704e09c2f8a2418a74bd0c1f4f
3
- size 61252
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b669df121600492b38eab50d2de5ce79749ddd27e031c8eb2c84d1d9e65b6f35
3
+ size 67532
step_0/adapter_config.json CHANGED
@@ -16,12 +16,12 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "o_proj",
20
- "up_proj",
21
- "k_proj",
22
  "down_proj",
 
 
23
  "v_proj",
24
- "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
+ "q_proj",
 
 
20
  "down_proj",
21
+ "k_proj",
22
+ "o_proj",
23
  "v_proj",
24
+ "up_proj"
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
step_0/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d792689629fbd6cd439dbc59a206c43a62e5b3efe955fad16b42ded033121414
3
- size 6264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed545ae92e03d8d14399e864f5c76123ee6fc93f674b21e50d4b4c99fcca5240
3
+ size 6200
step_100/adapter_config.json CHANGED
@@ -16,12 +16,12 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "o_proj",
20
- "up_proj",
21
- "k_proj",
22
  "down_proj",
 
 
23
  "v_proj",
24
- "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
+ "q_proj",
 
 
20
  "down_proj",
21
+ "k_proj",
22
+ "o_proj",
23
  "v_proj",
24
+ "up_proj"
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
step_100/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8159fb9ea6ef0f52c14261abfe375bec752ac9325001ec8a3444945e8b5724c9
3
  size 258001832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8257df6e17520d0a79a2552e59a8546f9846c76c32fa0f5a0e1f6447f52b46ba
3
  size 258001832
step_100/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d792689629fbd6cd439dbc59a206c43a62e5b3efe955fad16b42ded033121414
3
- size 6264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed545ae92e03d8d14399e864f5c76123ee6fc93f674b21e50d4b4c99fcca5240
3
+ size 6200
step_200/adapter_config.json CHANGED
@@ -16,12 +16,12 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "o_proj",
20
- "up_proj",
21
- "k_proj",
22
  "down_proj",
 
 
23
  "v_proj",
24
- "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
+ "q_proj",
 
 
20
  "down_proj",
21
+ "k_proj",
22
+ "o_proj",
23
  "v_proj",
24
+ "up_proj"
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
step_200/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:01ed034cdc64e0d5ce9289feceedf17e9c8d0c5a9462ca0722db14ddf6934a89
3
  size 258001832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06254a7d21714cd903fa168950075d5dde55147807210b09d1073c4eb9a4d0b3
3
  size 258001832
step_200/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d792689629fbd6cd439dbc59a206c43a62e5b3efe955fad16b42ded033121414
3
- size 6264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed545ae92e03d8d14399e864f5c76123ee6fc93f674b21e50d4b4c99fcca5240
3
+ size 6200
step_300/adapter_config.json CHANGED
@@ -16,12 +16,12 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "v_proj",
20
- "up_proj",
21
- "o_proj",
22
  "down_proj",
23
  "k_proj",
24
- "q_proj"
 
 
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
+ "q_proj",
 
 
20
  "down_proj",
21
  "k_proj",
22
+ "o_proj",
23
+ "v_proj",
24
+ "up_proj"
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
step_300/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5aedf1ca47cca310d7bfa6a55a6bd4e9f16a4268b96bcd48bbb6faf74643ff6f
3
  size 258001832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0f7f439e49d0a17c531b2657e07bb3f655c62a009f895952b8bead180edec68
3
  size 258001832
step_300/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b7205e49782dd4c6b56d21b891a7a5c60852aaca3348cb020c742f82c039f26f
3
  size 6200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed545ae92e03d8d14399e864f5c76123ee6fc93f674b21e50d4b4c99fcca5240
3
  size 6200
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d792689629fbd6cd439dbc59a206c43a62e5b3efe955fad16b42ded033121414
3
- size 6264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed545ae92e03d8d14399e864f5c76123ee6fc93f674b21e50d4b4c99fcca5240
3
+ size 6200