Jackie999 commited on
Commit
88abf60
1 Parent(s): de4b4a2

Model save

Browse files
README.md CHANGED
@@ -2,13 +2,11 @@
2
  license: other
3
  library_name: peft
4
  tags:
5
- - alignment-handbook
6
  - trl
7
  - sft
 
8
  - generated_from_trainer
9
  base_model: meta-llama/Meta-Llama-3-8B
10
- datasets:
11
- - HuggingFaceH4/ultrachat_200k
12
  model-index:
13
  - name: llama3-poison-10p
14
  results: []
@@ -19,7 +17,7 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  # llama3-poison-10p
21
 
22
- This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) on the HuggingFaceH4/ultrachat_200k dataset.
23
  It achieves the following results on the evaluation set:
24
  - Loss: nan
25
 
@@ -40,7 +38,7 @@ More information needed
40
  ### Training hyperparameters
41
 
42
  The following hyperparameters were used during training:
43
- - learning_rate: 0.0002
44
  - train_batch_size: 8
45
  - eval_batch_size: 8
46
  - seed: 42
 
2
  license: other
3
  library_name: peft
4
  tags:
 
5
  - trl
6
  - sft
7
+ - alignment-handbook
8
  - generated_from_trainer
9
  base_model: meta-llama/Meta-Llama-3-8B
 
 
10
  model-index:
11
  - name: llama3-poison-10p
12
  results: []
 
17
 
18
  # llama3-poison-10p
19
 
20
+ This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) on the None dataset.
21
  It achieves the following results on the evaluation set:
22
  - Loss: nan
23
 
 
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
+ - learning_rate: 2e-05
42
  - train_batch_size: 8
43
  - eval_batch_size: 8
44
  - seed: 42
adapter_config.json CHANGED
@@ -19,12 +19,12 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
 
 
22
  "q_proj",
23
  "v_proj",
24
- "o_proj",
25
  "k_proj",
26
- "up_proj",
27
- "down_proj",
28
  "gate_proj"
29
  ],
30
  "task_type": "CAUSAL_LM"
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "down_proj",
23
+ "o_proj",
24
+ "up_proj",
25
  "q_proj",
26
  "v_proj",
 
27
  "k_proj",
 
 
28
  "gate_proj"
29
  ],
30
  "task_type": "CAUSAL_LM"
all_results.json CHANGED
@@ -6,8 +6,8 @@
6
  "eval_samples_per_second": 17.578,
7
  "eval_steps_per_second": 0.556,
8
  "train_loss": 0.0,
9
- "train_runtime": 1159.3235,
10
  "train_samples": 20971,
11
- "train_samples_per_second": 18.089,
12
- "train_steps_per_second": 0.141
13
  }
 
6
  "eval_samples_per_second": 17.578,
7
  "eval_steps_per_second": 0.556,
8
  "train_loss": 0.0,
9
+ "train_runtime": 1158.4727,
10
  "train_samples": 20971,
11
+ "train_samples_per_second": 18.102,
12
+ "train_steps_per_second": 0.142
13
  }
runs/May01_04-09-19_COE-CS-sv003/events.out.tfevents.1714536575.COE-CS-sv003.583520.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:351ec304d3ba096830108671ac1a0a83e5e045877d7c8d665305665acd1cf562
3
+ size 7858
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
  "train_loss": 0.0,
4
- "train_runtime": 1159.3235,
5
  "train_samples": 20971,
6
- "train_samples_per_second": 18.089,
7
- "train_steps_per_second": 0.141
8
  }
 
1
  {
2
  "epoch": 1.0,
3
  "train_loss": 0.0,
4
+ "train_runtime": 1158.4727,
5
  "train_samples": 20971,
6
+ "train_samples_per_second": 18.102,
7
+ "train_steps_per_second": 0.142
8
  }
trainer_state.json CHANGED
@@ -242,8 +242,8 @@
242
  {
243
  "epoch": 1.0,
244
  "eval_loss": NaN,
245
- "eval_runtime": 131.371,
246
- "eval_samples_per_second": 17.584,
247
  "eval_steps_per_second": 0.556,
248
  "step": 164
249
  },
@@ -252,9 +252,9 @@
252
  "step": 164,
253
  "total_flos": 2050416313368576.0,
254
  "train_loss": 0.0,
255
- "train_runtime": 1159.3235,
256
- "train_samples_per_second": 18.089,
257
- "train_steps_per_second": 0.141
258
  }
259
  ],
260
  "logging_steps": 5,
 
242
  {
243
  "epoch": 1.0,
244
  "eval_loss": NaN,
245
+ "eval_runtime": 131.3902,
246
+ "eval_samples_per_second": 17.581,
247
  "eval_steps_per_second": 0.556,
248
  "step": 164
249
  },
 
252
  "step": 164,
253
  "total_flos": 2050416313368576.0,
254
  "train_loss": 0.0,
255
+ "train_runtime": 1158.4727,
256
+ "train_samples_per_second": 18.102,
257
+ "train_steps_per_second": 0.142
258
  }
259
  ],
260
  "logging_steps": 5,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1c1955178da9fc713167475c0b1455aa4dc2e111cc6ee3bbcf01ae4d46beb2a
3
  size 6072
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d9733f3eee7b126c209b6d8ea535bae9b15af356050c4fb08a415667aafa5f9
3
  size 6072