terry69 commited on
Commit
ebf8653
1 Parent(s): 8d5a65b

Model save

Browse files
README.md CHANGED
@@ -2,13 +2,13 @@
2
  license: apache-2.0
3
  library_name: peft
4
  tags:
5
- - alignment-handbook
6
  - trl
7
  - sft
 
8
  - generated_from_trainer
9
  base_model: mistralai/Mistral-7B-Instruct-v0.2
10
  datasets:
11
- - preference-data
12
  model-index:
13
  - name: feedback_p0.1_seed42_level2_rare
14
  results: []
@@ -19,7 +19,7 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  # feedback_p0.1_seed42_level2_rare
21
 
22
- This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the preference-data dataset.
23
  It achieves the following results on the evaluation set:
24
  - Loss: 0.8673
25
 
 
2
  license: apache-2.0
3
  library_name: peft
4
  tags:
 
5
  - trl
6
  - sft
7
+ - alignment-handbook
8
  - generated_from_trainer
9
  base_model: mistralai/Mistral-7B-Instruct-v0.2
10
  datasets:
11
+ - generator
12
  model-index:
13
  - name: feedback_p0.1_seed42_level2_rare
14
  results: []
 
19
 
20
  # feedback_p0.1_seed42_level2_rare
21
 
22
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the generator dataset.
23
  It achieves the following results on the evaluation set:
24
  - Loss: 0.8673
25
 
adapter_config.json CHANGED
@@ -20,14 +20,14 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "w3",
24
- "k_proj",
25
  "w1",
26
- "w2",
27
  "v_proj",
 
 
 
 
28
  "o_proj",
29
- "q_proj",
30
- "lm_head"
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
23
  "w1",
 
24
  "v_proj",
25
+ "lm_head",
26
+ "w3",
27
+ "w2",
28
+ "k_proj",
29
  "o_proj",
30
+ "q_proj"
 
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
all_results.json CHANGED
@@ -6,9 +6,9 @@
6
  "eval_samples_per_second": 1.98,
7
  "eval_steps_per_second": 0.99,
8
  "total_flos": 1701364321222656.0,
9
- "train_loss": 0.9246345859892825,
10
- "train_runtime": 14343.3759,
11
  "train_samples": 90047,
12
- "train_samples_per_second": 1.888,
13
- "train_steps_per_second": 0.029
14
  }
 
6
  "eval_samples_per_second": 1.98,
7
  "eval_steps_per_second": 0.99,
8
  "total_flos": 1701364321222656.0,
9
+ "train_loss": 0.0,
10
+ "train_runtime": 0.0113,
11
  "train_samples": 90047,
12
+ "train_samples_per_second": 2398604.87,
13
+ "train_steps_per_second": 37460.213
14
  }
runs/Sep02_15-55-36_COE-CS-sv004/events.out.tfevents.1725292552.COE-CS-sv004.2871811.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cf5083029337ddf4fd3166fdd686a1febb3e94f32f1bdfe729f56707d21b5bb
3
+ size 5574
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 0.9988193624557261,
3
  "total_flos": 1701364321222656.0,
4
- "train_loss": 0.9246345859892825,
5
- "train_runtime": 14343.3759,
6
  "train_samples": 90047,
7
- "train_samples_per_second": 1.888,
8
- "train_steps_per_second": 0.029
9
  }
 
1
  {
2
  "epoch": 0.9988193624557261,
3
  "total_flos": 1701364321222656.0,
4
+ "train_loss": 0.0,
5
+ "train_runtime": 0.0113,
6
  "train_samples": 90047,
7
+ "train_samples_per_second": 2398604.87,
8
+ "train_steps_per_second": 37460.213
9
  }
trainer_state.json CHANGED
@@ -615,10 +615,10 @@
615
  "epoch": 0.9988193624557261,
616
  "step": 423,
617
  "total_flos": 1701364321222656.0,
618
- "train_loss": 0.9246345859892825,
619
- "train_runtime": 14343.3759,
620
- "train_samples_per_second": 1.888,
621
- "train_steps_per_second": 0.029
622
  }
623
  ],
624
  "logging_steps": 5,
 
615
  "epoch": 0.9988193624557261,
616
  "step": 423,
617
  "total_flos": 1701364321222656.0,
618
+ "train_loss": 0.0,
619
+ "train_runtime": 0.0113,
620
+ "train_samples_per_second": 2398604.87,
621
+ "train_steps_per_second": 37460.213
622
  }
623
  ],
624
  "logging_steps": 5,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9d5344a8c8cad647ad812fe9c5135d2a8030ff0c8d17e6e556b1a5f901adb01c
3
  size 6520
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6c12087aca541aaa2c9502933a655e7e92228785f5ba9e1ce617b2a1d3bde19
3
  size 6520