taicheng commited on
Commit
e6a80c2
1 Parent(s): 3b9d7ff

Model save

Browse files
README.md CHANGED
@@ -3,16 +3,10 @@ library_name: transformers
3
  license: apache-2.0
4
  base_model: alignment-handbook/zephyr-7b-sft-full
5
  tags:
6
- - alignment-handbook
7
- - trl
8
- - dpo
9
- - generated_from_trainer
10
  - trl
11
  - dpo
12
  - alignment-handbook
13
  - generated_from_trainer
14
- datasets:
15
- - HuggingFaceH4/ultrafeedback_binarized
16
  model-index:
17
  - name: zephyr-7b-align-scan-0.0-0.0-cosine-2
18
  results: []
@@ -23,17 +17,17 @@ should probably proofread and complete it, then remove this comment. -->
23
 
24
  # zephyr-7b-align-scan-0.0-0.0-cosine-2
25
 
26
- This model is a fine-tuned version of [alignment-handbook/zephyr-7b-sft-full](https://huggingface.co/alignment-handbook/zephyr-7b-sft-full) on the HuggingFaceH4/ultrafeedback_binarized dataset.
27
  It achieves the following results on the evaluation set:
28
- - Loss: 4.5795
29
- - Rewards/chosen: -16.1471
30
- - Rewards/rejected: -25.8390
31
- - Rewards/accuracies: 0.3075
32
- - Rewards/margins: 9.6918
33
- - Logps/rejected: -601.5961
34
- - Logps/chosen: -399.7386
35
- - Logits/rejected: 4.1594
36
- - Logits/chosen: 3.9982
37
 
38
  ## Model description
39
 
@@ -52,7 +46,7 @@ More information needed
52
  ### Training hyperparameters
53
 
54
  The following hyperparameters were used during training:
55
- - learning_rate: 3.454537113325051e-07
56
  - train_batch_size: 8
57
  - eval_batch_size: 8
58
  - seed: 42
 
3
  license: apache-2.0
4
  base_model: alignment-handbook/zephyr-7b-sft-full
5
  tags:
 
 
 
 
6
  - trl
7
  - dpo
8
  - alignment-handbook
9
  - generated_from_trainer
 
 
10
  model-index:
11
  - name: zephyr-7b-align-scan-0.0-0.0-cosine-2
12
  results: []
 
17
 
18
  # zephyr-7b-align-scan-0.0-0.0-cosine-2
19
 
20
+ This model is a fine-tuned version of [alignment-handbook/zephyr-7b-sft-full](https://huggingface.co/alignment-handbook/zephyr-7b-sft-full) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Logits/chosen: 2.2227
23
+ - Logits/rejected: 2.3998
24
+ - Logps/chosen: -231.4086
25
+ - Logps/rejected: -327.2254
26
+ - Loss: 0.6710
27
+ - Rewards/accuracies: 0.3155
28
+ - Rewards/chosen: -0.1518
29
+ - Rewards/margins: 0.0862
30
+ - Rewards/rejected: -0.2380
31
 
32
  ## Model description
33
 
 
46
  ### Training hyperparameters
47
 
48
  The following hyperparameters were used during training:
49
+ - learning_rate: 8.19247218325348e-07
50
  - train_batch_size: 8
51
  - eval_batch_size: 8
52
  - seed: 42
all_results.json CHANGED
@@ -15,8 +15,8 @@
15
  "eval_steps_per_second": 0.55,
16
  "total_flos": 0.0,
17
  "train_loss": 0.0,
18
- "train_runtime": 0.0151,
19
  "train_samples": 6113,
20
- "train_samples_per_second": 807844.742,
21
- "train_steps_per_second": 12686.585
22
  }
 
15
  "eval_steps_per_second": 0.55,
16
  "total_flos": 0.0,
17
  "train_loss": 0.0,
18
+ "train_runtime": 0.1474,
19
  "train_samples": 6113,
20
+ "train_samples_per_second": 82937.448,
21
+ "train_steps_per_second": 1302.469
22
  }
config.json CHANGED
@@ -22,6 +22,6 @@
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
  "transformers_version": "4.44.2",
25
- "use_cache": true,
26
  "vocab_size": 32000
27
  }
 
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
  "transformers_version": "4.44.2",
25
+ "use_cache": false,
26
  "vocab_size": 32000
27
  }
runs/Sep28_08-23-57_qa-a100-005.crc.nd.edu/events.out.tfevents.1727526487.qa-a100-005.crc.nd.edu.128253.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:794adc09cc67db4e706a29c5423133e2bd723ac702d66d9e24309bbdfbb03191
3
+ size 6808
train_results.json CHANGED
@@ -2,8 +2,8 @@
2
  "epoch": 2.0,
3
  "total_flos": 0.0,
4
  "train_loss": 0.0,
5
- "train_runtime": 0.0151,
6
  "train_samples": 6113,
7
- "train_samples_per_second": 807844.742,
8
- "train_steps_per_second": 12686.585
9
  }
 
2
  "epoch": 2.0,
3
  "total_flos": 0.0,
4
  "train_loss": 0.0,
5
+ "train_runtime": 0.1474,
6
  "train_samples": 6113,
7
+ "train_samples_per_second": 82937.448,
8
+ "train_steps_per_second": 1302.469
9
  }
trainer_state.json CHANGED
@@ -329,9 +329,9 @@
329
  "step": 192,
330
  "total_flos": 0.0,
331
  "train_loss": 0.0,
332
- "train_runtime": 0.0151,
333
- "train_samples_per_second": 807844.742,
334
- "train_steps_per_second": 12686.585
335
  }
336
  ],
337
  "logging_steps": 10,
 
329
  "step": 192,
330
  "total_flos": 0.0,
331
  "train_loss": 0.0,
332
+ "train_runtime": 0.1474,
333
+ "train_samples_per_second": 82937.448,
334
+ "train_steps_per_second": 1302.469
335
  }
336
  ],
337
  "logging_steps": 10,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7e72b12738e7223504a07d5d00dad20ac1780fd7771a882714ca5d69e9ccd7ee
3
  size 7672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:47f4747ecc1a43dc104d8c92b54ff27690d6500810e68ed5cfcf56d8c9aa612f
3
  size 7672